WorldWideScience

Sample records for algorithm identifies community

  1. An efficient community detection algorithm using greedy surprise maximization

    International Nuclear Information System (INIS)

    Jiang, Yawen; Jia, Caiyan; Yu, Jian

    2014-01-01

    Community detection is an important and crucial problem in complex network analysis. Although classical modularity function optimization approaches are widely used for identifying communities, the modularity function (Q) suffers from its resolution limit. Recently, the surprise function (S) was experimentally proved to be better than the Q function. However, up until now, there has been no algorithm available to perform searches to directly determine the maximal surprise values. In this paper, considering the superiority of the S function over the Q function, we propose an efficient community detection algorithm called AGSO (algorithm based on greedy surprise optimization) and its improved version FAGSO (fast-AGSO), which are based on greedy surprise optimization and do not suffer from the resolution limit. In addition, (F)AGSO does not need the number of communities K to be specified in advance. Tests on experimental networks show that (F)AGSO is able to detect optimal partitions in both simple and even more complex networks. Moreover, algorithms based on surprise maximization perform better than those algorithms based on modularity maximization, including Blondel–Guillaume–Lambiotte–Lefebvre (BGLL), Clauset–Newman–Moore (CNM) and the other state-of-the-art algorithms such as Infomap, order statistics local optimization method (OSLOM) and label propagation algorithm (LPA). (paper)

  2. Information dynamics algorithm for detecting communities in networks

    Science.gov (United States)

    Massaro, Emanuele; Bagnoli, Franco; Guazzini, Andrea; Lió, Pietro

    2012-11-01

    The problem of community detection is relevant in many scientific disciplines, from social science to statistical physics. Given the impact of community detection in many areas, such as psychology and social sciences, we have addressed the issue of modifying existing well performing algorithms by incorporating elements of the domain application fields, i.e. domain-inspired. We have focused on a psychology and social network-inspired approach which may be useful for further strengthening the link between social network studies and mathematics of community detection. Here we introduce a community-detection algorithm derived from the van Dongen's Markov Cluster algorithm (MCL) method [4] by considering networks' nodes as agents capable to take decisions. In this framework we have introduced a memory factor to mimic a typical human behavior such as the oblivion effect. The method is based on information diffusion and it includes a non-linear processing phase. We test our method on two classical community benchmark and on computer generated networks with known community structure. Our approach has three important features: the capacity of detecting overlapping communities, the capability of identifying communities from an individual point of view and the fine tuning the community detectability with respect to prior knowledge of the data. Finally we discuss how to use a Shannon entropy measure for parameter estimation in complex networks.

  3. A Greedy Algorithm for Neighborhood Overlap-Based Community Detection

    Directory of Open Access Journals (Sweden)

    Natarajan Meghanathan

    2016-01-01

    Full Text Available The neighborhood overlap (NOVER of an edge u-v is defined as the ratio of the number of nodes who are neighbors for both u and v to that of the number of nodes who are neighbors of at least u or v. In this paper, we hypothesize that an edge u-v with a lower NOVER score bridges two or more sets of vertices, with very few edges (other than u-v connecting vertices from one set to another set. Accordingly, we propose a greedy algorithm of iteratively removing the edges of a network in the increasing order of their neighborhood overlap and calculating the modularity score of the resulting network component(s after the removal of each edge. The network component(s that have the largest cumulative modularity score are identified as the different communities of the network. We evaluate the performance of the proposed NOVER-based community detection algorithm on nine real-world network graphs and compare the performance against the multi-level aggregation-based Louvain algorithm, as well as the original and time-efficient versions of the edge betweenness-based Girvan-Newman (GN community detection algorithm.

  4. Local Community Detection Algorithm Based on Minimal Cluster

    Directory of Open Access Journals (Sweden)

    Yong Zhou

    2016-01-01

    Full Text Available In order to discover the structure of local community more effectively, this paper puts forward a new local community detection algorithm based on minimal cluster. Most of the local community detection algorithms begin from one node. The agglomeration ability of a single node must be less than multiple nodes, so the beginning of the community extension of the algorithm in this paper is no longer from the initial node only but from a node cluster containing this initial node and nodes in the cluster are relatively densely connected with each other. The algorithm mainly includes two phases. First it detects the minimal cluster and then finds the local community extended from the minimal cluster. Experimental results show that the quality of the local community detected by our algorithm is much better than other algorithms no matter in real networks or in simulated networks.

  5. Community detection algorithm evaluation with ground-truth data

    Science.gov (United States)

    Jebabli, Malek; Cherifi, Hocine; Cherifi, Chantal; Hamouda, Atef

    2018-02-01

    Community structure is of paramount importance for the understanding of complex networks. Consequently, there is a tremendous effort in order to develop efficient community detection algorithms. Unfortunately, the issue of a fair assessment of these algorithms is a thriving open question. If the ground-truth community structure is available, various clustering-based metrics are used in order to compare it versus the one discovered by these algorithms. However, these metrics defined at the node level are fairly insensitive to the variation of the overall community structure. To overcome these limitations, we propose to exploit the topological features of the 'community graphs' (where the nodes are the communities and the links represent their interactions) in order to evaluate the algorithms. To illustrate our methodology, we conduct a comprehensive analysis of overlapping community detection algorithms using a set of real-world networks with known a priori community structure. Results provide a better perception of their relative performance as compared to classical metrics. Moreover, they show that more emphasis should be put on the topology of the community structure. We also investigate the relationship between the topological properties of the community structure and the alternative evaluation measures (quality metrics and clustering metrics). It appears clearly that they present different views of the community structure and that they must be combined in order to evaluate the effectiveness of community detection algorithms.

  6. A Modularity Degree Based Heuristic Community Detection Algorithm

    Directory of Open Access Journals (Sweden)

    Dongming Chen

    2014-01-01

    Full Text Available A community in a complex network can be seen as a subgroup of nodes that are densely connected. Discovery of community structures is a basic problem of research and can be used in various areas, such as biology, computer science, and sociology. Existing community detection methods usually try to expand or collapse the nodes partitions in order to optimize a given quality function. These optimization function based methods share the same drawback of inefficiency. Here we propose a heuristic algorithm (MDBH algorithm based on network structure which employs modularity degree as a measure function. Experiments on both synthetic benchmarks and real-world networks show that our algorithm gives competitive accuracy with previous modularity optimization methods, even though it has less computational complexity. Furthermore, due to the use of modularity degree, our algorithm naturally improves the resolution limit in community detection.

  7. Identifying and characterizing key nodes among communities based on electrical-circuit networks.

    Science.gov (United States)

    Zhu, Fenghui; Wang, Wenxu; Di, Zengru; Fan, Ying

    2014-01-01

    Complex networks with community structures are ubiquitous in the real world. Despite many approaches developed for detecting communities, we continue to lack tools for identifying overlapping and bridging nodes that play crucial roles in the interactions and communications among communities in complex networks. Here we develop an algorithm based on the local flow conservation to effectively and efficiently identify and distinguish the two types of nodes. Our method is applicable in both undirected and directed networks without a priori knowledge of the community structure. Our method bypasses the extremely challenging problem of partitioning communities in the presence of overlapping nodes that may belong to multiple communities. Due to the fact that overlapping and bridging nodes are of paramount importance in maintaining the function of many social and biological networks, our tools open new avenues towards understanding and controlling real complex networks with communities accompanied with the key nodes.

  8. Fast detection of the fuzzy communities based on leader-driven algorithm

    Science.gov (United States)

    Fang, Changjian; Mu, Dejun; Deng, Zhenghong; Hu, Jun; Yi, Chen-He

    2018-03-01

    In this paper, we present the leader-driven algorithm (LDA) for learning community structure in networks. The algorithm allows one to find overlapping clusters in a network, an important aspect of real networks, especially social networks. The algorithm requires no input parameters and learns the number of clusters naturally from the network. It accomplishes this using leadership centrality in a clever manner. It identifies local minima of leadership centrality as followers which belong only to one cluster, and the remaining nodes are leaders which connect clusters. In this way, the number of clusters can be learned using only the network structure. The LDA is also an extremely fast algorithm, having runtime linear in the network size. Thus, this algorithm can be used to efficiently cluster extremely large networks.

  9. Performance of an electronic health record-based phenotype algorithm to identify community associated methicillin-resistant Staphylococcus aureus cases and controls for genetic association studies

    Directory of Open Access Journals (Sweden)

    Kathryn L. Jackson

    2016-11-01

    Full Text Available Abstract Background Community associated methicillin-resistant Staphylococcus aureus (CA-MRSA is one of the most common causes of skin and soft tissue infections in the United States, and a variety of genetic host factors are suspected to be risk factors for recurrent infection. Based on the CDC definition, we have developed and validated an electronic health record (EHR based CA-MRSA phenotype algorithm utilizing both structured and unstructured data. Methods The algorithm was validated at three eMERGE consortium sites, and positive predictive value, negative predictive value and sensitivity, were calculated. The algorithm was then run and data collected across seven total sites. The resulting data was used in GWAS analysis. Results Across seven sites, the CA-MRSA phenotype algorithm identified a total of 349 cases and 7761 controls among the genotyped European and African American biobank populations. PPV ranged from 68 to 100% for cases and 96 to 100% for controls; sensitivity ranged from 94 to 100% for cases and 75 to 100% for controls. Frequency of cases in the populations varied widely by site. There were no plausible GWAS-significant (p < 5 E −8 findings. Conclusions Differences in EHR data representation and screening patterns across sites may have affected identification of cases and controls and accounted for varying frequencies across sites. Future work identifying these patterns is necessary.

  10. An Efficient Hierarchy Algorithm for Community Detection in Complex Networks

    Directory of Open Access Journals (Sweden)

    Lili Zhang

    2014-01-01

    Full Text Available Community structure is one of the most fundamental and important topology characteristics of complex networks. The research on community structure has wide applications and is very important for analyzing the topology structure, understanding the functions, finding the hidden properties, and forecasting the time-varying of the networks. This paper analyzes some related algorithms and proposes a new algorithm—CN agglomerative algorithm based on graph theory and the local connectedness of network to find communities in network. We show this algorithm is distributed and polynomial; meanwhile the simulations show it is accurate and fine-grained. Furthermore, we modify this algorithm to get one modified CN algorithm and apply it to dynamic complex networks, and the simulations also verify that the modified CN algorithm has high accuracy too.

  11. Discrete particle swarm optimization for identifying community structures in signed social networks.

    Science.gov (United States)

    Cai, Qing; Gong, Maoguo; Shen, Bo; Ma, Lijia; Jiao, Licheng

    2014-10-01

    Modern science of networks has facilitated us with enormous convenience to the understanding of complex systems. Community structure is believed to be one of the notable features of complex networks representing real complicated systems. Very often, uncovering community structures in networks can be regarded as an optimization problem, thus, many evolutionary algorithms based approaches have been put forward. Particle swarm optimization (PSO) is an artificial intelligent algorithm originated from social behavior such as birds flocking and fish schooling. PSO has been proved to be an effective optimization technique. However, PSO was originally designed for continuous optimization which confounds its applications to discrete contexts. In this paper, a novel discrete PSO algorithm is suggested for identifying community structures in signed networks. In the suggested method, particles' status has been redesigned in discrete form so as to make PSO proper for discrete scenarios, and particles' updating rules have been reformulated by making use of the topology of the signed network. Extensive experiments compared with three state-of-the-art approaches on both synthetic and real-world signed networks demonstrate that the proposed method is effective and promising. Copyright © 2014 Elsevier Ltd. All rights reserved.

  12. Asymmetric intimacy and algorithm for detecting communities in bipartite networks

    Science.gov (United States)

    Wang, Xingyuan; Qin, Xiaomeng

    2016-11-01

    In this paper, an algorithm to choose a good partition in bipartite networks has been proposed. Bipartite networks have more theoretical significance and broader prospect of application. In view of distinctive structure of bipartite networks, in our method, two parameters are defined to show the relationships between the same type nodes and heterogeneous nodes respectively. Moreover, our algorithm employs a new method of finding and expanding the core communities in bipartite networks. Two kinds of nodes are handled separately and merged, and then the sub-communities are obtained. After that, objective communities will be found according to the merging rule. The proposed algorithm has been simulated in real-world networks and artificial networks, and the result verifies the accuracy and reliability of the parameters on intimacy for our algorithm. Eventually, comparisons with similar algorithms depict that the proposed algorithm has better performance.

  13. Community Clustering Algorithm in Complex Networks Based on Microcommunity Fusion

    Directory of Open Access Journals (Sweden)

    Jin Qi

    2015-01-01

    Full Text Available With the further research on physical meaning and digital features of the community structure in complex networks in recent years, the improvement of effectiveness and efficiency of the community mining algorithms in complex networks has become an important subject in this area. This paper puts forward a concept of the microcommunity and gets final mining results of communities through fusing different microcommunities. This paper starts with the basic definition of the network community and applies Expansion to the microcommunity clustering which provides prerequisites for the microcommunity fusion. The proposed algorithm is more efficient and has higher solution quality compared with other similar algorithms through the analysis of test results based on network data set.

  14. An Overlapping Communities Detection Algorithm via Maxing Modularity in Opportunistic Networks

    Directory of Open Access Journals (Sweden)

    Gao Zhi-Peng

    2016-01-01

    Full Text Available Community detection in opportunistic networks has been a significant and hot issue, which is used to understand characteristics of networks through analyzing structure of it. Community is used to represent a group of nodes in a network where nodes inside the community have more internal connections than external connections. However, most of the existing community detection algorithms focus on binary networks or disjoint community detection. In this paper, we propose a novel algorithm via maxing modularity of communities (MMCto find overlapping community structure in opportunistic networks. It utilizes contact history of nodes to calculate the relation intensity between nodes. It finds nodes with high relation intensity as the initial community and extend the community with nodes of higher belong degree. The algorithm achieves a rapid and efficient overlapping community detection method by maxing the modularity of community continuously. The experiments prove that MMC is effective for uncovering overlapping communities and it achieves better performance than COPRA and Conductance.

  15. Heuristic Artificial Bee Colony Algorithm for Uncovering Community in Complex Networks

    Directory of Open Access Journals (Sweden)

    Yuquan Guo

    2017-01-01

    Full Text Available Community structure is important for us to understand the functions and structure of the complex networks. In this paper, Heuristic Artificial Bee Colony (HABC algorithm based on swarm intelligence is proposed for uncovering community. The proposed HABC includes initialization, employed bee searching, onlooker searching, and scout bee searching. In initialization stage, the nectar sources with simple community structure are generated through network dynamic algorithm associated with complete subgraph. In employed bee searching and onlooker searching stages, the searching function is redefined to address the community problem. The efficiency of searching progress can be improved by a heuristic function which is an average agglomerate probability of two neighbor communities. Experiments are carried out on artificial and real world networks, and the results demonstrate that HABC will have better performance in terms of comparing with the state-of-the-art algorithms.

  16. An Automated Summarization Assessment Algorithm for Identifying Summarizing Strategies.

    Directory of Open Access Journals (Sweden)

    Asad Abdi

    Full Text Available Summarization is a process to select important information from a source text. Summarizing strategies are the core cognitive processes in summarization activity. Since summarization can be important as a tool to improve comprehension, it has attracted interest of teachers for teaching summary writing through direct instruction. To do this, they need to review and assess the students' summaries and these tasks are very time-consuming. Thus, a computer-assisted assessment can be used to help teachers to conduct this task more effectively.This paper aims to propose an algorithm based on the combination of semantic relations between words and their syntactic composition to identify summarizing strategies employed by students in summary writing. An innovative aspect of our algorithm lies in its ability to identify summarizing strategies at the syntactic and semantic levels. The efficiency of the algorithm is measured in terms of Precision, Recall and F-measure. We then implemented the algorithm for the automated summarization assessment system that can be used to identify the summarizing strategies used by students in summary writing.

  17. An Improved Particle Swarm Optimization Algorithm and Its Application in the Community Division

    Directory of Open Access Journals (Sweden)

    Jiang Hao

    2016-01-01

    Full Text Available With the deepening of the research on complex networks, the method of detecting and classifying social network is springing up. In this essay, the basic particle swarm algorithm is improved based on the GN algorithm. Modularity is taken as a measure of community division [1]. In view of the dynamic network community division, scrolling calculation method is put forward. Experiments show that using the improved particle swarm optimization algorithm can improve the accuracy of the community division and can also get higher value of the modularity in the dynamic community

  18. A New Multiobjective Evolutionary Algorithm for Community Detection in Dynamic Complex Networks

    Directory of Open Access Journals (Sweden)

    Guoqiang Chen

    2013-01-01

    Full Text Available Community detection in dynamic networks is an important research topic and has received an enormous amount of attention in recent years. Modularity is selected as a measure to quantify the quality of the community partition in previous detection methods. But, the modularity has been exposed to resolution limits. In this paper, we propose a novel multiobjective evolutionary algorithm for dynamic networks community detection based on the framework of nondominated sorting genetic algorithm. Modularity density which can address the limitations of modularity function is adopted to measure the snapshot cost, and normalized mutual information is selected to measure temporal cost, respectively. The characteristics knowledge of the problem is used in designing the genetic operators. Furthermore, a local search operator was designed, which can improve the effectiveness and efficiency of community detection. Experimental studies based on synthetic datasets show that the proposed algorithm can obtain better performance than the compared algorithms.

  19. Identifying vital edges in Chinese air route network via memetic algorithm

    Directory of Open Access Journals (Sweden)

    Wenbo Du

    2017-02-01

    Full Text Available Due to rapid development in the past decade, air transportation system has attracted considerable research attention from diverse communities. While most of the previous studies focused on airline networks, here we systematically explore the robustness of the Chinese air route network, and identify the vital edges which form the backbone of Chinese air transportation system. Specifically, we employ a memetic algorithm to minimize the network robustness after removing certain edges, and hence the solution of this model is the set of vital edges. Counterintuitively, our results show that the most vital edges are not necessarily the edges of the highest topological importance, for which we provide an extensive explanation from the microscope view. Our findings also offer new insights to understanding and optimizing other real-world network systems.

  20. A fast algorithm for identifying friends-of-friends halos

    Science.gov (United States)

    Feng, Y.; Modi, C.

    2017-07-01

    We describe a simple and fast algorithm for identifying friends-of-friends features and prove its correctness. The algorithm avoids unnecessary expensive neighbor queries, uses minimal memory overhead, and rejects slowdown in high over-density regions. We define our algorithm formally based on pair enumeration, a problem that has been heavily studied in fast 2-point correlation codes and our reference implementation employs a dual KD-tree correlation function code. We construct features in a hierarchical tree structure, and use a splay operation to reduce the average cost of identifying the root of a feature from O [ log L ] to O [ 1 ] (L is the size of a feature) without additional memory costs. This reduces the overall time complexity of merging trees from O [ L log L ] to O [ L ] , reducing the number of operations per splay by orders of magnitude. We next introduce a pruning operation that skips merge operations between two fully self-connected KD-tree nodes. This improves the robustness of the algorithm, reducing the number of merge operations in high density peaks from O [δ2 ] to O [ δ ] . We show that for cosmological data set the algorithm eliminates more than half of merge operations for typically used linking lengths b ∼ 0 . 2 (relative to mean separation). Furthermore, our algorithm is extremely simple and easy to implement on top of an existing pair enumeration code, reusing the optimization effort that has been invested in fast correlation function codes.

  1. Semi-supervised spectral algorithms for community detection in complex networks based on equivalence of clustering methods

    Science.gov (United States)

    Ma, Xiaoke; Wang, Bingbo; Yu, Liang

    2018-01-01

    Community detection is fundamental for revealing the structure-functionality relationship in complex networks, which involves two issues-the quantitative function for community as well as algorithms to discover communities. Despite significant research on either of them, few attempt has been made to establish the connection between the two issues. To attack this problem, a generalized quantification function is proposed for community in weighted networks, which provides a framework that unifies several well-known measures. Then, we prove that the trace optimization of the proposed measure is equivalent with the objective functions of algorithms such as nonnegative matrix factorization, kernel K-means as well as spectral clustering. It serves as the theoretical foundation for designing algorithms for community detection. On the second issue, a semi-supervised spectral clustering algorithm is developed by exploring the equivalence relation via combining the nonnegative matrix factorization and spectral clustering. Different from the traditional semi-supervised algorithms, the partial supervision is integrated into the objective of the spectral algorithm. Finally, through extensive experiments on both artificial and real world networks, we demonstrate that the proposed method improves the accuracy of the traditional spectral algorithms in community detection.

  2. An improvement of the fast uncovering community algorithm

    International Nuclear Information System (INIS)

    Wang Li; Wang Jiang; Shen Hua-Wei; Cheng Xue-Qi

    2013-01-01

    Community detection methods have been used in computer, sociology, physics, biology, and brain information science areas. Many methods are based on the optimization of modularity. The algorithm proposed by Blondel et al. (Blondel V D, Guillaume J L, Lambiotte R and Lefebvre E 2008 J. Stat. Mech. 10 10008) is one of the most widely used methods because of its good performance, especially in the big data era. In this paper we make some improvements to this algorithm in correctness and performance. By tests we see that different node orders bring different performances and different community structures. We find some node swings in different communities that influence the performance. So we design some strategies on the sweeping order of node to reduce the computing cost made by repetition swing. We introduce a new concept of overlapping degree (OV) that shows the strength of connection between nodes. Three improvement strategies are proposed that are based on constant OV, adaptive OV, and adaptive weighted OV, respectively. Experiments on synthetic datasets and real datasets are made, showing that our improved strategies can improve the performance and correctness. (interdisciplinary physics and related areas of science and technology)

  3. A game theoretic algorithm to detect overlapping community structure in networks

    Science.gov (United States)

    Zhou, Xu; Zhao, Xiaohui; Liu, Yanheng; Sun, Geng

    2018-04-01

    Community detection can be used as an important technique for product and personalized service recommendation. A game theory based approach to detect overlapping community structure is introduced in this paper. The process of the community formation is converted into a game, when all agents (nodes) cannot improve their own utility, the game process will be terminated. The utility function is composed of a gain and a loss function and we present a new gain function in this paper. In addition, different from choosing action randomly among join, quit and switch for each agent to get new label, two new strategies for each agent to update its label are designed during the game, and the strategies are also evaluated and compared for each agent in order to find its best result. The overlapping community structure is naturally presented when the stop criterion is satisfied. The experimental results demonstrate that the proposed algorithm outperforms other similar algorithms for detecting overlapping communities in networks.

  4. LP-LPA: A link influence-based label propagation algorithm for discovering community structures in networks

    Science.gov (United States)

    Berahmand, Kamal; Bouyer, Asgarali

    2018-03-01

    Community detection is an essential approach for analyzing the structural and functional properties of complex networks. Although many community detection algorithms have been recently presented, most of them are weak and limited in different ways. Label Propagation Algorithm (LPA) is a well-known and efficient community detection technique which is characterized by the merits of nearly-linear running time and easy implementation. However, LPA has some significant problems such as instability, randomness, and monster community detection. In this paper, an algorithm, namely node’s label influence policy for label propagation algorithm (LP-LPA) was proposed for detecting efficient community structures. LP-LPA measures link strength value for edges and nodes’ label influence value for nodes in a new label propagation strategy with preference on link strength and for initial nodes selection, avoid of random behavior in tiebreak states, and efficient updating order and rule update. These procedures can sort out the randomness issue in an original LPA and stabilize the discovered communities in all runs of the same network. Experiments on synthetic networks and a wide range of real-world social networks indicated that the proposed method achieves significant accuracy and high stability. Indeed, it can obviously solve monster community problem with regard to detecting communities in networks.

  5. Automatable algorithms to identify nonmedical opioid use using electronic data: a systematic review.

    Science.gov (United States)

    Canan, Chelsea; Polinski, Jennifer M; Alexander, G Caleb; Kowal, Mary K; Brennan, Troyen A; Shrank, William H

    2017-11-01

    Improved methods to identify nonmedical opioid use can help direct health care resources to individuals who need them. Automated algorithms that use large databases of electronic health care claims or records for surveillance are a potential means to achieve this goal. In this systematic review, we reviewed the utility, attempts at validation, and application of such algorithms to detect nonmedical opioid use. We searched PubMed and Embase for articles describing automatable algorithms that used electronic health care claims or records to identify patients or prescribers with likely nonmedical opioid use. We assessed algorithm development, validation, and performance characteristics and the settings where they were applied. Study variability precluded a meta-analysis. Of 15 included algorithms, 10 targeted patients, 2 targeted providers, 2 targeted both, and 1 identified medications with high abuse potential. Most patient-focused algorithms (67%) used prescription drug claims and/or medical claims, with diagnosis codes of substance abuse and/or dependence as the reference standard. Eleven algorithms were developed via regression modeling. Four used natural language processing, data mining, audit analysis, or factor analysis. Automated algorithms can facilitate population-level surveillance. However, there is no true gold standard for determining nonmedical opioid use. Users must recognize the implications of identifying false positives and, conversely, false negatives. Few algorithms have been applied in real-world settings. Automated algorithms may facilitate identification of patients and/or providers most likely to need more intensive screening and/or intervention for nonmedical opioid use. Additional implementation research in real-world settings would clarify their utility. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  6. Coding algorithms for identifying patients with cirrhosis and hepatitis B or C virus using administrative data.

    Science.gov (United States)

    Niu, Bolin; Forde, Kimberly A; Goldberg, David S

    2015-01-01

    Despite the use of administrative data to perform epidemiological and cost-effectiveness research on patients with hepatitis B or C virus (HBV, HCV), there are no data outside of the Veterans Health Administration validating whether International Classification of Disease, Ninth Revision, Clinical Modification (ICD-9-CM) codes can accurately identify cirrhotic patients with HBV or HCV. The validation of such algorithms is necessary for future epidemiological studies. We evaluated the positive predictive value (PPV) of ICD-9-CM codes for identifying chronic HBV or HCV among cirrhotic patients within the University of Pennsylvania Health System, a large network that includes a tertiary care referral center, a community-based hospital, and multiple outpatient practices across southeastern Pennsylvania and southern New Jersey. We reviewed a random sample of 200 cirrhotic patients with ICD-9-CM codes for HCV and 150 cirrhotic patients with ICD-9-CM codes for HBV. The PPV of 1 inpatient or 2 outpatient HCV codes was 88.0% (168/191, 95% CI: 82.5-92.2%), while the PPV of 1 inpatient or 2 outpatient HBV codes was 81.3% (113/139, 95% CI: 73.8-87.4%). Several variations of the primary coding algorithm were evaluated to determine if different combinations of inpatient and/or outpatient ICD-9-CM codes could increase the PPV of the coding algorithm. ICD-9-CM codes can identify chronic HBV or HCV in cirrhotic patients with a high PPV and can be used in future epidemiologic studies to examine disease burden and the proper allocation of resources. Copyright © 2014 John Wiley & Sons, Ltd.

  7. Searching for Communities in Bipartite Networks

    OpenAIRE

    Barber, Michael J.; Faria, Margarida; Streit, Ludwig; Strogan, Oleg

    2008-01-01

    Bipartite networks are a useful tool for representing and investigating interaction networks. We consider methods for identifying communities in bipartite networks. Intuitive notions of network community groups are made explicit using Newman's modularity measure. A specialized version of the modularity, adapted to be appropriate for bipartite networks, is presented; a corresponding algorithm is described for identifying community groups through maximizing this measure. The algorithm is applie...

  8. Optimal Design of a Hydrogen Community by Genetic Algorithms

    International Nuclear Information System (INIS)

    Rodolfo Dufo Lopez; Jose Luis Bernal Agustin; Luis Correas Uson; Ismael Aso Aguarta

    2006-01-01

    A study was conducted for the implementation of two Hydrogen Communities, following the recommendations of the HY-COM initiative of the European Commission. The proposed communities find their place in the municipality of Sabinanigo (Aragon, Spain). Two cases are analyzed, one off-grid village house near Sabinanigo, and a house situated in the town proper. The study was carried out with the HOGA program, Hybrid Optimization by Genetic Algorithms. A description is provided for the algorithms. The off-grid study deals with a hybrid pv-wind system with hydrogen storage for AC supply to an isolated house. The urban study is related to hydrogen production by means of hybrid renewable sources available locally (photovoltaic, wind and hydro). These complement the existing industrial electrolysis processes, in order to cater for the energy requirements of a small fleet of municipal hydrogen-powered vehicles. HOGA was used to optimize both hybrid systems. Dimensioning and deployment estimations are also provided. (authors)

  9. Optimal Design of a Hydrogen Community by Genetic Algorithms

    International Nuclear Information System (INIS)

    Rodolfo Dufo Lopeza; Jose Luis Bernal Agustin; Luis Correas Uson; Ismael Aso Aguarta

    2006-01-01

    A study was conducted for the implementation of two Hydrogen Communities, following the recommendations of the HY-COM initiative of the European Commission. The proposed communities find their place in the municipality of Sabinanigo (Aragon, Spain). Two cases are analyzed, one off-grid village house near Sabinanigo, and a house situated in the town proper. The study was carried out with the HOGA program, Hybrid Optimization by Genetic Algorithms. A description is provided for the algorithms. The off-grid study deals with a hybrid PV-wind system with hydrogen storage for AC supply to an isolated house. The urban study is related to hydrogen production by means of hybrid renewable sources available locally (photovoltaic, wind and hydro). These complement the existing industrial electrolysis processes, in order to cater for the energy requirements of a small fleet of municipal hydrogen-powered vehicles. HOGA was used to optimize both hybrid systems. Dimensioning and deployment estimations are also provided. (authors)

  10. An improved label propagation algorithm based on node importance and random walk for community detection

    Science.gov (United States)

    Ma, Tianren; Xia, Zhengyou

    2017-05-01

    Currently, with the rapid development of information technology, the electronic media for social communication is becoming more and more popular. Discovery of communities is a very effective way to understand the properties of complex networks. However, traditional community detection algorithms consider the structural characteristics of a social organization only, with more information about nodes and edges wasted. In the meanwhile, these algorithms do not consider each node on its merits. Label propagation algorithm (LPA) is a near linear time algorithm which aims to find the community in the network. It attracts many scholars owing to its high efficiency. In recent years, there are more improved algorithms that were put forward based on LPA. In this paper, an improved LPA based on random walk and node importance (NILPA) is proposed. Firstly, a list of node importance is obtained through calculation. The nodes in the network are sorted in descending order of importance. On the basis of random walk, a matrix is constructed to measure the similarity of nodes and it avoids the random choice in the LPA. Secondly, a new metric IAS (importance and similarity) is calculated by node importance and similarity matrix, which we can use to avoid the random selection in the original LPA and improve the algorithm stability. Finally, a test in real-world and synthetic networks is given. The result shows that this algorithm has better performance than existing methods in finding community structure.

  11. Community detection using preference networks

    Science.gov (United States)

    Tasgin, Mursel; Bingol, Haluk O.

    2018-04-01

    Community detection is the task of identifying clusters or groups of nodes in a network where nodes within the same group are more connected with each other than with nodes in different groups. It has practical uses in identifying similar functions or roles of nodes in many biological, social and computer networks. With the availability of very large networks in recent years, performance and scalability of community detection algorithms become crucial, i.e. if time complexity of an algorithm is high, it cannot run on large networks. In this paper, we propose a new community detection algorithm, which has a local approach and is able to run on large networks. It has a simple and effective method; given a network, algorithm constructs a preference network of nodes where each node has a single outgoing edge showing its preferred node to be in the same community with. In such a preference network, each connected component is a community. Selection of the preferred node is performed using similarity based metrics of nodes. We use two alternatives for this purpose which can be calculated in 1-neighborhood of nodes, i.e. number of common neighbors of selector node and its neighbors and, the spread capability of neighbors around the selector node which is calculated by the gossip algorithm of Lind et.al. Our algorithm is tested on both computer generated LFR networks and real-life networks with ground-truth community structure. It can identify communities accurately in a fast way. It is local, scalable and suitable for distributed execution on large networks.

  12. Identifying influential user communities on the social network

    Science.gov (United States)

    Hu, Weishu; Gong, Zhiguo; Hou U, Leong; Guo, Jingzhi

    2015-10-01

    Nowadays social network services have been popularly used in electronic commerce systems. Users on the social network can develop different relationships based on their common interests and activities. In order to promote the business, it is interesting to explore hidden relationships among users developed on the social network. Such knowledge can be used to locate target users for different advertisements and to provide effective product recommendations. In this paper, we define and study a novel community detection problem that is to discover the hidden community structure in large social networks based on their common interests. We observe that the users typically pay more attention to those users who share similar interests, which enable a way to partition the users into different communities according to their common interests. We propose two algorithms to detect influential communities using common interests in large social networks efficiently and effectively. We conduct our experimental evaluation using a data set from Epinions, which demonstrates that our method achieves 4-11.8% accuracy improvement over the state-of-the-art method.

  13. Detecting highly overlapping community structure by greedy clique expansion

    OpenAIRE

    Lee, Conrad; Reid, Fergal; McDaid, Aaron; Hurley, Neil

    2010-01-01

    In complex networks it is common for each node to belong to several communities, implying a highly overlapping community structure. Recent advances in benchmarking indicate that existing community assignment algorithms that are capable of detecting overlapping communities perform well only when the extent of community overlap is kept to modest levels. To overcome this limitation, we introduce a new community assignment algorithm called Greedy Clique Expansion (GCE). The algorithm identifies d...

  14. Positive predictive value of a register-based algorithm using the Danish National Registries to identify suicidal events.

    Science.gov (United States)

    Gasse, Christiane; Danielsen, Andreas Aalkjaer; Pedersen, Marianne Giørtz; Pedersen, Carsten Bøcker; Mors, Ole; Christensen, Jakob

    2018-04-17

    It is not possible to fully assess intention of self-harm and suicidal events using information from administrative databases. We conducted a validation study of intention of suicide attempts/self-harm contacts identified by a commonly applied Danish register-based algorithm (DK-algorithm) based on hospital discharge diagnosis and emergency room contacts. Of all 101 530 people identified with an incident suicide attempt/self-harm contact at Danish hospitals between 1995 and 2012 using the DK-algorithm, we selected a random sample of 475 people. We validated the DK-algorithm against medical records applying the definitions and terminology of the Columbia Classification Algorithm of Suicide Assessment of suicidal events, nonsuicidal events, and indeterminate or potentially suicidal events. We calculated positive predictive values (PPVs) of the DK-algorithm to identify suicidal events overall, by gender, age groups, and calendar time. We retrieved medical records for 357 (75%) people. The PPV of the DK-algorithm to identify suicidal events was 51.5% (95% CI: 46.4-56.7) overall, 42.7% (95% CI: 35.2-50.5) in males, and 58.5% (95% CI: 51.6-65.1) in females. The PPV varied further across age groups and calendar time. After excluding cases identified via the DK-algorithm by unspecific codes of intoxications and injury, the PPV improved slightly (56.8% [95% CI: 50.0-63.4]). The DK-algorithm can reliably identify self-harm with suicidal intention in 52% of the identified cases of suicide attempts/self-harm. The PPVs could be used for quantitative bias analysis and implemented as weights in future studies to estimate the proportion of suicidal events among cases identified via the DK-algorithm. Copyright © 2018 John Wiley & Sons, Ltd.

  15. Comparative evaluation of community detection algorithms: a topological approach

    International Nuclear Information System (INIS)

    Orman, Günce Keziban; Labatut, Vincent; Cherifi, Hocine

    2012-01-01

    Community detection is one of the most active fields in complex network analysis, due to its potential value in practical applications. Many works inspired by different paradigms are devoted to the development of algorithmic solutions allowing the network structure in such cohesive subgroups to be revealed. Comparative studies reported in the literature usually rely on a performance measure considering the community structure as a partition (Rand index, normalized mutual information, etc). However, this type of comparison neglects the topological properties of the communities. In this paper, we present a comprehensive comparative study of a representative set of community detection methods, in which we adopt both types of evaluation. Community-oriented topological measures are used to qualify the communities and evaluate their deviation from the reference structure. In order to mimic real-world systems, we use artificially generated realistic networks. It turns out there is no equivalence between the two approaches: a high performance does not necessarily correspond to correct topological properties, and vice versa. They can therefore be considered as complementary, and we recommend applying both of them in order to perform a complete and accurate assessment. (paper)

  16. Identifying Students’ Misconceptions on Basic Algorithmic Concepts Through Flowchart Analysis

    NARCIS (Netherlands)

    Rahimi, E.; Barendsen, E.; Henze, I.; Dagienė, V.; Hellas, A.

    2017-01-01

    In this paper, a flowchart-based approach to identifying secondary school students’ misconceptions (in a broad sense) on basic algorithm concepts is introduced. This approach uses student-generated flowcharts as the units of analysis and examines them against plan composition and construct-based

  17. Identifying multiple influential spreaders by a heuristic clustering algorithm

    International Nuclear Information System (INIS)

    Bao, Zhong-Kui; Liu, Jian-Guo; Zhang, Hai-Feng

    2017-01-01

    The problem of influence maximization in social networks has attracted much attention. However, traditional centrality indices are suitable for the case where a single spreader is chosen as the spreading source. Many times, spreading process is initiated by simultaneously choosing multiple nodes as the spreading sources. In this situation, choosing the top ranked nodes as multiple spreaders is not an optimal strategy, since the chosen nodes are not sufficiently scattered in networks. Therefore, one ideal situation for multiple spreaders case is that the spreaders themselves are not only influential but also they are dispersively distributed in networks, but it is difficult to meet the two conditions together. In this paper, we propose a heuristic clustering (HC) algorithm based on the similarity index to classify nodes into different clusters, and finally the center nodes in clusters are chosen as the multiple spreaders. HC algorithm not only ensures that the multiple spreaders are dispersively distributed in networks but also avoids the selected nodes to be very “negligible”. Compared with the traditional methods, our experimental results on synthetic and real networks indicate that the performance of HC method on influence maximization is more significant. - Highlights: • A heuristic clustering algorithm is proposed to identify the multiple influential spreaders in complex networks. • The algorithm can not only guarantee the selected spreaders are sufficiently scattered but also avoid to be “insignificant”. • The performance of our algorithm is generally better than other methods, regardless of real networks or synthetic networks.

  18. Identifying multiple influential spreaders by a heuristic clustering algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Bao, Zhong-Kui [School of Mathematical Science, Anhui University, Hefei 230601 (China); Liu, Jian-Guo [Data Science and Cloud Service Research Center, Shanghai University of Finance and Economics, Shanghai, 200133 (China); Zhang, Hai-Feng, E-mail: haifengzhang1978@gmail.com [School of Mathematical Science, Anhui University, Hefei 230601 (China); Department of Communication Engineering, North University of China, Taiyuan, Shan' xi 030051 (China)

    2017-03-18

    The problem of influence maximization in social networks has attracted much attention. However, traditional centrality indices are suitable for the case where a single spreader is chosen as the spreading source. Many times, spreading process is initiated by simultaneously choosing multiple nodes as the spreading sources. In this situation, choosing the top ranked nodes as multiple spreaders is not an optimal strategy, since the chosen nodes are not sufficiently scattered in networks. Therefore, one ideal situation for multiple spreaders case is that the spreaders themselves are not only influential but also they are dispersively distributed in networks, but it is difficult to meet the two conditions together. In this paper, we propose a heuristic clustering (HC) algorithm based on the similarity index to classify nodes into different clusters, and finally the center nodes in clusters are chosen as the multiple spreaders. HC algorithm not only ensures that the multiple spreaders are dispersively distributed in networks but also avoids the selected nodes to be very “negligible”. Compared with the traditional methods, our experimental results on synthetic and real networks indicate that the performance of HC method on influence maximization is more significant. - Highlights: • A heuristic clustering algorithm is proposed to identify the multiple influential spreaders in complex networks. • The algorithm can not only guarantee the selected spreaders are sufficiently scattered but also avoid to be “insignificant”. • The performance of our algorithm is generally better than other methods, regardless of real networks or synthetic networks.

  19. Identification of overlapping communities and their hierarchy by locally calculating community-changing resolution levels

    International Nuclear Information System (INIS)

    Havemann, Frank; Heinz, Michael; Struck, Alexander; Gläser, Jochen

    2011-01-01

    We propose a new local, deterministic and parameter-free algorithm that detects fuzzy and crisp overlapping communities in a weighted network and simultaneously reveals their hierarchy. Using a local fitness function, the algorithm greedily expands natural communities of seeds until the whole graph is covered. The hierarchy of communities is obtained analytically by calculating resolution levels at which communities grow rather than numerically by testing different resolution levels. This analytic procedure is not only more exact than its numerical alternatives such as LFM and GCE but also much faster. Critical resolution levels can be identified by searching for intervals in which large changes of the resolution do not lead to growth of communities. We tested our algorithm on benchmark graphs and on a network of 492 papers in information science. Combined with a specific post-processing, the algorithm gives much more precise results on LFR benchmarks with high overlap compared to other algorithms and performs very similarly to GCE

  20. Identification of overlapping communities and their hierarchy by locally calculating community-changing resolution levels

    Science.gov (United States)

    Havemann, Frank; Heinz, Michael; Struck, Alexander; Gläser, Jochen

    2011-01-01

    We propose a new local, deterministic and parameter-free algorithm that detects fuzzy and crisp overlapping communities in a weighted network and simultaneously reveals their hierarchy. Using a local fitness function, the algorithm greedily expands natural communities of seeds until the whole graph is covered. The hierarchy of communities is obtained analytically by calculating resolution levels at which communities grow rather than numerically by testing different resolution levels. This analytic procedure is not only more exact than its numerical alternatives such as LFM and GCE but also much faster. Critical resolution levels can be identified by searching for intervals in which large changes of the resolution do not lead to growth of communities. We tested our algorithm on benchmark graphs and on a network of 492 papers in information science. Combined with a specific post-processing, the algorithm gives much more precise results on LFR benchmarks with high overlap compared to other algorithms and performs very similarly to GCE.

  1. New algorithms for identifying the flavour of [Formula: see text] mesons using pions and protons.

    Science.gov (United States)

    Aaij, R; Adeva, B; Adinolfi, M; Ajaltouni, Z; Akar, S; Albrecht, J; Alessio, F; Alexander, M; Ali, S; Alkhazov, G; Alvarez Cartelle, P; Alves, A A; Amato, S; Amerio, S; Amhis, Y; An, L; Anderlini, L; Andreassi, G; Andreotti, M; Andrews, J E; Appleby, R B; Archilli, F; d'Argent, P; Arnau Romeu, J; Artamonov, A; Artuso, M; Aslanides, E; Auriemma, G; Baalouch, M; Babuschkin, I; Bachmann, S; Back, J J; Badalov, A; Baesso, C; Baker, S; Baldini, W; Barlow, R J; Barschel, C; Barsuk, S; Barter, W; Baszczyk, M; Batozskaya, V; Batsukh, B; Battista, V; Bay, A; Beaucourt, L; Beddow, J; Bedeschi, F; Bediaga, I; Bel, L J; Bellee, V; Belloli, N; Belous, K; Belyaev, I; Ben-Haim, E; Bencivenni, G; Benson, S; Benton, J; Berezhnoy, A; Bernet, R; Bertolin, A; Betti, F; Bettler, M-O; van Beuzekom, M; Bezshyiko, Ia; Bifani, S; Billoir, P; Bird, T; Birnkraut, A; Bitadze, A; Bizzeti, A; Blake, T; Blanc, F; Blouw, J; Blusk, S; Bocci, V; Boettcher, T; Bondar, A; Bondar, N; Bonivento, W; Bordyuzhin, I; Borgheresi, A; Borghi, S; Borisyak, M; Borsato, M; Bossu, F; Boubdir, M; Bowcock, T J V; Bowen, E; Bozzi, C; Braun, S; Britsch, M; Britton, T; Brodzicka, J; Buchanan, E; Burr, C; Bursche, A; Buytaert, J; Cadeddu, S; Calabrese, R; Calvi, M; Calvo Gomez, M; Camboni, A; Campana, P; Campora Perez, D; Campora Perez, D H; Capriotti, L; Carbone, A; Carboni, G; Cardinale, R; Cardini, A; Carniti, P; Carson, L; Carvalho Akiba, K; Casse, G; Cassina, L; Castillo Garcia, L; Cattaneo, M; Cauet, Ch; Cavallero, G; Cenci, R; Charles, M; Charpentier, Ph; Chatzikonstantinidis, G; Chefdeville, M; Chen, S; Cheung, S F; Chobanova, V; Chrzaszcz, M; Cid Vidal, X; Ciezarek, G; Clarke, P E L; Clemencic, M; Cliff, H V; Closier, J; Coco, V; Cogan, J; Cogneras, E; Cogoni, V; Cojocariu, L; Collins, P; Comerma-Montells, A; Contu, A; Cook, A; Coombs, G; Coquereau, S; Corti, G; Corvo, M; Costa Sobral, C M; Couturier, B; Cowan, G A; Craik, D C; Crocombe, A; Cruz Torres, M; Cunliffe, S; Currie, R; D'Ambrosio, C; Da Cunha Marinho, F; Dall'Occo, E; Dalseno, J; David, P N Y; Davis, A; De Aguiar Francisco, O; De Bruyn, K; De Capua, S; De Cian, M; De Miranda, J M; De Paula, L; De Serio, M; De Simone, P; Dean, C T; Decamp, D; Deckenhoff, M; Del Buono, L; Demmer, M; Dendek, A; Derkach, D; Deschamps, O; Dettori, F; Dey, B; Di Canto, A; Dijkstra, H; Dordei, F; Dorigo, M; Dosil Suárez, A; Dovbnya, A; Dreimanis, K; Dufour, L; Dujany, G; Dungs, K; Durante, P; Dzhelyadin, R; Dziurda, A; Dzyuba, A; Déléage, N; Easo, S; Ebert, M; Egede, U; Egorychev, V; Eidelman, S; Eisenhardt, S; Eitschberger, U; Ekelhof, R; Eklund, L; Elsasser, Ch; Ely, S; Esen, S; Evans, H M; Evans, T; Falabella, A; Farley, N; Farry, S; Fay, R; Fazzini, D; Ferguson, D; Fernandez Prieto, A; Ferrari, F; Ferreira Rodrigues, F; Ferro-Luzzi, M; Filippov, S; Fini, R A; Fiore, M; Fiorini, M; Firlej, M; Fitzpatrick, C; Fiutowski, T; Fleuret, F; Fohl, K; Fontana, M; Fontanelli, F; Forshaw, D C; Forty, R; Franco Lima, V; Frank, M; Frei, C; Fu, J; Furfaro, E; Färber, C; Gallas Torreira, A; Galli, D; Gallorini, S; Gambetta, S; Gandelman, M; Gandini, P; Gao, Y; Garcia Martin, L M; García Pardiñas, J; Garra Tico, J; Garrido, L; Garsed, P J; Gascon, D; Gaspar, C; Gavardi, L; Gazzoni, G; Gerick, D; Gersabeck, E; Gersabeck, M; Gershon, T; Ghez, Ph; Gianì, S; Gibson, V; Girard, O G; Giubega, L; Gizdov, K; Gligorov, V V; Golubkov, D; Golutvin, A; Gomes, A; Gorelov, I V; Gotti, C; Grabalosa Gándara, M; Graciani Diaz, R; Granado Cardoso, L A; Graugés, E; Graverini, E; Graziani, G; Grecu, A; Griffith, P; Grillo, L; Gruberg Cazon, B R; Grünberg, O; Gushchin, E; Guz, Yu; Gys, T; Göbel, C; Hadavizadeh, T; Hadjivasiliou, C; Haefeli, G; Haen, C; Haines, S C; Hall, S; Hamilton, B; Han, X; Hansmann-Menzemer, S; Harnew, N; Harnew, S T; Harrison, J; Hatch, M; He, J; Head, T; Heister, A; Hennessy, K; Henrard, P; Henry, L; Hernando Morata, J A; van Herwijnen, E; Heß, M; Hicheur, A; Hill, D; Hombach, C; Hopchev, P H; Hulsbergen, W; Humair, T; Hushchyn, M; Hussain, N; Hutchcroft, D; Idzik, M; Ilten, P; Jacobsson, R; Jaeger, A; Jalocha, J; Jans, E; Jawahery, A; Jiang, F; John, M; Johnson, D; Jones, C R; Joram, C; Jost, B; Jurik, N; Kandybei, S; Kanso, W; Karacson, M; Kariuki, J M; Karodia, S; Kecke, M; Kelsey, M; Kenyon, I R; Kenzie, M; Ketel, T; Khairullin, E; Khanji, B; Khurewathanakul, C; Kirn, T; Klaver, S; Klimaszewski, K; Koliiev, S; Kolpin, M; Komarov, I; Koopman, R F; Koppenburg, P; Kosmyntseva, A; Kozeiha, M; Kravchuk, L; Kreplin, K; Kreps, M; Krokovny, P; Kruse, F; Krzemien, W; Kucewicz, W; Kucharczyk, M; Kudryavtsev, V; Kuonen, A K; Kurek, K; Kvaratskheliya, T; Lacarrere, D; Lafferty, G; Lai, A; Lambert, D; Lanfranchi, G; Langenbruch, C; Latham, T; Lazzeroni, C; Le Gac, R; van Leerdam, J; Lees, J-P; Leflat, A; Lefrançois, J; Lefèvre, R; Lemaitre, F; Lemos Cid, E; Leroy, O; Lesiak, T; Leverington, B; Li, Y; Likhomanenko, T; Lindner, R; Linn, C; Lionetto, F; Liu, B; Liu, X; Loh, D; Longstaff, I; Lopes, J H; Lucchesi, D; Lucio Martinez, M; Luo, H; Lupato, A; Luppi, E; Lupton, O; Lusiani, A; Lyu, X; Machefert, F; Maciuc, F; Maev, O; Maguire, K; Malde, S; Malinin, A; Maltsev, T; Manca, G; Mancinelli, G; Manning, P; Maratas, J; Marchand, J F; Marconi, U; Marin Benito, C; Marino, P; Marks, J; Martellotti, G; Martin, M; Martinelli, M; Martinez Santos, D; Martinez Vidal, F; Martins Tostes, D; Massacrier, L M; Massafferri, A; Matev, R; Mathad, A; Mathe, Z; Matteuzzi, C; Mauri, A; Maurin, B; Mazurov, A; McCann, M; McCarthy, J; McNab, A; McNulty, R; Meadows, B; Meier, F; Meissner, M; Melnychuk, D; Merk, M; Merli, A; Michielin, E; Milanes, D A; Minard, M-N; Mitzel, D S; Mogini, A; Molina Rodriguez, J; Monroy, I A; Monteil, S; Morandin, M; Morawski, P; Mordà, A; Morello, M J; Moron, J; Morris, A B; Mountain, R; Muheim, F; Mulder, M; Mussini, M; Müller, D; Müller, J; Müller, K; Müller, V; Naik, P; Nakada, T; Nandakumar, R; Nandi, A; Nasteva, I; Needham, M; Neri, N; Neubert, S; Neufeld, N; Neuner, M; Nguyen, A D; Nguyen, T D; Nguyen-Mau, C; Nieswand, S; Niet, R; Nikitin, N; Nikodem, T; Novoselov, A; O'Hanlon, D P; Oblakowska-Mucha, A; Obraztsov, V; Ogilvy, S; Oldeman, R; Onderwater, C J G; Otalora Goicochea, J M; Otto, A; Owen, P; Oyanguren, A; Pais, P R; Palano, A; Palombo, F; Palutan, M; Panman, J; Papanestis, A; Pappagallo, M; Pappalardo, L L; Parker, W; Parkes, C; Passaleva, G; Pastore, A; Patel, G D; Patel, M; Patrignani, C; Pearce, A; Pellegrino, A; Penso, G; Pepe Altarelli, M; Perazzini, S; Perret, P; Pescatore, L; Petridis, K; Petrolini, A; Petrov, A; Petruzzo, M; Picatoste Olloqui, E; Pietrzyk, B; Pikies, M; Pinci, D; Pistone, A; Piucci, A; Playfer, S; Plo Casasus, M; Poikela, T; Polci, F; Poluektov, A; Polyakov, I; Polycarpo, E; Pomery, G J; Popov, A; Popov, D; Popovici, B; Poslavskii, S; Potterat, C; Price, E; Price, J D; Prisciandaro, J; Pritchard, A; Prouve, C; Pugatch, V; Puig Navarro, A; Punzi, G; Qian, W; Quagliani, R; Rachwal, B; Rademacker, J H; Rama, M; Ramos Pernas, M; Rangel, M S; Raniuk, I; Ratnikov, F; Raven, G; Redi, F; Reichert, S; Dos Reis, A C; Remon Alepuz, C; Renaudin, V; Ricciardi, S; Richards, S; Rihl, M; Rinnert, K; Rives Molina, V; Robbe, P; Rodrigues, A B; Rodrigues, E; Rodriguez Lopez, J A; Rodriguez Perez, P; Rogozhnikov, A; Roiser, S; Rollings, A; Romanovskiy, V; Romero Vidal, A; Ronayne, J W; Rotondo, M; Rudolph, M S; Ruf, T; Ruiz Valls, P; Saborido Silva, J J; Sadykhov, E; Sagidova, N; Saitta, B; Salustino Guimaraes, V; Sanchez Mayordomo, C; Sanmartin Sedes, B; Santacesaria, R; Santamarina Rios, C; Santimaria, M; Santovetti, E; Sarti, A; Satriano, C; Satta, A; Saunders, D M; Savrina, D; Schael, S; Schellenberg, M; Schiller, M; Schindler, H; Schlupp, M; Schmelling, M; Schmelzer, T; Schmidt, B; Schneider, O; Schopper, A; Schubert, K; Schubiger, M; Schune, M-H; Schwemmer, R; Sciascia, B; Sciubba, A; Semennikov, A; Sergi, A; Serra, N; Serrano, J; Sestini, L; Seyfert, P; Shapkin, M; Shapoval, I; Shcheglov, Y; Shears, T; Shekhtman, L; Shevchenko, V; Shires, A; Siddi, B G; Silva Coutinho, R; Silva de Oliveira, L; Simi, G; Simone, S; Sirendi, M; Skidmore, N; Skwarnicki, T; Smith, E; Smith, I T; Smith, J; Smith, M; Snoek, H; Sokoloff, M D; Soler, F J P; Souza De Paula, B; Spaan, B; Spradlin, P; Sridharan, S; Stagni, F; Stahl, M; Stahl, S; Stefko, P; Stefkova, S; Steinkamp, O; Stemmle, S; Stenyakin, O; Stevenson, S; Stoica, S; Stone, S; Storaci, B; Stracka, S; Straticiuc, M; Straumann, U; Sun, L; Sutcliffe, W; Swientek, K; Syropoulos, V; Szczekowski, M; Szumlak, T; T'Jampens, S; Tayduganov, A; Tekampe, T; Teklishyn, M; Tellarini, G; Teubert, F; Thomas, E; van Tilburg, J; Tilley, M J; Tisserand, V; Tobin, M; Tolk, S; Tomassetti, L; Tonelli, D; Topp-Joergensen, S; Toriello, F; Tournefier, E; Tourneur, S; Trabelsi, K; Traill, M; Tran, M T; Tresch, M; Trisovic, A; Tsaregorodtsev, A; Tsopelas, P; Tully, A; Tuning, N; Ukleja, A; Ustyuzhanin, A; Uwer, U; Vacca, C; Vagnoni, V; Valassi, A; Valat, S; Valenti, G; Vallier, A; Vazquez Gomez, R; Vazquez Regueiro, P; Vecchi, S; van Veghel, M; Velthuis, J J; Veltri, M; Veneziano, G; Venkateswaran, A; Vernet, M; Vesterinen, M; Viaud, B; Vieira, D; Vieites Diaz, M; Vilasis-Cardona, X; Volkov, V; Vollhardt, A; Voneki, B; Vorobyev, A; Vorobyev, V; Voß, C; de Vries, J A; Vázquez Sierra, C; Waldi, R; Wallace, C; Wallace, R; Walsh, J; Wang, J; Ward, D R; Wark, H M; Watson, N K; Websdale, D; Weiden, A; Whitehead, M; Wicht, J; Wilkinson, G; Wilkinson, M; Williams, M; Williams, M P; Williams, M; Williams, T; Wilson, F F; Wimberley, J; Wishahi, J; Wislicki, W; Witek, M; Wormser, G; Wotton, S A; Wraight, K; Wyllie, K; Xie, Y; Xu, Z; Yang, Z; Yin, H; Yu, J; Yuan, X; Yushchenko, O; Zarebski, K A; Zavertyaev, M; Zhang, L; Zhang, Y; Zhelezov, A; Zheng, Y; Zhokhov, A; Zhu, X; Zhukov, V; Zucchelli, S

    2017-01-01

    Two new algorithms for use in the analysis of [Formula: see text] collision are developed to identify the flavour of [Formula: see text] mesons at production using pions and protons from the hadronization process. The algorithms are optimized and calibrated on data, using [Formula: see text] decays from [Formula: see text] collision data collected by LHCb at centre-of-mass energies of 7 and 8 TeV . The tagging power of the new pion algorithm is 60% greater than the previously available one; the algorithm using protons to identify the flavour of a [Formula: see text] meson is the first of its kind.

  2. Research on Community Structure in Bus Transport Networks

    International Nuclear Information System (INIS)

    Yang Xuhua; Wang Bo; Sun Youxian

    2009-01-01

    We abstract the bus transport networks (BTNs) to two kinds of complex networks with space L and space P methods respectively. Using improved community detecting algorithm (PKM agglomerative algorithm), we analyze the community property of two kinds of BTNs graphs. The results show that the BTNs graph described with space L method have obvious community property, but the other kind of BTNs graph described with space P method have not. The reason is that the BTNs graph described with space P method have the intense overlapping community property and general community division algorithms can not identify this kind of community structure. To overcome this problem, we propose a novel community structure called N-depth community and present a corresponding community detecting algorithm, which can detect overlapping community. Applying the novel community structure and detecting algorithm to a BTN evolution model described with space P, whose network property agrees well with real BTNs', we get obvious community property. (general)

  3. Identifying Septal Support Reconstructions for Saddle Nose Deformity: The Cakmak Algorithm.

    Science.gov (United States)

    Cakmak, Ozcan; Emre, Ismet Emrah; Ozkurt, Fazil Emre

    2015-01-01

    The saddle nose deformity is one of the most challenging problems in nasal surgery with a less predictable and reproducible result than other nasal procedures. The main feature of this deformity is loss of septal support with both functional and aesthetic implications. Most reports on saddle nose have focused on aesthetic improvement and neglected the reestablishment of septal support to improve airway. To explain how the Cakmak algorithm, an algorithm that describes various fixation techniques and grafts in different types of saddle nose deformities, aids in identifying saddle nose reconstructions that restore supportive nasal framework and provide the aesthetic improvements typically associated with procedures to correct saddle nose deformities. This algorithm presents septal support reconstruction of patients with saddle nose deformity based on the experience of the senior author in 206 patients with saddle nose deformity. Preoperative examination, intraoperative assessment, reconstruction techniques, graft materials, and patient evaluation of aesthetic success were documented, and 4 different types of saddle nose deformities were defined. The Cakmak algorithm classifies varying degrees of saddle nose deformity from type 0 to type 4 and helps identify the most appropriate surgical procedure to restore the supportive nasal framework and aesthetic dorsum. Among the 206 patients, 110 women and 96 men, mean (range) age was 39.7 years (15-68 years), and mean (range) of follow-up was 32 months (6-148 months). All but 12 patients had a history of previous nasal surgeries. Application of the Cakmak algorithm resulted in 36 patients categorized with type 0 saddle nose deformities; 79, type 1; 50, type 2; 20, type 3a; 7, type 3b; and 14, type 4. Postoperative photographs showed improvement of deformities, and patient surveys revealed aesthetic improvement in 201 patients and improvement in nasal breathing in 195 patients. Three patients developed postoperative infection

  4. A similarity based agglomerative clustering algorithm in networks

    Science.gov (United States)

    Liu, Zhiyuan; Wang, Xiujuan; Ma, Yinghong

    2018-04-01

    The detection of clusters is benefit for understanding the organizations and functions of networks. Clusters, or communities, are usually groups of nodes densely interconnected but sparsely linked with any other clusters. To identify communities, an efficient and effective community agglomerative algorithm based on node similarity is proposed. The proposed method initially calculates similarities between each pair of nodes, and form pre-partitions according to the principle that each node is in the same community as its most similar neighbor. After that, check each partition whether it satisfies community criterion. For the pre-partitions who do not satisfy, incorporate them with others that having the biggest attraction until there are no changes. To measure the attraction ability of a partition, we propose an attraction index that based on the linked node's importance in networks. Therefore, our proposed method can better exploit the nodes' properties and network's structure. To test the performance of our algorithm, both synthetic and empirical networks ranging in different scales are tested. Simulation results show that the proposed algorithm can obtain superior clustering results compared with six other widely used community detection algorithms.

  5. Global identifiability of linear compartmental models--a computer algebra algorithm.

    Science.gov (United States)

    Audoly, S; D'Angiò, L; Saccomani, M P; Cobelli, C

    1998-01-01

    A priori global identifiability deals with the uniqueness of the solution for the unknown parameters of a model and is, thus, a prerequisite for parameter estimation of biological dynamic models. Global identifiability is however difficult to test, since it requires solving a system of algebraic nonlinear equations which increases both in nonlinearity degree and number of terms and unknowns with increasing model order. In this paper, a computer algebra tool, GLOBI (GLOBal Identifiability) is presented, which combines the topological transfer function method with the Buchberger algorithm, to test global identifiability of linear compartmental models. GLOBI allows for the automatic testing of a priori global identifiability of general structure compartmental models from general multi input-multi output experiments. Examples of usage of GLOBI to analyze a priori global identifiability of some complex biological compartmental models are provided.

  6. Improving the recommender algorithms with the detected communities in bipartite networks

    Science.gov (United States)

    Zhang, Peng; Wang, Duo; Xiao, Jinghua

    2017-04-01

    Recommender system offers a powerful tool to make information overload problem well solved and thus gains wide concerns of scholars and engineers. A key challenge is how to make recommendations more accurate and personalized. We notice that community structures widely exist in many real networks, which could significantly affect the recommendation results. By incorporating the information of detected communities in the recommendation algorithms, an improved recommendation approach for the networks with communities is proposed. The approach is examined in both artificial and real networks, the results show that the improvement on accuracy and diversity can be 20% and 7%, respectively. This reveals that it is beneficial to classify the nodes based on the inherent properties in recommender systems.

  7. The development of gamma energy identify algorithm for compact radiation sensors using stepwise refinement technique

    Energy Technology Data Exchange (ETDEWEB)

    Yoo, Hyun Jun [Div. of Radiation Regulation, Korea Institute of Nuclear Safety, Daejeon (Korea, Republic of); Kim, Ye Won; Kim, Hyun Duk; Cho, Gyu Seong [Dept. of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon (Korea, Republic of); Yi, Yun [Dept. of of Electronics and Information Engineering, Korea University, Seoul (Korea, Republic of)

    2017-06-15

    A gamma energy identifying algorithm using spectral decomposition combined with smoothing method was suggested to confirm the existence of the artificial radio isotopes. The algorithm is composed by original pattern recognition method and smoothing method to enhance the performance to identify gamma energy of radiation sensors that have low energy resolution. The gamma energy identifying algorithm for the compact radiation sensor is a three-step of refinement process. Firstly, the magnitude set is calculated by the original spectral decomposition. Secondly, the magnitude of modeling error in the magnitude set is reduced by the smoothing method. Thirdly, the expected gamma energy is finally decided based on the enhanced magnitude set as a result of the spectral decomposition with the smoothing method. The algorithm was optimized for the designed radiation sensor composed of a CsI (Tl) scintillator and a silicon pin diode. The two performance parameters used to estimate the algorithm are the accuracy of expected gamma energy and the number of repeated calculations. The original gamma energy was accurately identified with the single energy of gamma radiation by adapting this modeling error reduction method. Also the average error decreased by half with the multi energies of gamma radiation in comparison to the original spectral decomposition. In addition, the number of repeated calculations also decreased by half even in low fluence conditions under 104 (/0.09 cm{sup 2} of the scintillator surface). Through the development of this algorithm, we have confirmed the possibility of developing a product that can identify artificial radionuclides nearby using inexpensive radiation sensors that are easy to use by the public. Therefore, it can contribute to reduce the anxiety of the public exposure by determining the presence of artificial radionuclides in the vicinity.

  8. An Improved Topology-Potential-Based Community Detection Algorithm for Complex Network

    Directory of Open Access Journals (Sweden)

    Zhixiao Wang

    2014-01-01

    Full Text Available Topology potential theory is a new community detection theory on complex network, which divides a network into communities by spreading outward from each local maximum potential node. At present, almost all topology-potential-based community detection methods ignore node difference and assume that all nodes have the same mass. This hypothesis leads to inaccuracy of topology potential calculation and then decreases the precision of community detection. Inspired by the idea of PageRank algorithm, this paper puts forward a novel mass calculation method for complex network nodes. A node’s mass obtained by our method can effectively reflect its importance and influence in complex network. The more important the node is, the bigger its mass is. Simulation experiment results showed that, after taking node mass into consideration, the topology potential of node is more accurate, the distribution of topology potential is more reasonable, and the results of community detection are more precise.

  9. Validity of administrative database code algorithms to identify vascular access placement, surgical revisions, and secondary patency.

    Science.gov (United States)

    Al-Jaishi, Ahmed A; Moist, Louise M; Oliver, Matthew J; Nash, Danielle M; Fleet, Jamie L; Garg, Amit X; Lok, Charmaine E

    2018-03-01

    We assessed the validity of physician billing codes and hospital admission using International Classification of Diseases 10th revision codes to identify vascular access placement, secondary patency, and surgical revisions in administrative data. We included adults (≥18 years) with a vascular access placed between 1 April 2004 and 31 March 2013 at the University Health Network, Toronto. Our reference standard was a prospective vascular access database (VASPRO) that contains information on vascular access type and dates of placement, dates for failure, and any revisions. We used VASPRO to assess the validity of different administrative coding algorithms by calculating the sensitivity, specificity, and positive predictive values of vascular access events. The sensitivity (95% confidence interval) of the best performing algorithm to identify arteriovenous access placement was 86% (83%, 89%) and specificity was 92% (89%, 93%). The corresponding numbers to identify catheter insertion were 84% (82%, 86%) and 84% (80%, 87%), respectively. The sensitivity of the best performing coding algorithm to identify arteriovenous access surgical revisions was 81% (67%, 90%) and specificity was 89% (87%, 90%). The algorithm capturing arteriovenous access placement and catheter insertion had a positive predictive value greater than 90% and arteriovenous access surgical revisions had a positive predictive value of 20%. The duration of arteriovenous access secondary patency was on average 578 (553, 603) days in VASPRO and 555 (530, 580) days in administrative databases. Administrative data algorithms have fair to good operating characteristics to identify vascular access placement and arteriovenous access secondary patency. Low positive predictive values for surgical revisions algorithm suggest that administrative data should only be used to rule out the occurrence of an event.

  10. The efficiency of average linkage hierarchical clustering algorithm associated multi-scale bootstrap resampling in identifying homogeneous precipitation catchments

    Science.gov (United States)

    Chuan, Zun Liang; Ismail, Noriszura; Shinyie, Wendy Ling; Lit Ken, Tan; Fam, Soo-Fen; Senawi, Azlyna; Yusoff, Wan Nur Syahidah Wan

    2018-04-01

    Due to the limited of historical precipitation records, agglomerative hierarchical clustering algorithms widely used to extrapolate information from gauged to ungauged precipitation catchments in yielding a more reliable projection of extreme hydro-meteorological events such as extreme precipitation events. However, identifying the optimum number of homogeneous precipitation catchments accurately based on the dendrogram resulted using agglomerative hierarchical algorithms are very subjective. The main objective of this study is to propose an efficient regionalized algorithm to identify the homogeneous precipitation catchments for non-stationary precipitation time series. The homogeneous precipitation catchments are identified using average linkage hierarchical clustering algorithm associated multi-scale bootstrap resampling, while uncentered correlation coefficient as the similarity measure. The regionalized homogeneous precipitation is consolidated using K-sample Anderson Darling non-parametric test. The analysis result shows the proposed regionalized algorithm performed more better compared to the proposed agglomerative hierarchical clustering algorithm in previous studies.

  11. Effect of size heterogeneity on community identification in complex networks

    Energy Technology Data Exchange (ETDEWEB)

    Danon, L.; Diaz-Guilera, A.; Arenas, A.

    2008-01-01

    Identifying community structure can be a potent tool in the analysis and understanding of the structure of complex networks. Up to now, methods for evaluating the performance of identification algorithms use ad-hoc networks with communities of equal size. We show that inhomogeneities in community sizes can and do affect the performance of algorithms considerably, and propose an alternative method which takes these factors into account. Furthermore, we propose a simple modification of the algorithm proposed by Newman for community detection (Phys. Rev. E 69 066133) which treats communities of different sizes on an equal footing, and show that it outperforms the original algorithm while retaining its speed.

  12. Positive predictive value of a register-based algorithm using the Danish National Registries to identify suicidal events

    DEFF Research Database (Denmark)

    Gasse, Christiane; Danielsen, Andreas Aalkjaer; Pedersen, Marianne Giørtz

    2018-01-01

    events overall, by gender, age groups, and calendar time. RESULTS: We retrieved medical records for 357 (75%) people. The PPV of the DK-algorithm to identify suicidal events was 51.5% (95% CI: 46.4-56.7) overall, 42.7% (95% CI: 35.2-50.5) in males, and 58.5% (95% CI: 51.6-65.1) in females. The PPV varied...... further across age groups and calendar time. After excluding cases identified via the DK-algorithm by unspecific codes of intoxications and injury, the PPV improved slightly (56.8% [95% CI: 50.0-63.4]). CONCLUSIONS: The DK-algorithm can reliably identify self-harm with suicidal intention in 52......PURPOSE: It is not possible to fully assess intention of self-harm and suicidal events using information from administrative databases. We conducted a validation study of intention of suicide attempts/self-harm contacts identified by a commonly applied Danish register-based algorithm (DK...

  13. Identifying and Analyzing Novel Epilepsy-Related Genes Using Random Walk with Restart Algorithm

    Directory of Open Access Journals (Sweden)

    Wei Guo

    2017-01-01

    Full Text Available As a pathological condition, epilepsy is caused by abnormal neuronal discharge in brain which will temporarily disrupt the cerebral functions. Epilepsy is a chronic disease which occurs in all ages and would seriously affect patients’ personal lives. Thus, it is highly required to develop effective medicines or instruments to treat the disease. Identifying epilepsy-related genes is essential in order to understand and treat the disease because the corresponding proteins encoded by the epilepsy-related genes are candidates of the potential drug targets. In this study, a pioneering computational workflow was proposed to predict novel epilepsy-related genes using the random walk with restart (RWR algorithm. As reported in the literature RWR algorithm often produces a number of false positive genes, and in this study a permutation test and functional association tests were implemented to filter the genes identified by RWR algorithm, which greatly reduce the number of suspected genes and result in only thirty-three novel epilepsy genes. Finally, these novel genes were analyzed based upon some recently published literatures. Our findings implicate that all novel genes were closely related to epilepsy. It is believed that the proposed workflow can also be applied to identify genes related to other diseases and deepen our understanding of the mechanisms of these diseases.

  14. An administrative data validation study of the accuracy of algorithms for identifying rheumatoid arthritis: the influence of the reference standard on algorithm performance.

    Science.gov (United States)

    Widdifield, Jessica; Bombardier, Claire; Bernatsky, Sasha; Paterson, J Michael; Green, Diane; Young, Jacqueline; Ivers, Noah; Butt, Debra A; Jaakkimainen, R Liisa; Thorne, J Carter; Tu, Karen

    2014-06-23

    We have previously validated administrative data algorithms to identify patients with rheumatoid arthritis (RA) using rheumatology clinic records as the reference standard. Here we reassessed the accuracy of the algorithms using primary care records as the reference standard. We performed a retrospective chart abstraction study using a random sample of 7500 adult patients under the care of 83 family physicians contributing to the Electronic Medical Record Administrative data Linked Database (EMRALD) in Ontario, Canada. Using physician-reported diagnoses as the reference standard, we computed and compared the sensitivity, specificity, and predictive values for over 100 administrative data algorithms for RA case ascertainment. We identified 69 patients with RA for a lifetime RA prevalence of 0.9%. All algorithms had excellent specificity (>97%). However, sensitivity varied (75-90%) among physician billing algorithms. Despite the low prevalence of RA, most algorithms had adequate positive predictive value (PPV; 51-83%). The algorithm of "[1 hospitalization RA diagnosis code] or [3 physician RA diagnosis codes with ≥1 by a specialist over 2 years]" had a sensitivity of 78% (95% CI 69-88), specificity of 100% (95% CI 100-100), PPV of 78% (95% CI 69-88) and NPV of 100% (95% CI 100-100). Administrative data algorithms for detecting RA patients achieved a high degree of accuracy amongst the general population. However, results varied slightly from our previous report, which can be attributed to differences in the reference standards with respect to disease prevalence, spectrum of disease, and type of comparator group.

  15. Evaluation of algorithms to identify incident cancer cases by using French health administrative databases.

    Science.gov (United States)

    Ajrouche, Aya; Estellat, Candice; De Rycke, Yann; Tubach, Florence

    2017-08-01

    Administrative databases are increasingly being used in cancer observational studies. Identifying incident cancer in these databases is crucial. This study aimed to develop algorithms to estimate cancer incidence by using health administrative databases and to examine the accuracy of the algorithms in terms of national cancer incidence rates estimated from registries. We identified a cohort of 463 033 participants on 1 January 2012 in the Echantillon Généraliste des Bénéficiaires (EGB; a representative sample of the French healthcare insurance system). The EGB contains data on long-term chronic disease (LTD) status, reimbursed outpatient treatments and procedures, and hospitalizations (including discharge diagnoses, and costly medical procedures and drugs). After excluding cases of prevalent cancer, we applied 15 algorithms to estimate the cancer incidence rates separately for men and women in 2012 and compared them to the national cancer incidence rates estimated from French registries by indirect age and sex standardization. The most accurate algorithm for men combined information from LTD status, outpatient anticancer drugs, radiotherapy sessions and primary or related discharge diagnosis of cancer, although it underestimated the cancer incidence (standardized incidence ratio (SIR) 0.85 [0.80-0.90]). For women, the best algorithm used the same definition of the algorithm for men but restricted hospital discharge to only primary or related diagnosis with an additional inpatient procedure or drug reimbursement related to cancer and gave comparable estimates to those from registries (SIR 1.00 [0.94-1.06]). The algorithms proposed could be used for cancer incidence monitoring and for future etiological cancer studies involving French healthcare databases. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.

  16. Derivation and validation of the Personal Support Algorithm: an evidence-based framework to inform allocation of personal support services in home and community care.

    Science.gov (United States)

    Sinn, Chi-Ling Joanna; Jones, Aaron; McMullan, Janet Legge; Ackerman, Nancy; Curtin-Telegdi, Nancy; Eckel, Leslie; Hirdes, John P

    2017-11-25

    Personal support services enable many individuals to stay in their homes, but there are no standard ways to classify need for functional support in home and community care settings. The goal of this project was to develop an evidence-based clinical tool to inform service planning while allowing for flexibility in care coordinator judgment in response to patient and family circumstances. The sample included 128,169 Ontario home care patients assessed in 2013 and 25,800 Ontario community support clients assessed between 2014 and 2016. Independent variables were drawn from the Resident Assessment Instrument-Home Care and interRAI Community Health Assessment that are standardised, comprehensive, and fully compatible clinical assessments. Clinical expertise and regression analyses identified candidate variables that were entered into decision tree models. The primary dependent variable was the weekly hours of personal support calculated based on the record of billed services. The Personal Support Algorithm classified need for personal support into six groups with a 32-fold difference in average billed hours of personal support services between the highest and lowest group. The algorithm explained 30.8% of the variability in billed personal support services. Care coordinators and managers reported that the guidelines based on the algorithm classification were consistent with their clinical judgment and current practice. The Personal Support Algorithm provides a structured yet flexible decision-support framework that may facilitate a more transparent and equitable approach to the allocation of personal support services.

  17. Derivation and validation of the Personal Support Algorithm: an evidence-based framework to inform allocation of personal support services in home and community care

    Directory of Open Access Journals (Sweden)

    Chi-Ling Joanna Sinn

    2017-11-01

    Full Text Available Abstract Background Personal support services enable many individuals to stay in their homes, but there are no standard ways to classify need for functional support in home and community care settings. The goal of this project was to develop an evidence-based clinical tool to inform service planning while allowing for flexibility in care coordinator judgment in response to patient and family circumstances. Methods The sample included 128,169 Ontario home care patients assessed in 2013 and 25,800 Ontario community support clients assessed between 2014 and 2016. Independent variables were drawn from the Resident Assessment Instrument-Home Care and interRAI Community Health Assessment that are standardised, comprehensive, and fully compatible clinical assessments. Clinical expertise and regression analyses identified candidate variables that were entered into decision tree models. The primary dependent variable was the weekly hours of personal support calculated based on the record of billed services. Results The Personal Support Algorithm classified need for personal support into six groups with a 32-fold difference in average billed hours of personal support services between the highest and lowest group. The algorithm explained 30.8% of the variability in billed personal support services. Care coordinators and managers reported that the guidelines based on the algorithm classification were consistent with their clinical judgment and current practice. Conclusions The Personal Support Algorithm provides a structured yet flexible decision-support framework that may facilitate a more transparent and equitable approach to the allocation of personal support services.

  18. Identifying Priorities for Aging Policies in Two Portuguese Communities.

    Science.gov (United States)

    Bárrios, Maria João; Fernandes, Ana Alexandre; Fonseca, António Manuel

    2018-02-20

    The aging population has an impact on health, social, and economic issues in regard to individuals, communities, and organizations. The challenge for local policies in response to aging is to create sufficient resources to meet the population's needs, wishes, and rights as people age. Active aging constitutes one of the guiding perspectives on policies. Taking into account the local governance perspective, the Model for Aging Local Policies Analysis (MALPA) was created in order to convert the active aging paradigm into a practical approach, as a technique to evaluate and analyze local aging policies. In this research, the MALPA instrument was applied in two Portuguese communities (Coruche and Oeiras). The objective was to identify the intervention priorities of aging policies in both communities, determining whether the instrument can facilitate the development of proposals for the improvement of local aging policies. It was possible to evaluate the communities and programs, identifying the least appropriate policy actions regarding the intervention priorities. The results allowed us to identify 10 priorities about collaborative governance, involvement of the elderly in the policy-making process, lifelong learning, economic hardship, policies for all ages, isolated vulnerable and fragile groups, intergenerational contacts, safety in all policies, labor opportunities, and conditions and transport network improvement.

  19. A spectral method to detect community structure based on distance modularity matrix

    Science.gov (United States)

    Yang, Jin-Xuan; Zhang, Xiao-Dong

    2017-08-01

    There are many community organizations in social and biological networks. How to identify these community structure in complex networks has become a hot issue. In this paper, an algorithm to detect community structure of networks is proposed by using spectra of distance modularity matrix. The proposed algorithm focuses on the distance of vertices within communities, rather than the most weakly connected vertex pairs or number of edges between communities. The experimental results show that our method achieves better effectiveness to identify community structure for a variety of real-world networks and computer generated networks with a little more time-consumption.

  20. A divisive spectral method for network community detection

    International Nuclear Information System (INIS)

    Cheng, Jianjun; Li, Longjie; Yao, Yukai; Chen, Xiaoyun; Leng, Mingwei; Lu, Weiguo

    2016-01-01

    Community detection is a fundamental problem in the domain of complex network analysis. It has received great attention, and many community detection methods have been proposed in the last decade. In this paper, we propose a divisive spectral method for identifying community structures from networks which utilizes a sparsification operation to pre-process the networks first, and then uses a repeated bisection spectral algorithm to partition the networks into communities. The sparsification operation makes the community boundaries clearer and sharper, so that the repeated spectral bisection algorithm extract high-quality community structures accurately from the sparsified networks. Experiments show that the combination of network sparsification and a spectral bisection algorithm is highly successful, the proposed method is more effective in detecting community structures from networks than the others. (paper: interdisciplinary statistical mechanics)

  1. A framework for detecting communities of unbalanced sizes in networks

    Science.gov (United States)

    Žalik, Krista Rizman; Žalik, Borut

    2018-01-01

    Community detection in large networks has been a focus of recent research in many of fields, including biology, physics, social sciences, and computer science. Most community detection methods partition the entire network into communities, groups of nodes that have many connections within communities and few connections between them and do not identify different roles that nodes can have in communities. We propose a community detection model that integrates more different measures that can fast identify communities of different sizes and densities. We use node degree centrality, strong similarity with one node from community, maximal similarity of node to community, compactness of communities and separation between communities. Each measure has its own strength and weakness. Thus, combining different measures can benefit from the strengths of each one and eliminate encountered problems of using an individual measure. We present a fast local expansion algorithm for uncovering communities of different sizes and densities and reveals rich information on input networks. Experimental results show that the proposed algorithm is better or as effective as the other community detection algorithms for both real-world and synthetic networks while it requires less time.

  2. An Evaluation of Algorithms for Identifying Metastatic Breast, Lung, or Colorectal Cancer in Administrative Claims Data.

    Science.gov (United States)

    Whyte, Joanna L; Engel-Nitz, Nicole M; Teitelbaum, April; Gomez Rey, Gabriel; Kallich, Joel D

    2015-07-01

    Administrative health care claims data are used for epidemiologic, health services, and outcomes cancer research and thus play a significant role in policy. Cancer stage, which is often a major driver of cost and clinical outcomes, is not typically included in claims data. Evaluate algorithms used in a dataset of cancer patients to identify patients with metastatic breast (BC), lung (LC), or colorectal (CRC) cancer using claims data. Clinical data on BC, LC, or CRC patients (between January 1, 2007 and March 31, 2010) were linked to a health care claims database. Inclusion required health plan enrollment ≥3 months before initial cancer diagnosis date. Algorithms were used in the claims database to identify patients' disease status, which was compared with physician-reported metastases. Generic and tumor-specific algorithms were evaluated using ICD-9 codes, varying diagnosis time frames, and including/excluding other tumors. Positive and negative predictive values, sensitivity, and specificity were assessed. The linked databases included 14,480 patients; of whom, 32%, 17%, and 14.2% had metastatic BC, LC, and CRC, respectively, at diagnosis and met inclusion criteria. Nontumor-specific algorithms had lower specificity than tumor-specific algorithms. Tumor-specific algorithms' sensitivity and specificity were 53% and 99% for BC, 55% and 85% for LC, and 59% and 98% for CRC, respectively. Algorithms to distinguish metastatic BC, LC, and CRC from locally advanced disease should use tumor-specific primary cancer codes with 2 claims for the specific primary cancer >30-42 days apart to reduce misclassification. These performed best overall in specificity, positive predictive values, and overall accuracy to identify metastatic cancer in a health care claims database.

  3. Identifying Perceived Neighborhood Stressors Across Diverse Communities in New York City.

    Science.gov (United States)

    Shmool, Jessie L C; Yonas, Michael A; Newman, Ogonnaya Dotson; Kubzansky, Laura D; Joseph, Evelyn; Parks, Ana; Callaway, Charles; Chubb, Lauren G; Shepard, Peggy; Clougherty, Jane E

    2015-09-01

    There is growing interest in the role of psychosocial stress in health disparities. Identifying which social stressors are most important to community residents is critical for accurately incorporating stressor exposures into health research. Using a community-academic partnered approach, we designed a multi-community study across the five boroughs of New York City to characterize resident perceptions of key neighborhood stressors. We conducted 14 community focus groups; two to three in each borough, with one adolescent group and one Spanish-speaking group per borough. We then used systematic content analysis and participant ranking data to describe prominent neighborhood stressors and identify dominant themes. Three inter-related themes regarding the social and structural sources of stressful experiences were most commonly identified across neighborhoods: (1) physical disorder and perceived neglect, (2) harassment by police and perceived safety and (3) gentrification and racial discrimination. Our findings suggest that multiple sources of distress, including social, political, physical and economic factors, should be considered when investigating health effects of community stressor exposures and psychological distress. Community expertise is essential for comprehensively characterizing the range of neighborhood stressors that may be implicated in psychosocial exposure pathways.

  4. Validation of case-finding algorithms derived from administrative data for identifying adults living with human immunodeficiency virus infection.

    Directory of Open Access Journals (Sweden)

    Tony Antoniou

    Full Text Available OBJECTIVE: We sought to validate a case-finding algorithm for human immunodeficiency virus (HIV infection using administrative health databases in Ontario, Canada. METHODS: We constructed 48 case-finding algorithms using combinations of physician billing claims, hospital and emergency room separations and prescription drug claims. We determined the test characteristics of each algorithm over various time frames for identifying HIV infection, using data abstracted from the charts of 2,040 randomly selected patients receiving care at two medical practices in Toronto, Ontario as the reference standard. RESULTS: With the exception of algorithms using only a single physician claim, the specificity of all algorithms exceeded 99%. An algorithm consisting of three physician claims over a three year period had a sensitivity and specificity of 96.2% (95% CI 95.2%-97.9% and 99.6% (95% CI 99.1%-99.8%, respectively. Application of the algorithm to the province of Ontario identified 12,179 HIV-infected patients in care for the period spanning April 1, 2007 to March 31, 2009. CONCLUSIONS: Case-finding algorithms generated from administrative data can accurately identify adults living with HIV. A relatively simple "3 claims in 3 years" definition can be used for assembling a population-based cohort and facilitating future research examining trends in health service use and outcomes among HIV-infected adults in Ontario.

  5. New algorithms for identifying the flavour of B0 mesons using pions and protons

    NARCIS (Netherlands)

    Aaij, R.; Adeva, B.; Adinolfi, M.; Ajaltouni, Z.; Akar, S.; Albrecht, J.; Alessio, F.; Alexander, M.; Ali, S.; Alkhazov, G.; Alvarez Cartelle, P.; Alves, A. A.; Amato, S.; Amerio, S.; Amhis, Y.; Everse, LA; Anderlini, L.; Andreassi, G.; Andreotti, M.; Andrews, J.E.; Appleby, R. B.; Archilli, F.; d’Argent, P.; Arnau Romeu, J.; Artamonov, A.; Artuso, M.; Aslanides, E.; Auriemma, G.; Baalouch, M.; Babuschkin, I.; Bachmann, S.; Back, J. J.; Badalov, A.; Baesso, C.; Baker, S.C.; Baldini, W.; Barlow, R. J.; Barschel, C.; Barsuk, S.; Barter, W.; Baszczyk, M.; Batozskaya, V.; Batsukh, B.; Battista, V.; Bay, A.; Beaucourt, L.; Beddow, J.; Bedeschi, F.; Bediaga, I.; Bel, L. J.; Bellee, V.; Belloli, N.; Belous, K.; Belyaev, I.; Ben-Haim, E.; Bencivenni, G.; Benson, S.; Benton, J.; Berezhnoy, A.; Bernet, R.; Bertolin, A.; Betti, F.; Bettler, M-O.; Van Beuzekom, Martin; Bezshyiko, Ia; Bifani, S.; Billoir, P.; Bird, T.D.; Birnkraut, A.; Bitadze, A.; Bizzeti, A.; Blake, T.; Blanc, F.; Blouw, J.; Blusk, S.; Bocci, V.; Boettcher, T.; Bondar, A.; Bondar, N.; Bonivento, W.; Bordyuzhin, I.; Borgheresi, A.; Borghi, S.; Borisyak, M.; Borsato, M.; Bossu, F.; Boubdir, M.; Bowcock, T. J. V.; Bowen, E.; Bozzi, C.; Braun, S.; Britsch, M.; Britton, T.; Brodzicka, J.; Buchanan, E.; Burr, C.; Bursche, A.; Buytaert, J.; Cadeddu, S.; Calabrese, R.; Calvi, M.; Calvo Gomez, M.; Camboni, A.; Campana, P.; Campora Perez, D.; Campora Perez, D. H.; Capriotti, L.; Carbone, A.; Carboni, G.; Cardinale, R.; Cardini, A.; Carniti, P.; Carson, L.; Carvalho Akiba, K.; Casse, G.; Cassina, L.; Castillo Garcia, L.; Cattaneo, M.; Cauet, Ch; Cavallero, G.; Cenci, R.; Charles, M.; Charpentier, Ph; Chatzikonstantinidis, G.; Chefdeville, M.; Chen, S.; Cheung, S-F.; Chobanova, V.; Chrzaszcz, M.; Cid Vidal, X.; Ciezarek, G.; Clarke, P. E. L.; Clemencic, M.; Cliff, H. V.; Closier, J.; Coco, V.; Cogan, J.; Cogneras, E.; Cogoni, V.; Cojocariu, L.; Collins, P.; Comerma-Montells, A.; Contu, A.; Cook, A.; Coombs, G.; Coquereau, S.; Corti, G.; Corvo, M.; Costa Sobral, C. M.; Couturier, B.; Cowan, G. A.; Craik, D. C.; Crocombe, A.; Cruz Torres, M.; Cunliffe, S.; Currie, C.R.; D’Ambrosio, C.; Da Cunha Marinho, F.; Dall’Occo, E.; Dalseno, J.; David, P. N.Y.; Davis, A.; De Aguiar Francisco, O.; De Bruyn, K.; De Capua, S.; De Cian, M.; de Miranda, J. M.; Paula, L.E.; De Serio, M.; De Simone, P.; Dean, C-T.; Decamp, D.; Deckenhoff, M.; Del Buono, L.; Demmer, M.; Dendek, A.; Derkach, D.; Deschamps, O.; Dettori, F.; Dey, B.; Di Canto, A.; Dijkstra, H.; Dordei, F.; Dorigo, M.; Dosil Suárez, A.; Dovbnya, A.; Dreimanis, K.; Dufour, L.; Dujany, G.; Dungs, K.; Durante, P.; Dzhelyadin, R.; Dziurda, A.; Dzyuba, A.; Déléage, N.; Easo, S.; Ebert, M.; Egede, U.; Egorychev, V.; Eidelman, S.; Eisenhardt, S.; Eitschberger, U.; Ekelhof, R.; Eklund, L.; Elsasser, Ch.; Ely, S.; Esen, S.; Evans, H. M.; Evans, T. M.; Falabella, A.; Farley, N.; Farry, S.; Fay, R.; Fazzini, D.; Ferguson, D.; Fernandez Prieto, A.; Ferrari, F.; Ferreira Rodrigues, F.; Ferro-Luzzi, M.; Filippov, S.; Fini, R. A.; Fiore, M.; Fiorini, M.; Firlej, M.; Fitzpatrick, C.; Fiutowski, T.; Fleuret, F.; Fohl, K.; Fontana, Mark; Fontanelli, F.; Forshaw, D. C.; Forty, R.; Franco Lima, V.; Frank, M.; Frei, C.; Fu, J.; Furfaro, E.; Färber, C.; Gallas Torreira, A.; Galli, D.; Gallorini, S.; Gambetta, S.; Gandelman, M.; Gandini, P.; Gao, Y.; Garcia Martin, L. M.; García Pardiñas, J.; Garra Tico, J.; Garrido, L.; Garsed, P. J.; Gascon, D.; Carvalho-Gaspar, M.; Gavardi, L.; Gazzoni, G.; Gerick, D.; Gersabeck, E.; Gersabeck, M.; Gershon, T. J.; Ghez, Ph; Gianì, S.; Gibson, V.; Girard, O. G.; Giubega, L.; Gizdov, K.; Gligorov, V. V.; Golubkov, D.; Golutvin, A.; Gomes, A.Q.; Gorelov, I. V.; Gotti, C.; Grabalosa Gándara, M.; Graciani Diaz, R.; Granado Cardoso, L. A.; Graugés, E.; Graverini, E.; Graziani, G.; Grecu, A.; Griffith, P.; Grillo, L.; Gruberg Cazon, B. R.; Grünberg, O.; Gushchin, E.; Guz, Yu; Gys, T.; Göbel, C.; Hadavizadeh, T.; Hadjivasiliou, C.; Haefeli, G.; Haen, C.; Haines, S. C.; Hall, S.; Hamilton, B.; Han, X.; Hansmann-Menzemer, S.; Harnew, N.; Harnew, S. T.; Harrison, J.; Hatch, M.J.; He, J.; Head, T.; Heister, A.J.G.A.M.; Hennessy, K.; Henrard, P.; Henry, L.; Hernando Morata, J. A.; van Herwijnen, E.; Heß, M.; Hicheur, A.; Hill, D.; Hombach, C.; Hopchev, P. H.; Hulsbergen, W.; Humair, T.; Hushchyn, M.; Hussain, N.; Hutchcroft, D. E.; Idzik, M.; Ilten, P.; Jacobsson, R.; Jaeger, A.; Jalocha, J.; Jans, E.; Jawahery, A.; Jiang, F.; John, M.; Johnson, D.; Jones, C. R.; Joram, C.; Jost, B.; Jurik, N.; Kandybei, S.; Kanso, W.; Karacson, M.; Kariuki, J. M.; Karodia, S.; Kecke, M.; Kelsey, M. H.; Kenyon, I. R.; Kenzie, M.; Ketel, T.; Khairullin, E.; Khanji, B.; Khurewathanakul, C.; Kirn, T.; Klaver, S.M.; Klimaszewski, K.; Koliiev, S.; Kolpin, M.; Komarov, I.; Koopman, R. F.; Koppenburg, P.; Kosmyntseva, A.; Kozeiha, M.; Kravchuk, L.; Kreplin, K.; Kreps, M.; Krokovny, P.; Kruse, F.; Krzemien, W.; Kucewicz, W.; Kucharczyk, M.; Kudryavtsev, V.; Kuonen, A. K.; Kurek, K.; Kvaratskheliya, T.; Lacarrere, D.; Lafferty, G. D.; Lai, A.; Lambert, D.M.; Lanfranchi, G.; Langenbruch, C.; Latham, T. E.; Lazzeroni, C.; Le Gac, R.; van Leerdam, J.; Lees, J. P.; Leflat, A.; Lefrançois, J.; Lefèvre, R.; Lemaitre, F.; Lemos Cid, E.; Leroy, O.; Lesiak, T.; Leverington, B.; Li, Y.; Likhomanenko, T.; Lindner, R.; Linn, S.C.; Lionetto, F.; Liu, B.; Liu, X.; Loh, D.; Longstaff, I.; Lopes, J. H.; Lucchesi, D.; Lucio Martinez, M.; Luo, H.; Lupato, A.; Luppi, E.; Lupton, O.; Lusiani, A.; Lyu, X.; Machefert, F.; Maciuc, F.; Maev, O.; Maguire, K.; Malde, S.; Malinin, A.; Maltsev, T.; Manca, G.; Mancinelli, G.; Manning, P.; Maratas, J.; Marchand, J. F.; Marconi, U.; Marin Benito, C.; Marino, P.; Marks, J.; Martellotti, G.; Martin, M.; Martinelli-Boneschi, F.; Martinez-Santos, D.; Martinez-Vidal, F.; Martins Tostes, D.; Massacrier, L. M.; Massafferri, A.; Matev, R.; Mathad, A.; Mathe, Z.; Matteuzzi, C.; Mauri, A.; Maurin, B.; Mazurov, A.; McCann, M.; McCarthy, J.; Mcnab, A.; McNulty, R.; Meadows, B. T.; Meier, F.; Meissner, M.; Melnychuk, D.; Merk, M.; Merli, A.; Michielin, E.; Milanes, D. A.; Minard, M. N.; Mitzel, D. S.; Mogini, A.; Molina Rodriguez, J.; Monroy, I. A.; Monteil, S.; Morandin, M.; Morawski, P.; Mordà, A.; Morello, M. J.; Moron, J.; Morris, A. B.; Mountain, R.; Muheim, F.; Mulder, M.; Mussini, M.; Müller, D.; Müller, J.; Müller, Karl; von Müller, L.; Naik, P.; Nakada, T.; Nandakumar, R.; Nandi, A.; Nasteva, I.; Needham, M.; Neri, N.; Neubert, S.; Neufeld, N.; Neuner, M.; Nguyen, A. D.; Nguyen, T. D.; Nguyen-Mau, C.; Nieswand, S.; Niet, R.; Nikitin, N.; Nikodem, T.; Novoselov, A.; O’Hanlon, D. P.; Oblakowska-Mucha, A.; Obraztsov, V.; Ogilvy, S.; Oldeman, R.; Onderwater, C. J.G.; Otalora Goicochea, J. M.; Otto, E.A.; Owen, R.P.; Oyanguren, A.; Pais, P. R.; Palano, A.; Palombo, F.; Palutan, M.; Panman, J.; Papanestis, A.; Pappagallo, M.; Pappalardo, L.L.; Parker, W.S; Parkes, C.; Passaleva, G.; Pastore, A.; Patel, G. D.; Patel, M.; Patrignani, C.; Pearce, D.A.; Pellegrino, A.; Penso, G.; Pepe Altarelli, M.; Perazzini, S.; Perret, P.; Pescatore, L.; Petridis, K.; Petrolini, A.; Petrov, A.; Petruzzo, M.; Picatoste Olloqui, E.; Pietrzyk, B.; Pikies, M.; Pinci, D.; Pistone, A.; Piucci, A.; Playfer, S.; Plo Casasus, M.; Poikela, T.; Polci, F.; Poluektov, A.; Polyakov, I.; Polycarpo, E.; Pomery, G. J.; Popov, A.; Popov, D.; Popovici, B.; Poslavskii, S.; Potterat, C.; Price, M. E.; Price, J.D.; Prisciandaro, J.; Pritchard, C.A.; Prouve, C.; Pugatch, V.; Puig Navarro, A.; Punzi, G.; Qian, Y.W.; Quagliani, R.; Rachwal, B.; Rademacker, J. H.; Rama, M.; Ramos Pernas, M.; Rangel, M. S.; Raniuk, I.; Ratnikov, F.; Raven, G.; Redi, F.; Reichert, S.; dos Reis, A. C.; Remon Alepuz, C.; Renaudin, V.; Ricciardi, S.; Richards, Jennifer S; Rihl, M.; Rinnert, K.; Rives Molina, V.; Robbe, P.; Rodrigues, A. B.; Rodrigues, L.E.T.; Rodriguez Lopez, J. A.; Rodriguez Perez, P.; Rogozhnikov, A.; Roiser, S.; Rollings, A.; Romanovskiy, V.; Romero Vidal, A.; Ronayne, J. W.; Rotondo, M.; Rudolph, M. S.; Ruf, T.; Ruiz Valls, P.; Saborido Silva, J. J.; Sadykhov, E.; Sagidova, N.; Saitta, B.; Salustino Guimaraes, V.; Sanchez Mayordomo, C.; Sanmartin Sedes, B.; Santacesaria, R.; Santamarina Rios, C.; Santimaria, M.; Santovetti, E.; Sarti, A.; Satriano, C.; Satta, A.; Saunders, D. M.; Savrina, D.; Schael, S.; Schellenberg, M.; Schiller, M.; Schindler, R. H.; Schlupp, M.; Schmelling, M.; Schmelzer, T.; Schmidt, B.; Schneider, O.; Schopper, A.; Schubert, K.; Schubiger, M.; Schune, M. H.; Schwemmer, R.; Sciascia, B.; Sciubba, A.; Semennikov, A.; Sergi, A; Serra, N.; Serrano, J.; Sestini, L.; Seyfert, P.; Shapkin, M.; Shapoval, I.; Shcheglov, Y.; Shears, T.; Shekhtman, L.; Shevchenko, V.; Shires, A.; Siddi, B. G.; Silva Coutinho, R.; Silva de Oliveira, L.; Simi, G.; Simone, S.; Sirendi, M.; Skidmore, N.; Skwarnicki, T.; Smith, E.; Smith, I. T.; Smith, J; Smith, M.; Snoek, H.; Sokoloff, M. D.; Soler, F. J. P.; Souza De Paula, B.; Spaan, B.; Spradlin, P.; Sridharan, S.; Stagni, F.; Stahl, M.; Stahl, S.; Stefko, P.; Stefkova, S.; Steinkamp, O.; Stemmle, S.; Stenyakin, O.; Stevenson-Moore, P.; Stoica, S.; Stone, S.; Storaci, B.; Stracka, S.; Straticiuc, M.; Straumann, U.; Sun, L.; Sutcliffe, W.; Swientek, K.; Syropoulos, V.; Szczekowski, M.; Szumlak, T.; T’Jampens, S.; Tayduganov, A.; Tekampe, T.; Teklishyn, M.; Tellarini, G.; Teubert, F.; Thomas, E.; van Tilburg, J.; Tilley, M. J.; Tisserand, V.; Tobin, M. N.; Tolk, S.; Tomassetti, L.; Tonelli, D.; Topp-Joergensen, S.; Toriello, F.; Tournefier, E.; Tourneur, S.; Trabelsi, K.; Traill, M.; Tran, N.T.M.T.; Tresch, M.; Trisovic, A.; Tsaregorodtsev, A.; Tsopelas, P.; Tully, M.A.; Tuning, N.; Ukleja, A.; Ustyuzhanin, A.; Uwer, U.; Vacca, C.; Vagnoni, V.; Valassi, A.; Valat, S.; Valenti, G.; Vallier, A.; Vazquez Gomez, R.; Vazquez Regueiro, P.; Vecchi, S.; van Veghel-Plandsoen, M.M.; Velthuis, M.J.; Veltri, M.; Veneziano, G.; Venkateswaran, A.; Vernet, M.; Vesterinen, M.; Viaud, B.; Vieira, D.; Vieites Diaz, M.; Vilasis-Cardona, X.; Volkov, V.; Vollhardt, A.; Voneki, B.; Vorobyev, A.; Vorobyev, V.; Voß, C.; de Vries, J. A.; Vázquez Sierra, C.; Waldi, R.; Wallace, C.; Wallace, R.; Walsh, John; Wang, J.; Ward, D. R.; Wark, H. M.; Watson, N. K.; Websdale, D.; Weiden, A.; Whitehead, M.; Wicht, J.; Wilkinson, G.; Wilkinson, M.; Williams, M.; Williams, M.P.; Williams, M.; Williams, T.; Wilson, James F; Wimberley, J.; Wishahi, J.; Wislicki, W.; Witek, M.; Wormser, G.; Wotton, S. A.; Wraight, K.; Wyllie, K.; Xie, Y.; Xu, Z.; Yang, Z.; Yin, H; Yu, J.; Yuan, X.; Yushchenko, O.; Zarebski, K. A.; Zavertyaev, M.; Zhang, L.; Zhang, Y.; Zhelezov, A.; Zheng, Y.; Zhokhov, A.; Zhu, X.; Zhukov, V.; Zucchelli, S.

    2017-01-01

    Two new algorithms for use in the analysis of pp collision are developed to identify the flavour of B0 mesons at production using pions and protons from the hadronization process. The algorithms are optimized and calibrated on data, using B0→D-π+ decays from pp collision data collected by LHCb at

  6. ContextD: an algorithm to identify contextual properties of medical terms in a Dutch clinical corpus.

    Science.gov (United States)

    Afzal, Zubair; Pons, Ewoud; Kang, Ning; Sturkenboom, Miriam C J M; Schuemie, Martijn J; Kors, Jan A

    2014-11-29

    In order to extract meaningful information from electronic medical records, such as signs and symptoms, diagnoses, and treatments, it is important to take into account the contextual properties of the identified information: negation, temporality, and experiencer. Most work on automatic identification of these contextual properties has been done on English clinical text. This study presents ContextD, an adaptation of the English ConText algorithm to the Dutch language, and a Dutch clinical corpus. We created a Dutch clinical corpus containing four types of anonymized clinical documents: entries from general practitioners, specialists' letters, radiology reports, and discharge letters. Using a Dutch list of medical terms extracted from the Unified Medical Language System, we identified medical terms in the corpus with exact matching. The identified terms were annotated for negation, temporality, and experiencer properties. To adapt the ConText algorithm, we translated English trigger terms to Dutch and added several general and document specific enhancements, such as negation rules for general practitioners' entries and a regular expression based temporality module. The ContextD algorithm utilized 41 unique triggers to identify the contextual properties in the clinical corpus. For the negation property, the algorithm obtained an F-score from 87% to 93% for the different document types. For the experiencer property, the F-score was 99% to 100%. For the historical and hypothetical values of the temporality property, F-scores ranged from 26% to 54% and from 13% to 44%, respectively. The ContextD showed good performance in identifying negation and experiencer property values across all Dutch clinical document types. Accurate identification of the temporality property proved to be difficult and requires further work. The anonymized and annotated Dutch clinical corpus can serve as a useful resource for further algorithm development.

  7. Decoding communities in networks.

    Science.gov (United States)

    Radicchi, Filippo

    2018-02-01

    According to a recent information-theoretical proposal, the problem of defining and identifying communities in networks can be interpreted as a classical communication task over a noisy channel: memberships of nodes are information bits erased by the channel, edges and nonedges in the network are parity bits introduced by the encoder but degraded through the channel, and a community identification algorithm is a decoder. The interpretation is perfectly equivalent to the one at the basis of well-known statistical inference algorithms for community detection. The only difference in the interpretation is that a noisy channel replaces a stochastic network model. However, the different perspective gives the opportunity to take advantage of the rich set of tools of coding theory to generate novel insights on the problem of community detection. In this paper, we illustrate two main applications of standard coding-theoretical methods to community detection. First, we leverage a state-of-the-art decoding technique to generate a family of quasioptimal community detection algorithms. Second and more important, we show that the Shannon's noisy-channel coding theorem can be invoked to establish a lower bound, here named as decodability bound, for the maximum amount of noise tolerable by an ideal decoder to achieve perfect detection of communities. When computed for well-established synthetic benchmarks, the decodability bound explains accurately the performance achieved by the best community detection algorithms existing on the market, telling us that only little room for their improvement is still potentially left.

  8. Decoding communities in networks

    Science.gov (United States)

    Radicchi, Filippo

    2018-02-01

    According to a recent information-theoretical proposal, the problem of defining and identifying communities in networks can be interpreted as a classical communication task over a noisy channel: memberships of nodes are information bits erased by the channel, edges and nonedges in the network are parity bits introduced by the encoder but degraded through the channel, and a community identification algorithm is a decoder. The interpretation is perfectly equivalent to the one at the basis of well-known statistical inference algorithms for community detection. The only difference in the interpretation is that a noisy channel replaces a stochastic network model. However, the different perspective gives the opportunity to take advantage of the rich set of tools of coding theory to generate novel insights on the problem of community detection. In this paper, we illustrate two main applications of standard coding-theoretical methods to community detection. First, we leverage a state-of-the-art decoding technique to generate a family of quasioptimal community detection algorithms. Second and more important, we show that the Shannon's noisy-channel coding theorem can be invoked to establish a lower bound, here named as decodability bound, for the maximum amount of noise tolerable by an ideal decoder to achieve perfect detection of communities. When computed for well-established synthetic benchmarks, the decodability bound explains accurately the performance achieved by the best community detection algorithms existing on the market, telling us that only little room for their improvement is still potentially left.

  9. Dynamics in microbial communities: Unraveling mechanisms to identify principles

    Energy Technology Data Exchange (ETDEWEB)

    Konopka, Allan; Lindemann, Stephen R.; Fredrickson, Jim K.

    2015-07-01

    Diversity begets higher order properties such as functional stability and robustness in microbial communities, but principles that inform conceptual (and eventually predictive) models of community dynamics are lacking. Recent work has shown that selection as well as dispersal and drift shape communities, but the mechanistic bases for assembly of communities and the forces that maintain their function in the face of environmental perturbation are not well understood. Conceptually, some interactions among community members could generate endogenous dynamics in composition, even in the absence of environmental changes. These endogenous dynamics are further perturbed by exogenous forcing factors to produce a richer network of community interactions, and it is this “system” that is the basis for higher order community properties. Elucidation of principles that follow from this conceptual model requires identifying the mechanisms that (a) optimize diversity within a community and (b) impart community stability. The network of interactions between organisms can be an important element by providing a buffer against disturbance beyond the effect of functional redundancy, as alternative pathways with different combinations of microbes can be recruited to fulfill specific functions.

  10. A new neuro-fuzzy training algorithm for identifying dynamic characteristics of smart dampers

    International Nuclear Information System (INIS)

    Nguyen, Sy Dzung; Choi, Seung-Bok

    2012-01-01

    This paper proposes a new algorithm, named establishing neuro-fuzzy system (ENFS), to identify dynamic characteristics of smart dampers such as magnetorheological (MR) and electrorheological (ER) dampers. In the ENFS, data clustering is performed based on the proposed algorithm named partitioning data space (PDS). Firstly, the PDS builds data clusters in joint input–output data space with appropriate constraints. The role of these constraints is to create reasonable data distribution in clusters. The ENFS then uses these clusters to perform the following tasks. Firstly, the fuzzy sets expressing characteristics of data clusters are established. The structure of the fuzzy sets is adjusted to be suitable for features of the data set. Secondly, an appropriate structure of neuro-fuzzy (NF) expressed by an optimal number of labeled data clusters and the fuzzy-set groups is determined. After the ENFS is introduced, its effectiveness is evaluated by a prediction-error-comparative work between the proposed method and some other methods in identifying numerical data sets such as ‘daily data of stock A’, or in identifying a function. The ENFS is then applied to identify damping force characteristics of the smart dampers. In order to evaluate the effectiveness of the ENFS in identifying the damping forces of the smart dampers, the prediction errors are presented by comparing with experimental results. (paper)

  11. A new neuro-fuzzy training algorithm for identifying dynamic characteristics of smart dampers

    Science.gov (United States)

    Dzung Nguyen, Sy; Choi, Seung-Bok

    2012-08-01

    This paper proposes a new algorithm, named establishing neuro-fuzzy system (ENFS), to identify dynamic characteristics of smart dampers such as magnetorheological (MR) and electrorheological (ER) dampers. In the ENFS, data clustering is performed based on the proposed algorithm named partitioning data space (PDS). Firstly, the PDS builds data clusters in joint input-output data space with appropriate constraints. The role of these constraints is to create reasonable data distribution in clusters. The ENFS then uses these clusters to perform the following tasks. Firstly, the fuzzy sets expressing characteristics of data clusters are established. The structure of the fuzzy sets is adjusted to be suitable for features of the data set. Secondly, an appropriate structure of neuro-fuzzy (NF) expressed by an optimal number of labeled data clusters and the fuzzy-set groups is determined. After the ENFS is introduced, its effectiveness is evaluated by a prediction-error-comparative work between the proposed method and some other methods in identifying numerical data sets such as ‘daily data of stock A’, or in identifying a function. The ENFS is then applied to identify damping force characteristics of the smart dampers. In order to evaluate the effectiveness of the ENFS in identifying the damping forces of the smart dampers, the prediction errors are presented by comparing with experimental results.

  12. Leveraging disjoint communities for detecting overlapping community structure

    International Nuclear Information System (INIS)

    Chakraborty, Tanmoy

    2015-01-01

    Network communities represent mesoscopic structure for understanding the organization of real-world networks, where nodes often belong to multiple communities and form overlapping community structure in the network. Due to non-triviality in finding the exact boundary of such overlapping communities, this problem has become challenging, and therefore huge effort has been devoted to detect overlapping communities from the network.In this paper, we present PVOC (Permanence based Vertex-replication algorithm for Overlapping Community detection), a two-stage framework to detect overlapping community structure. We build on a novel observation that non-overlapping community structure detected by a standard disjoint community detection algorithm from a network has high resemblance with its actual overlapping community structure, except the overlapping part. Based on this observation, we posit that there is perhaps no need of building yet another overlapping community finding algorithm; but one can efficiently manipulate the output of any existing disjoint community finding algorithm to obtain the required overlapping structure. We propose a new post-processing technique that by combining with any existing disjoint community detection algorithm, can suitably process each vertex using a new vertex-based metric, called permanence, and thereby finds out overlapping candidates with their community memberships. Experimental results on both synthetic and large real-world networks show that PVOC significantly outperforms six state-of-the-art overlapping community detection algorithms in terms of high similarity of the output with the ground-truth structure. Thus our framework not only finds meaningful overlapping communities from the network, but also allows us to put an end to the constant effort of building yet another overlapping community detection algorithm. (paper)

  13. An Automated Algorithm for Identifying and Tracking Transverse Waves in Solar Images

    Science.gov (United States)

    Weberg, Micah J.; Morton, Richard J.; McLaughlin, James A.

    2018-01-01

    Recent instrumentation has demonstrated that the solar atmosphere supports omnipresent transverse waves, which could play a key role in energizing the solar corona. Large-scale studies are required in order to build up an understanding of the general properties of these transverse waves. To help facilitate this, we present an automated algorithm for identifying and tracking features in solar images and extracting the wave properties of any observed transverse oscillations. We test and calibrate our algorithm using a set of synthetic data, which includes noise and rotational effects. The results indicate an accuracy of 1%–2% for displacement amplitudes and 4%–10% for wave periods and velocity amplitudes. We also apply the algorithm to data from the Atmospheric Imaging Assembly on board the Solar Dynamics Observatory and find good agreement with previous studies. Of note, we find that 35%–41% of the observed plumes exhibit multiple wave signatures, which indicates either the superposition of waves or multiple independent wave packets observed at different times within a single structure. The automated methods described in this paper represent a significant improvement on the speed and quality of direct measurements of transverse waves within the solar atmosphere. This algorithm unlocks a wide range of statistical studies that were previously impractical.

  14. An Advanced Coupled Genetic Algorithm for Identifying Unknown Moving Loads on Bridge Decks

    Directory of Open Access Journals (Sweden)

    Sang-Youl Lee

    2014-01-01

    Full Text Available This study deals with an inverse method to identify moving loads on bridge decks using the finite element method (FEM and a coupled genetic algorithm (c-GA. We developed the inverse technique using a coupled genetic algorithm that can make global solution searches possible as opposed to classical gradient-based optimization techniques. The technique described in this paper allows us to not only detect the weight of moving vehicles but also find their moving velocities. To demonstrate the feasibility of the method, the algorithm is applied to a bridge deck model with beam elements. In addition, 1D and 3D finite element models are simulated to study the influence of measurement errors and model uncertainty between numerical and real structures. The results demonstrate the excellence of the method from the standpoints of computation efficiency and avoidance of premature convergence.

  15. Medical chart validation of an algorithm for identifying multiple sclerosis relapse in healthcare claims.

    Science.gov (United States)

    Chastek, Benjamin J; Oleen-Burkey, Merrikay; Lopez-Bresnahan, Maria V

    2010-01-01

    Relapse is a common measure of disease activity in relapsing-remitting multiple sclerosis (MS). The objective of this study was to test the content validity of an operational algorithm for detecting relapse in claims data. A claims-based relapse detection algorithm was tested by comparing its detection rate over a 1-year period with relapses identified based on medical chart review. According to the algorithm, MS patients in a US healthcare claims database who had either (1) a primary claim for MS during hospitalization or (2) a corticosteroid claim following a MS-related outpatient visit were designated as having a relapse. Patient charts were examined for explicit indication of relapse or care suggestive of relapse. Positive and negative predictive values were calculated. Medical charts were reviewed for 300 MS patients, half of whom had a relapse according to the algorithm. The claims-based criteria correctly classified 67.3% of patients with relapses (positive predictive value) and 70.0% of patients without relapses (negative predictive value; kappa 0.373: p value of the operational algorithm. Limitations of the algorithm include lack of differentiation between relapsing-remitting MS and other types, and that it does not incorporate measures of function and disability. The claims-based algorithm appeared to successfully detect moderate-to-severe MS relapse. This validated definition can be applied to future claims-based MS studies.

  16. Geographically Modified PageRank Algorithms: Identifying the Spatial Concentration of Human Movement in a Geospatial Network.

    Science.gov (United States)

    Chin, Wei-Chien-Benny; Wen, Tzai-Hung

    2015-01-01

    A network approach, which simplifies geographic settings as a form of nodes and links, emphasizes the connectivity and relationships of spatial features. Topological networks of spatial features are used to explore geographical connectivity and structures. The PageRank algorithm, a network metric, is often used to help identify important locations where people or automobiles concentrate in the geographical literature. However, geographic considerations, including proximity and location attractiveness, are ignored in most network metrics. The objective of the present study is to propose two geographically modified PageRank algorithms-Distance-Decay PageRank (DDPR) and Geographical PageRank (GPR)-that incorporate geographic considerations into PageRank algorithms to identify the spatial concentration of human movement in a geospatial network. Our findings indicate that in both intercity and within-city settings the proposed algorithms more effectively capture the spatial locations where people reside than traditional commonly-used network metrics. In comparing location attractiveness and distance decay, we conclude that the concentration of human movement is largely determined by the distance decay. This implies that geographic proximity remains a key factor in human mobility.

  17. Biclique communities

    DEFF Research Database (Denmark)

    Jørgensen, Sune Lehmann; Hansen-Schwartz, Martin; Hansen, Lars Kai

    2008-01-01

    We present a method for detecting communities in bipartite networks. Based on an extension of the k-clique community detection algorithm, we demonstrate how modular structure in bipartite networks presents itself as overlapping bicliques. If bipartite information is available, the biclique...... community detection algorithm retains all of the advantages of the k-clique algorithm, but avoids discarding important structural information when performing a one-mode projection of the network. Further, the biclique community detection algorithm provides a level of flexibility by incorporating independent...

  18. Electronic Health Record Based Algorithm to Identify Patients with Autism Spectrum Disorder.

    Directory of Open Access Journals (Sweden)

    Todd Lingren

    Full Text Available Cohort selection is challenging for large-scale electronic health record (EHR analyses, as International Classification of Diseases 9th edition (ICD-9 diagnostic codes are notoriously unreliable disease predictors. Our objective was to develop, evaluate, and validate an automated algorithm for determining an Autism Spectrum Disorder (ASD patient cohort from EHR. We demonstrate its utility via the largest investigation to date of the co-occurrence patterns of medical comorbidities in ASD.We extracted ICD-9 codes and concepts derived from the clinical notes. A gold standard patient set was labeled by clinicians at Boston Children's Hospital (BCH (N = 150 and Cincinnati Children's Hospital and Medical Center (CCHMC (N = 152. Two algorithms were created: (1 rule-based implementing the ASD criteria from Diagnostic and Statistical Manual of Mental Diseases 4th edition, (2 predictive classifier. The positive predictive values (PPV achieved by these algorithms were compared to an ICD-9 code baseline. We clustered the patients based on grouped ICD-9 codes and evaluated subgroups.The rule-based algorithm produced the best PPV: (a BCH: 0.885 vs. 0.273 (baseline; (b CCHMC: 0.840 vs. 0.645 (baseline; (c combined: 0.864 vs. 0.460 (baseline. A validation at Children's Hospital of Philadelphia yielded 0.848 (PPV. Clustering analyses of comorbidities on the three-site large cohort (N = 20,658 ASD patients identified psychiatric, developmental, and seizure disorder clusters.In a large cross-institutional cohort, co-occurrence patterns of comorbidities in ASDs provide further hypothetical evidence for distinct courses in ASD. The proposed automated algorithms for cohort selection open avenues for other large-scale EHR studies and individualized treatment of ASD.

  19. Eigenspaces of networks reveal the overlapping and hierarchical community structure more precisely

    International Nuclear Information System (INIS)

    Ma, Xiaoke; Gao, Lin; Yong, Xuerong

    2010-01-01

    Identifying community structure is fundamental for revealing the structure–functionality relationship in complex networks, and spectral algorithms have been shown to be powerful for this purpose. In a traditional spectral algorithm, each vertex of a network is embedded into a spectral space by making use of the eigenvectors of the adjacency matrix or Laplacian matrix of the graph. In this paper, a novel spectral approach for revealing the overlapping and hierarchical community structure of complex networks is proposed by not only using the eigenvalues and eigenvectors but also the properties of eigenspaces of the networks involved. This gives us a better characterization of community. We first show that the communicability between a pair of vertices can be rewritten in term of eigenspaces of a network. An agglomerative clustering algorithm is then presented to discover the hierarchical communities using the communicability matrix. Finally, these overlapping vertices are discovered with the corresponding eigenspaces, based on the fact that the vertices more densely connected amongst one another are more likely to be linked through short cycles. Compared with the traditional spectral algorithms, our algorithm can identify both the overlapping and hierarchical community without increasing the time complexity O(n 3 ), where n is the size of the network. Furthermore, our algorithm can also distinguish the overlapping vertices from bridges. The method is tested by applying it to some computer-generated and real-world networks. The experimental results indicate that our algorithm can reveal community structure more precisely than the traditional spectral approaches

  20. Geographically Modified PageRank Algorithms: Identifying the Spatial Concentration of Human Movement in a Geospatial Network.

    Directory of Open Access Journals (Sweden)

    Wei-Chien-Benny Chin

    Full Text Available A network approach, which simplifies geographic settings as a form of nodes and links, emphasizes the connectivity and relationships of spatial features. Topological networks of spatial features are used to explore geographical connectivity and structures. The PageRank algorithm, a network metric, is often used to help identify important locations where people or automobiles concentrate in the geographical literature. However, geographic considerations, including proximity and location attractiveness, are ignored in most network metrics. The objective of the present study is to propose two geographically modified PageRank algorithms-Distance-Decay PageRank (DDPR and Geographical PageRank (GPR-that incorporate geographic considerations into PageRank algorithms to identify the spatial concentration of human movement in a geospatial network. Our findings indicate that in both intercity and within-city settings the proposed algorithms more effectively capture the spatial locations where people reside than traditional commonly-used network metrics. In comparing location attractiveness and distance decay, we conclude that the concentration of human movement is largely determined by the distance decay. This implies that geographic proximity remains a key factor in human mobility.

  1. A simple algorithm for identifying periods of snow accumulation on a radiometer

    Science.gov (United States)

    Lapo, Karl E.; Hinkelman, Laura M.; Landry, Christopher C.; Massmann, Adam K.; Lundquist, Jessica D.

    2015-09-01

    Downwelling solar, Qsi, and longwave, Qli, irradiances at the earth's surface are the primary energy inputs for many hydrologic processes, and uncertainties in measurements of these two terms confound evaluations of estimated irradiances and negatively impact hydrologic modeling. Observations of Qsi and Qli in cold environments are subject to conditions that create additional uncertainties not encountered in other climates, specifically the accumulation of snow on uplooking radiometers. To address this issue, we present an automated method for estimating these periods of snow accumulation. Our method is based on forest interception of snow and uses common meteorological observations. In this algorithm, snow accumulation must exceed a threshold to obscure the sensor and is only removed through scouring by wind or melting. The algorithm is evaluated at two sites representing different mountain climates: (1) Snoqualmie Pass, Washington (maritime) and (2) the Senator Beck Basin Study Area, Colorado (continental). The algorithm agrees well with time-lapse camera observations at the Washington site and with multiple measurements at the Colorado site, with 70-80% of observed snow accumulation events correctly identified. We suggest using the method for quality controlling irradiance observations in snow-dominated climates where regular, daily maintenance is not possible.

  2. Costs per Diagnosis of Acute HIV Infection in Community-based Screening Strategies: A Comparative Analysis of Four Screening Algorithms

    Science.gov (United States)

    Hoenigl, Martin; Graff-Zivin, Joshua; Little, Susan J.

    2016-01-01

    Background. In nonhealthcare settings, widespread screening for acute human immunodeficiency virus (HIV) infection (AHI) is limited by cost and decision algorithms to better prioritize use of resources. Comparative cost analyses for available strategies are lacking. Methods. To determine cost-effectiveness of community-based testing strategies, we evaluated annual costs of 3 algorithms that detect AHI based on HIV nucleic acid amplification testing (EarlyTest algorithm) or on HIV p24 antigen (Ag) detection via Architect (Architect algorithm) or Determine (Determine algorithm) as well as 1 algorithm that relies on HIV antibody testing alone (Antibody algorithm). The cost model used data on men who have sex with men (MSM) undergoing community-based AHI screening in San Diego, California. Incremental cost-effectiveness ratios (ICERs) per diagnosis of AHI were calculated for programs with HIV prevalence rates between 0.1% and 2.9%. Results. Among MSM in San Diego, EarlyTest was cost-savings (ie, ICERs per AHI diagnosis less than $13.000) when compared with the 3 other algorithms. Cost analyses relative to regional HIV prevalence showed that EarlyTest was cost-effective (ie, ICERs less than $69.547) for similar populations of MSM with an HIV prevalence rate >0.4%; Architect was the second best alternative for HIV prevalence rates >0.6%. Conclusions. Identification of AHI by the dual EarlyTest screening algorithm is likely to be cost-effective not only among at-risk MSM in San Diego but also among similar populations of MSM with HIV prevalence rates >0.4%. PMID:26508512

  3. A novel community detection method in bipartite networks

    Science.gov (United States)

    Zhou, Cangqi; Feng, Liang; Zhao, Qianchuan

    2018-02-01

    Community structure is a common and important feature in many complex networks, including bipartite networks, which are used as a standard model for many empirical networks comprised of two types of nodes. In this paper, we propose a two-stage method for detecting community structure in bipartite networks. Firstly, we extend the widely-used Louvain algorithm to bipartite networks. The effectiveness and efficiency of the Louvain algorithm have been proved by many applications. However, there lacks a Louvain-like algorithm specially modified for bipartite networks. Based on bipartite modularity, a measure that extends unipartite modularity and that quantifies the strength of partitions in bipartite networks, we fill the gap by developing the Bi-Louvain algorithm that iteratively groups the nodes in each part by turns. This algorithm in bipartite networks often produces a balanced network structure with equal numbers of two types of nodes. Secondly, for the balanced network yielded by the first algorithm, we use an agglomerative clustering method to further cluster the network. We demonstrate that the calculation of the gain of modularity of each aggregation, and the operation of joining two communities can be compactly calculated by matrix operations for all pairs of communities simultaneously. At last, a complete hierarchical community structure is unfolded. We apply our method to two benchmark data sets and a large-scale data set from an e-commerce company, showing that it effectively identifies community structure in bipartite networks.

  4. Community Mining Method of Label Propagation Based on Dense Pairs

    Directory of Open Access Journals (Sweden)

    WENG Wei

    2014-03-01

    Full Text Available In recent years, with the popularity of handheld Internet equipments like mobile phones, increasing numbers of people are becoming involved in the virtual social network. Because of its large amount of data and complex structure, the network faces new challenges of community mining. A label propagation algorithm with low time complexity and without prior parameters deals easily with a large networks. This study explored a new method of community mining, based on label propagation with two stages. The first stage involved identifying closely linked nodes according to their local adjacency relations that gave rise to a micro-community. The second stage involved expanding and adjusting this community through a label propagation algorithm (LPA to finally obtain the community structure of the entire social network. This algorithm reduced the number of initial labels and avoided the merging of small communities in general LPAs. Thus, the quality of community discovery was improved, and the linear time complexity of the LPA was maintained.

  5. Using Community-Based Participatory Research to Identify Environmental Justice Issues in an Inner-City Community and Inform Urban Planning.

    Science.gov (United States)

    Mansyur, Carol Leler; Jeng, Hueiwang Anna; Holloman, Erica; DeBrew, Linwood

    2016-01-01

    The Southeast CARE Coalition has been using community-based participatory research to examine environmental degradation in the Southeast Community, Newport News, Virginia. A survey was developed to collect assessment data. Up to 66% of respondents were concerned about environmental problems in their community. Those with health conditions were significantly more likely to identify specific environmental problems. The top 5 environmental concerns included coal dust, air quality, crime, water quality, and trash. The community-based participatory research process is building community capacity and participation, providing community input into strategic planning, and empowering community members to take control of environmental justice issues in their community.

  6. Identifying nuclear power plant transients using the Discrete Binary Artificial Bee Colony (DBABC) algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Oliveira, Iona M.S. de; Schirru, Roberto, E-mail: ioliveira@con.ufrj.br, E-mail: schirru@lmp.ufrj.br [Coordenacoa dos Programas de Pos-Graduacao em Engenharia (UFRJ/PEN/COPPE), Rio de Janeiro, RJ (Brazil). Programa de Engenharia Nuclear

    2011-07-01

    The identification of possible transients in a nuclear power plant is a highly relevant problem. This is mainly due to the fact that the operation of a nuclear power plant involves a large number of state variables whose behaviors are extremely dynamic. In risk situations, besides the huge cognitive overload that operators are submitted to, there is also the problem related with the considerable decrease in the effective time for correct decision making. To minimize these problems and help operators to make the corrective actions in due time, this paper presents a new contribution in this area and introduces an experimental transient identification system based exclusively on the abilities of the Discrete Binary Artificial Bee Colony (DBABC) algorithm to find the best centroid positions that correctly identifies a transient in a nuclear power plant. The DBABC is a reworking of the Artificial Bee Colony (ABC) algorithm which presents the advantage of operating in both continuous and discrete search spaces. Through the analysis of experimental results, the effective performance of the proposed DBABC algorithm is shown against some well known best performing algorithms from the literature. (author)

  7. Identifying nuclear power plant transients using the Discrete Binary Artificial Bee Colony (DBABC) algorithm

    International Nuclear Information System (INIS)

    Oliveira, Iona M.S. de; Schirru, Roberto

    2011-01-01

    The identification of possible transients in a nuclear power plant is a highly relevant problem. This is mainly due to the fact that the operation of a nuclear power plant involves a large number of state variables whose behaviors are extremely dynamic. In risk situations, besides the huge cognitive overload that operators are submitted to, there is also the problem related with the considerable decrease in the effective time for correct decision making. To minimize these problems and help operators to make the corrective actions in due time, this paper presents a new contribution in this area and introduces an experimental transient identification system based exclusively on the abilities of the Discrete Binary Artificial Bee Colony (DBABC) algorithm to find the best centroid positions that correctly identifies a transient in a nuclear power plant. The DBABC is a reworking of the Artificial Bee Colony (ABC) algorithm which presents the advantage of operating in both continuous and discrete search spaces. Through the analysis of experimental results, the effective performance of the proposed DBABC algorithm is shown against some well known best performing algorithms from the literature. (author)

  8. Communities in Large Networks: Identification and Ranking

    DEFF Research Database (Denmark)

    Olsen, Martin

    2008-01-01

    show that the problem of deciding whether a non trivial community exists is NP complete. Nevertheless, experiments show that a very simple greedy approach can identify members of a community in the Danish part of the web graph with time complexity only dependent on the size of the found community...... and its immediate surroundings. The members are ranked with a “local” variant of the PageRank algorithm. Results are reported from successful experiments on identifying and ranking Danish Computer Science sites and Danish Chess pages using only a few representatives....

  9. Communities in Large Networks: Identification and Ranking

    DEFF Research Database (Denmark)

    Olsen, Martin

    2008-01-01

    We study the problem of identifying and ranking the members of a community in a very large network with link analysis only, given a set of representatives of the community. We define the concept of a community justified by a formal analysis of a simple model of the evolution of a directed graph. ...... and its immediate surroundings. The members are ranked with a “local” variant of the PageRank algorithm. Results are reported from successful experiments on identifying and ranking Danish Computer Science sites and Danish Chess pages using only a few representatives....

  10. Identifying hidden sexual bridging communities in Chicago.

    Science.gov (United States)

    Youm, Yoosik; Mackesy-Amiti, Mary Ellen; Williams, Chyvette T; Ouellet, Lawrence J

    2009-07-01

    Bridge populations can play a central role in the spread of human immunodeficiency virus (HIV) by providing transmission links between higher and lower prevalence populations. While social network methods are well suited to the study of bridge populations, analyses tend to focus on dyads (i.e., risk between drug and/or sex partners) and ignore bridges between distinct subpopulations. This study takes initial steps toward moving the analysis of sexual network linkages beyond individual and risk group levels to a community level in which Chicago's 77 community areas are examined as subpopulations for the purpose of identifying potential bridging communities. Of particular interest are "hidden" bridging communities; that is, areas with above-average levels of sexual ties with other areas but whose below-average AIDS prevalence may hide their potential importance for HIV prevention. Data for this analysis came from the first wave of recruiting at the Chicago Sexual Acquisition and Transmission of HIV Cooperative Agreement Program site. Between August 2005 through October 2006, respondent-driven sampling was used to recruit users of heroin, cocaine, or methamphetamine, men who have sex with men regardless of drug use, the sex partners of these two groups, and sex partners of the sex partners. In this cross-sectional study of the sexual transmission of HIV, participants completed a network-focused computer-assisted self-administered interview, which included questions about the geographic locations of sexual contacts with up to six recent partners. Bridging scores for each area were determined using a matrix representing Chicago's 77 community areas and were assessed using two measures: non-redundant ties and flow betweenness. Bridging measures and acquired immunodeficiency syndrome (AIDS) case prevalence rates were plotted for each community area on charts representing four conditions: below-average bridging and AIDS prevalence, below-average bridging and above

  11. Hillslope characterization: Identifying key controls on local-scale plant communities' distribution using remote sensing and subsurface data fusion.

    Science.gov (United States)

    Falco, N.; Wainwright, H. M.; Dafflon, B.; Leger, E.; Peterson, J.; Steltzer, H.; Wilmer, C.; Williams, K. H.; Hubbard, S. S.

    2017-12-01

    Mountainous watershed systems are characterized by extreme heterogeneity in hydrological and pedological properties that influence biotic activities, plant communities and their dynamics. To gain predictive understanding of how ecosystem and watershed system evolve under climate change, it is critical to capture such heterogeneity and to quantify the effect of key environmental variables such as topography, and soil properties. In this study, we exploit advanced geophysical and remote sensing techniques - coupled with machine learning - to better characterize and quantify the interactions between plant communities' distribution and subsurface properties. First, we have developed a remote sensing data fusion framework based on the random forest (RF) classification algorithm to estimate the spatial distribution of plant communities. The framework allows the integration of both plant spectral and structural information, which are derived from multispectral satellite images and airborne LiDAR data. We then use the RF method to evaluate the estimated plant community map, exploiting the subsurface properties (such as bedrock depth, soil moisture and other properties) and geomorphological parameters (such as slope, curvature) as predictors. Datasets include high-resolution geophysical data (electrical resistivity tomography) and LiDAR digital elevation maps. We demonstrate our approach on a mountain hillslope and meadow within the East River watershed in Colorado, which is considered to be a representative headwater catchment in the Upper Colorado Basin. The obtained results show the existence of co-evolution between above and below-ground processes; in particular, dominant shrub communities in wet and flat areas. We show that successful integration of remote sensing data with geophysical measurements allows identifying and quantifying the key environmental controls on plant communities' distribution, and provides insights into their potential changes in the future

  12. [Study of algorithms to identify schizophrenia in the SNIIRAM database conducted by the REDSIAM network].

    Science.gov (United States)

    Quantin, C; Collin, C; Frérot, M; Besson, J; Cottenet, J; Corneloup, M; Soudry-Faure, A; Mariet, A-S; Roussot, A

    2017-10-01

    The aim of the REDSIAM network is to foster communication between users of French medico-administrative databases and to validate and promote analysis methods suitable for the data. Within this network, the working group "Mental and behavioral disorders" took an interest in algorithms to identify adult schizophrenia in the SNIIRAM database and inventoried identification criteria for patients with schizophrenia in these databases. The methodology was based on interviews with nine experts in schizophrenia concerning the procedures they use to identify patients with schizophrenia disorders in databases. The interviews were based on a questionnaire and conducted by telephone. The synthesis of the interviews showed that the SNIIRAM contains various tables which allow coders to identify patients suffering from schizophrenia: chronic disease status, drugs and hospitalizations. Taken separately, these criteria were not sufficient to recognize patients with schizophrenia, an algorithm should be based on all of them. Apparently, only one-third of people living with schizophrenia benefit from the longstanding disease status. Not all patients are hospitalized, and coding for diagnoses at the hospitalization, notably for short stays in medicine, surgery or obstetrics departments, is not exhaustive. As for treatment with antipsychotics, it is not specific enough as such treatments are also prescribed to patients with bipolar disorders, or even other disorders. It seems appropriate to combine these complementary criteria, while keeping in mind out-patient care (every year 80,000 patients are seen exclusively in an outpatient setting), even if these data are difficult to link with other information. Finally, the experts made three propositions for selection algorithms of patients with schizophrenia. Patients with schizophrenia can be relatively accurately identified using SNIIRAM data. Different combinations of the selected criteria must be used depending on the objectives and

  13. Identifying opportunities in online-communities

    DEFF Research Database (Denmark)

    Hienerth, C.; Lettl, Christopher

    how this phenomenon - as manifested in user communities - can be used to derive deeper insights into the prominent phases of opportunity identification, evaluation and exploitation. We also outline how user communities create new avenues for empirical research on these early entrepreneurial processes....... Based on our analysis, we develop a set of hypotheses on how processes in user communities affect the outcome of entrepreneurial activities....

  14. PedMine – A simulated annealing algorithm to identify maximally unrelated individuals in population isolates

    OpenAIRE

    Douglas, Julie A.; Sandefur, Conner I.

    2008-01-01

    In family-based genetic studies, it is often useful to identify a subset of unrelated individuals. When such studies are conducted in population isolates, however, most if not all individuals are often detectably related to each other. To identify a set of maximally unrelated (or equivalently, minimally related) individuals, we have implemented simulated annealing, a general-purpose algorithm for solving difficult combinatorial optimization problems. We illustrate our method on data from a ge...

  15. IDENTIFYING COMPETENCIES OF VOLUNTEER BOARD MEMBERS OF COMMUNITY SPORTS CLUBS

    OpenAIRE

    A. BALDUCK; A. VAN ROSSEM; M. BUELENS

    2009-01-01

    This study contributes to the emerging empirical studies on roles and responsibilities of boards in nonprofit organizations by identifying competencies of volunteer board members. We identified how two types of constituents—volunteer board members and sports members—perceived competencies of volunteer board members in community sports clubs. We used the repertory grid technique to draw cognitive maps and to reveal the perceived reality of these constituents. Our results suggest that constitue...

  16. Identifying elementary iterated systems through algorithmic inference: The Cantor set example

    Energy Technology Data Exchange (ETDEWEB)

    Apolloni, Bruno [Dipartimento di Scienze dell' Informazione, Universita degli Studi di Milano, Via Comelico 39/41, 20135 Milan (Italy)]. E-mail: apolloni@dsi.unimi.it; Bassis, Simone [Dipartimento di Scienze dell' Informazione, Universita degli Studi di Milano, Via Comelico 39/41, 20135 Milan (Italy)]. E-mail: bassis@dsi.unimi.it

    2006-10-15

    We come back to the old problem of fractal identification within the new framework of algorithmic Inference. The key points are: (i) to identify sufficient statistics to be put in connection with the unknown values of the fractal parameters, and (ii) to manage the timing of the iterated process through spatial statistics. We fill these tasks successfully with the Cantor sets. We are able to compute confidence intervals for both the scaling parameter {theta} and the iteration number n at which we are observing a set. We both check numerically the coverage of these intervals and delineate a general strategy for affording more complex iterated systems.

  17. An Empirical Comparison of Algorithms to Find Communities in Directed Graphs and Their Application in Web Data Analytics

    DEFF Research Database (Denmark)

    Agreste, Santa; De Meo, Pasquale; Fiumara, Giacomo

    2017-01-01

    Detecting communities in graphs is a fundamental tool to understand the structure of Web-based systems and predict their evolution. Many community detection algorithms are designed to process undirected graphs (i.e., graphs with bidirectional edges) but many graphs on the Web-e.g., microblogging ...... the best trade-off between accuracy and computational performance and, therefore, it has to be considered as a promising tool for Web Data Analytics purposes....

  18. Classifying spatially heterogeneous wetland communities using machine learning algorithms and spectral and textural features.

    Science.gov (United States)

    Szantoi, Zoltan; Escobedo, Francisco J; Abd-Elrahman, Amr; Pearlstine, Leonard; Dewitt, Bon; Smith, Scot

    2015-05-01

    Mapping of wetlands (marsh vs. swamp vs. upland) is a common remote sensing application.Yet, discriminating between similar freshwater communities such as graminoid/sedge fromremotely sensed imagery is more difficult. Most of this activity has been performed using medium to low resolution imagery. There are only a few studies using highspatial resolutionimagery and machine learning image classification algorithms for mapping heterogeneouswetland plantcommunities. This study addresses this void by analyzing whether machine learning classifierssuch as decisiontrees (DT) and artificial neural networks (ANN) can accurately classify graminoid/sedgecommunities usinghigh resolution aerial imagery and image texture data in the Everglades National Park, Florida.In addition tospectral bands, the normalized difference vegetation index, and first- and second-order texturefeatures derivedfrom the near-infrared band were analyzed. Classifier accuracies were assessed using confusiontablesand the calculated kappa coefficients of the resulting maps. The results indicated that an ANN(multilayerperceptron based on backpropagation) algorithm produced a statistically significantly higheraccuracy(82.04%) than the DT (QUEST) algorithm (80.48%) or the maximum likelihood (80.56%)classifier (αtexture features.

  19. Community detection by graph Voronoi diagrams

    Science.gov (United States)

    Deritei, Dávid; Lázár, Zsolt I.; Papp, István; Járai-Szabó, Ferenc; Sumi, Róbert; Varga, Levente; Ravasz Regan, Erzsébet; Ercsey-Ravasz, Mária

    2014-06-01

    Accurate and efficient community detection in networks is a key challenge for complex network theory and its applications. The problem is analogous to cluster analysis in data mining, a field rich in metric space-based methods. Common to these methods is a geometric, distance-based definition of clusters or communities. Here we propose a new geometric approach to graph community detection based on graph Voronoi diagrams. Our method serves as proof of principle that the definition of appropriate distance metrics on graphs can bring a rich set of metric space-based clustering methods to network science. We employ a simple edge metric that reflects the intra- or inter-community character of edges, and a graph density-based rule to identify seed nodes of Voronoi cells. Our algorithm outperforms most network community detection methods applicable to large networks on benchmark as well as real-world networks. In addition to offering a computationally efficient alternative for community detection, our method opens new avenues for adapting a wide range of data mining algorithms to complex networks from the class of centroid- and density-based clustering methods.

  20. An approach of community evolution based on gravitational relationship refactoring in dynamic networks

    International Nuclear Information System (INIS)

    Yin, Guisheng; Chi, Kuo; Dong, Yuxin; Dong, Hongbin

    2017-01-01

    In this paper, an approach of community evolution based on gravitational relationship refactoring between the nodes in a dynamic network is proposed, and it can be used to simulate the process of community evolution. A static community detection algorithm and a dynamic community evolution algorithm are included in the approach. At first, communities are initialized by constructing the core nodes chains, the nodes can be iteratively searched and divided into corresponding communities via the static community detection algorithm. For a dynamic network, an evolutionary process is divided into three phases, and behaviors of community evolution can be judged according to the changing situation of the core nodes chain in each community. Experiments show that the proposed approach can achieve accuracy and availability in the synthetic and real world networks. - Highlights: • The proposed approach considers both the static community detection and dynamic community evolution. • The approach of community evolution can identify the whole 6 common evolution events. • The proposed approach can judge the evolutionary events according to the variations of the core nodes chains.

  1. The CARPEDIEM Algorithm: A Rule-Based System for Identifying Heart Failure Phenotype with a Precision Public Health Approach

    Directory of Open Access Journals (Sweden)

    Michela Franchini

    2018-01-01

    Full Text Available Modern medicine remains dependent on the accurate evaluation of a patient’s health state, recognizing that disease is a process that evolves over time and interacts with many factors unique to that patient. The CARPEDIEM project represents a concrete attempt to address these issues by developing reproducible algorithms to support the accuracy in detection of complex diseases. This study aims to establish and validate the CARPEDIEM approach and algorithm for identifying those patients presenting with or at risk of heart failure (HF by studying 153,393 subjects in Italy, based on patient information flow databases and is not reliant on the electronic health record to accomplish its goals. The resulting algorithm has been validated in a two-stage process, comparing predicted results with (1 HF diagnosis as identified by general practitioners (GPs among the reference cohort and (2 HF diagnosis as identified by cardiologists within a randomly sampled subpopulation of 389 patients. The sources of data used to detect HF cases are numerous and were standardized for this study. The accuracy and the predictive values of the algorithm with respect to the GPs and the clinical standards are highly consistent with those from previous studies. In particular, the algorithm is more efficient in detecting the more severe cases of HF according to the GPs’ validation (specificity increases according to the number of comorbidities and external validation (NYHA: II–IV; HF severity index: 2, 3. Positive and negative predictive values reveal that the CARPEDIEM algorithm is most consistent with clinical evaluation performed in the specialist setting, while it presents a greater ability to rule out false-negative HF cases within the GP practice, probably as a consequence of the different HF prevalence in the two different care settings. Further development includes analyzing the clinical features of false-positive and -negative predictions, to explore the natural

  2. Overlapping communities from dense disjoint and high total degree clusters

    Science.gov (United States)

    Zhang, Hongli; Gao, Yang; Zhang, Yue

    2018-04-01

    Community plays an important role in the field of sociology, biology and especially in domains of computer science, where systems are often represented as networks. And community detection is of great importance in the domains. A community is a dense subgraph of the whole graph with more links between its members than between its members to the outside nodes, and nodes in the same community probably share common properties or play similar roles in the graph. Communities overlap when nodes in a graph belong to multiple communities. A vast variety of overlapping community detection methods have been proposed in the literature, and the local expansion method is one of the most successful techniques dealing with large networks. The paper presents a density-based seeding method, in which dense disjoint local clusters are searched and selected as seeds. The proposed method selects a seed by the total degree and density of local clusters utilizing merely local structures of the network. Furthermore, this paper proposes a novel community refining phase via minimizing the conductance of each community, through which the quality of identified communities is largely improved in linear time. Experimental results in synthetic networks show that the proposed seeding method outperforms other seeding methods in the state of the art and the proposed refining method largely enhances the quality of the identified communities. Experimental results in real graphs with ground-truth communities show that the proposed approach outperforms other state of the art overlapping community detection algorithms, in particular, it is more than two orders of magnitude faster than the existing global algorithms with higher quality, and it obtains much more accurate community structure than the current local algorithms without any priori information.

  3. Development and validation of an algorithm for identifying urinary retention in a cohort of patients with epilepsy in a large US administrative claims database.

    Science.gov (United States)

    Quinlan, Scott C; Cheng, Wendy Y; Ishihara, Lianna; Irizarry, Michael C; Holick, Crystal N; Duh, Mei Sheng

    2016-04-01

    The aim of this study was to develop and validate an insurance claims-based algorithm for identifying urinary retention (UR) in epilepsy patients receiving antiepileptic drugs to facilitate safety monitoring. Data from the HealthCore Integrated Research Database(SM) in 2008-2011 (retrospective) and 2012-2013 (prospective) were used to identify epilepsy patients with UR. During the retrospective phase, three algorithms identified potential UR: (i) UR diagnosis code with a catheterization procedure code; (ii) UR diagnosis code alone; or (iii) diagnosis with UR-related symptoms. Medical records for 50 randomly selected patients satisfying ≥1 algorithm were reviewed by urologists to ascertain UR status. Positive predictive value (PPV) and 95% confidence intervals (CI) were calculated for the three component algorithms and the overall algorithm (defined as satisfying ≥1 component algorithms). Algorithms were refined using urologist review notes. In the prospective phase, the UR algorithm was refined using medical records for an additional 150 cases. In the retrospective phase, the PPV of the overall algorithm was 72.0% (95%CI: 57.5-83.8%). Algorithm 3 performed poorly and was dropped. Algorithm 1 was unchanged; urinary incontinence and cystitis were added as exclusionary diagnoses to Algorithm 2. The PPV for the modified overall algorithm was 89.2% (74.6-97.0%). In the prospective phase, the PPV for the modified overall algorithm was 76.0% (68.4-82.6%). Upon adding overactive bladder, nocturia and urinary frequency as exclusionary diagnoses, the PPV for the final overall algorithm was 81.9% (73.7-88.4%). The current UR algorithm yielded a PPV > 80% and could be used for more accurate identification of UR among epilepsy patients in a large claims database. Copyright © 2016 John Wiley & Sons, Ltd.

  4. Coupling mode-destination accessibility with seismic risk assessment to identify at-risk communities

    International Nuclear Information System (INIS)

    Miller, Mahalia; Baker, Jack W.

    2016-01-01

    In this paper, we develop a framework for coupling mode-destination accessibility with quantitative seismic risk assessment to identify communities at high risk for travel disruptions after an earthquake. Mode-destination accessibility measures the ability of people to reach destinations they desire. We use a probabilistic seismic risk assessment procedure, including a stochastic set of earthquake events, ground-motion intensity maps, damage maps, and realizations of traffic and accessibility impacts. For a case study of the San Francisco Bay Area, we couple our seismic risk framework with a practical activity-based traffic model. As a result, we quantify accessibility risk probabilistically by community and household type. We find that accessibility varies more strongly as a function of travelers' geographic location than as a function of their income class, and we identify particularly at-risk communities. We also observe that communities more conducive to local trips by foot or bike are predicted to be less impacted by losses in accessibility. This work shows the potential to link quantitative risk assessment methodologies with high-resolution travel models used by transportation planners. Quantitative risk metrics of this type should have great utility for planners working to reduce risk to a region's infrastructure systems. - Highlights: • We couple mode-destination accessibility with probabilistic seismic risk assessment. • Results identify communities at high risk for post-earthquake travel disruptions. • Accessibility varies more as a function of home location than by income. • Our model predicts reduced accessibility risk for more walking-friendly communities.

  5. Identifying Effective and Sustainable Measures for Community-Based Environmental Monitoring

    Science.gov (United States)

    McKay, Ariana J.; Johnson, Chris J.

    2017-09-01

    Resource development projects typically result in monitoring programs that fail to fully consider the values and participation of surrounding communities. Also, monitoring protocols for single environmental values can be insufficient for addressing the cumulative impacts of resource development. Community-based environmental monitoring (CBEM) has emerged as a way to meaningfully include local citizens in the decision-making process and assessment of the development of natural resources. Our research explored how to develop effective and sustainable CBEM. Interviews were conducted with staff from 15 CBEM programs established across Canada to identify criteria of what constitutes effective CBEM. Results demonstrate that CBEM offers an effective, locally adapted, and culturally applicable approach to facilitate community participation in natural resource management and to track environmental change. Benefits of CBEM include: locally relevant monitoring protocols, inclusion of cumulative impacts, better informed decision-making, and increased awareness and collaboration amongst community, governments, and proponents. Challenges associated with CBEM are cost, capacity, longevity, distribution of results, and establishing credibility. This research validates the use of CBEM for improving resource management.

  6. Identifying barriers to mental health system improvements: an examination of community participation in assertive community treatment programs

    Directory of Open Access Journals (Sweden)

    Wakefield Patricia A

    2011-11-01

    Full Text Available Abstract Background Integrating the best available evidence into program standards is essential if system-wide improvements in the delivery of community-based mental health services are to be achieved. Since the beginning of the Assertive Community Treatment (ACT program movement, program standards have included a role for the community. In particular, ACT program standards have sought to ensure that members of the local community are involved in governance and that former clients participate in service delivery as "Peer Support Specialists". This paper reports on the extent to which ACT program standards related to community participation have been implemented and identifies barriers to full compliance. Methods Qualitative and quantitative data were collected through a telephone survey of ACT Program Coordinators in Ontario, Canada, using a census sample of the existing 66 ACT programs. A thematic approach to content analysis was used to analyze respondents' qualitative comments. Quantitative data were analyzed using SPSS 16.0 and included means, frequencies, independent t-tests and Pearson Correlations. Results An 85% response rate was achieved. Of the 33 program standards, the two that received the lowest perceived compliance ratings were the two standards directly concerning community participation. Specifically, the standard to have a functioning Community Advisory Body and the standard requiring the inclusion of a Peer Support Specialist. The three major themes that emerged from the survey data with respect to the barriers to fully implementing the Community Advisory Body were: external issues; standard related issues; and, organizational/structural related issues. The three major themes concerning barriers to implementing the Peer Support Specialist role were: human resource related issues; organizational/structural related issues; and, standard related issues. Conclusions The reasons for low compliance of ACT programs with community

  7. Path lumping: An efficient algorithm to identify metastable path channels for conformational dynamics of multi-body systems

    Science.gov (United States)

    Meng, Luming; Sheong, Fu Kit; Zeng, Xiangze; Zhu, Lizhe; Huang, Xuhui

    2017-07-01

    Constructing Markov state models from large-scale molecular dynamics simulation trajectories is a promising approach to dissect the kinetic mechanisms of complex chemical and biological processes. Combined with transition path theory, Markov state models can be applied to identify all pathways connecting any conformational states of interest. However, the identified pathways can be too complex to comprehend, especially for multi-body processes where numerous parallel pathways with comparable flux probability often coexist. Here, we have developed a path lumping method to group these parallel pathways into metastable path channels for analysis. We define the similarity between two pathways as the intercrossing flux between them and then apply the spectral clustering algorithm to lump these pathways into groups. We demonstrate the power of our method by applying it to two systems: a 2D-potential consisting of four metastable energy channels and the hydrophobic collapse process of two hydrophobic molecules. In both cases, our algorithm successfully reveals the metastable path channels. We expect this path lumping algorithm to be a promising tool for revealing unprecedented insights into the kinetic mechanisms of complex multi-body processes.

  8. Community Engagement for Identifying Cancer Education Needs in Puerto Rico.

    Science.gov (United States)

    Jiménez, Julio; Ramos, Axel; Ramos-Rivera, Francisco E; Gwede, Clement; Quinn, Gwendolyn P; Vadaparampil, Susan; Brandon, Thomas; Simmons, Vani; Castro, Eida

    2018-02-01

    Cancer is the leading cause of death in Puerto Rico, suggesting a need for improved strategies, programs, and resources devoted to cancer prevention. Enhanced prevention needs in Puerto Rico were initially identified in pilot studies conducted by the Ponce School of Medicine (PSM) in collaboration with the H. Lee Moffitt Cancer Center (MCC). In the current study, we used community engagement to identify specific needs in cancer prevention and education and strategies to create culturally attuned, effective cancer prevention education programs. A total of 37 participants attended a community forum and were assigned to one of three discussion groups: patients/survivors (n = 14); family/caregivers (n = 11); or healthcare providers (n = 12). Most participants were women (73 %), over 35 years of age, and a majority were married (58 %) and had a university education (81 %). The sessions were recorded and transcribed and analyzed for key themes. Participants wanted improved awareness of cancer prevention in Puerto Rico and believed cancer prevention education should start early, ideally in elementary school. Participants also stressed the importance of creating partnerships with private and government agencies to coordinate educational efforts. Suggested strategies included outreach to communities with limited resources, incorporating the testimony of cancer survivors, and utilizing social media to disseminate cancer prevention information.

  9. Identifying key hospital service quality factors in online health communities.

    Science.gov (United States)

    Jung, Yuchul; Hur, Cinyoung; Jung, Dain; Kim, Minki

    2015-04-07

    The volume of health-related user-created content, especially hospital-related questions and answers in online health communities, has rapidly increased. Patients and caregivers participate in online community activities to share their experiences, exchange information, and ask about recommended or discredited hospitals. However, there is little research on how to identify hospital service quality automatically from the online communities. In the past, in-depth analysis of hospitals has used random sampling surveys. However, such surveys are becoming impractical owing to the rapidly increasing volume of online data and the diverse analysis requirements of related stakeholders. As a solution for utilizing large-scale health-related information, we propose a novel approach to identify hospital service quality factors and overtime trends automatically from online health communities, especially hospital-related questions and answers. We defined social media-based key quality factors for hospitals. In addition, we developed text mining techniques to detect such factors that frequently occur in online health communities. After detecting these factors that represent qualitative aspects of hospitals, we applied a sentiment analysis to recognize the types of recommendations in messages posted within online health communities. Korea's two biggest online portals were used to test the effectiveness of detection of social media-based key quality factors for hospitals. To evaluate the proposed text mining techniques, we performed manual evaluations on the extraction and classification results, such as hospital name, service quality factors, and recommendation types using a random sample of messages (ie, 5.44% (9450/173,748) of the total messages). Service quality factor detection and hospital name extraction achieved average F1 scores of 91% and 78%, respectively. In terms of recommendation classification, performance (ie, precision) is 78% on average. Extraction and

  10. Overlapping community detection based on link graph using distance dynamics

    Science.gov (United States)

    Chen, Lei; Zhang, Jing; Cai, Li-Jun

    2018-01-01

    The distance dynamics model was recently proposed to detect the disjoint community of a complex network. To identify the overlapping structure of a network using the distance dynamics model, an overlapping community detection algorithm, called L-Attractor, is proposed in this paper. The process of L-Attractor mainly consists of three phases. In the first phase, L-Attractor transforms the original graph to a link graph (a new edge graph) to assure that one node has multiple distances. In the second phase, using the improved distance dynamics model, a dynamic interaction process is introduced to simulate the distance dynamics (shrink or stretch). Through the dynamic interaction process, all distances converge, and the disjoint community structure of the link graph naturally manifests itself. In the third phase, a recovery method is designed to convert the disjoint community structure of the link graph to the overlapping community structure of the original graph. Extensive experiments are conducted on the LFR benchmark networks as well as real-world networks. Based on the results, our algorithm demonstrates higher accuracy and quality than other state-of-the-art algorithms.

  11. Clustering analysis of water distribution systems: identifying critical components and community impacts.

    Science.gov (United States)

    Diao, K; Farmani, R; Fu, G; Astaraie-Imani, M; Ward, S; Butler, D

    2014-01-01

    Large water distribution systems (WDSs) are networks with both topological and behavioural complexity. Thereby, it is usually difficult to identify the key features of the properties of the system, and subsequently all the critical components within the system for a given purpose of design or control. One way is, however, to more explicitly visualize the network structure and interactions between components by dividing a WDS into a number of clusters (subsystems). Accordingly, this paper introduces a clustering strategy that decomposes WDSs into clusters with stronger internal connections than external connections. The detected cluster layout is very similar to the community structure of the served urban area. As WDSs may expand along with urban development in a community-by-community manner, the correspondingly formed distribution clusters may reveal some crucial configurations of WDSs. For verification, the method is applied to identify all the critical links during firefighting for the vulnerability analysis of a real-world WDS. Moreover, both the most critical pipes and clusters are addressed, given the consequences of pipe failure. Compared with the enumeration method, the method used in this study identifies the same group of the most critical components, and provides similar criticality prioritizations of them in a more computationally efficient time.

  12. Identifying Time Measurement Tampering in the Traversal Time and Hop Count Analysis (TTHCA Wormhole Detection Algorithm

    Directory of Open Access Journals (Sweden)

    Jonny Karlsson

    2013-05-01

    Full Text Available Traversal time and hop count analysis (TTHCA is a recent wormhole detection algorithm for mobile ad hoc networks (MANET which provides enhanced detection performance against all wormhole attack variants and network types. TTHCA involves each node measuring the processing time of routing packets during the route discovery process and then delivering the measurements to the source node. In a participation mode (PM wormhole where malicious nodes appear in the routing tables as legitimate nodes, the time measurements can potentially be altered so preventing TTHCA from successfully detecting the wormhole. This paper analyses the prevailing conditions for time tampering attacks to succeed for PM wormholes, before introducing an extension to the TTHCA detection algorithm called ∆T Vector which is designed to identify time tampering, while preserving low false positive rates. Simulation results confirm that the ∆T Vector extension is able to effectively detect time tampering attacks, thereby providing an important security enhancement to the TTHCA algorithm.

  13. Desired attributes of new graduate nurses as identified by the rural community.

    Science.gov (United States)

    Sivamalai, S

    2008-01-01

    Preparing nurse graduates for practice is challenging because of the diversity of skills expected of them. Increasingly consumers are more informed and expect quality care. To identify the attributes a rural community expect in new graduate nurses in order for them to provide quality care. A questionnaire was designed to assess the importance attached to a set of attributes of graduate nurses expected by a rural community. The community included a range of professionals working with government and hospitals, community volunteers and retired people. After pilot testing, the questionnaire was distributed using a cluster sampling technique. A total of 656 completed questionnaires were returned, giving a response rate of 69%. The respondents were asked to rate the importance of each item for the community on a five-point Likert scale (5 = extremely important, 4 = very important, 3 = moderately important, 2 = possibly importantly, and 1 = not important at all). Exploratory factor analysis was performed on the 38 items using SPSS (SPSS inc; Chicago, IL, USA). Principal Components Analysis was applied to identify the number of factors followed by Oblimin rotation. The sample of 656 respondents consisted of 68% females and 30% males (2% did not identify their gender). The majority of the respondents (75.6%) were born in Australia, while 3.2% were born in the UK Kingdom. Principal Components Analysis identified five factors with eigenvalues above one, explaining 47.4% of the total variance. Items that loaded greater than + or - 0.3, (approximately 10% of the common factor variance) was associated with the factor in question. Component 1 was labelled Sympathetic/ Patients' welfare with the item 'Nurses should be sensitive to the emotional needs of patients' showing the highest loading. Component 2 was called Contextual knowledge/ Interpersonal skills. It contained items indicating that nurses should have good personal skills and possess a broad contextual knowledge of issues

  14. Administrative Algorithms to identify Avascular necrosis of bone among patients undergoing upper or lower extremity magnetic resonance imaging: a validation study.

    Science.gov (United States)

    Barbhaiya, Medha; Dong, Yan; Sparks, Jeffrey A; Losina, Elena; Costenbader, Karen H; Katz, Jeffrey N

    2017-06-19

    Studies of the epidemiology and outcomes of avascular necrosis (AVN) require accurate case-finding methods. The aim of this study was to evaluate performance characteristics of a claims-based algorithm designed to identify AVN cases in administrative data. Using a centralized patient registry from a US academic medical center, we identified all adults aged ≥18 years who underwent magnetic resonance imaging (MRI) of an upper/lower extremity joint during the 1.5 year study period. A radiologist report confirming AVN on MRI served as the gold standard. We examined the sensitivity, specificity, positive predictive value (PPV) and positive likelihood ratio (LR + ) of four algorithms (A-D) using International Classification of Diseases, 9th edition (ICD-9) codes for AVN. The algorithms ranged from least stringent (Algorithm A, requiring ≥1 ICD-9 code for AVN [733.4X]) to most stringent (Algorithm D, requiring ≥3 ICD-9 codes, each at least 30 days apart). Among 8200 patients who underwent MRI, 83 (1.0% [95% CI 0.78-1.22]) had AVN by gold standard. Algorithm A yielded the highest sensitivity (81.9%, 95% CI 72.0-89.5), with PPV of 66.0% (95% CI 56.0-75.1). The PPV of algorithm D increased to 82.2% (95% CI 67.9-92.0), although sensitivity decreased to 44.6% (95% CI 33.7-55.9). All four algorithms had specificities >99%. An algorithm that uses a single billing code to screen for AVN among those who had MRI has the highest sensitivity and is best suited for studies in which further medical record review confirming AVN is feasible. Algorithms using multiple billing codes are recommended for use in administrative databases when further AVN validation is not feasible.

  15. An interactive algorithm for identifying multiattribute measurable value functions based on finite-order independence of structural difference

    International Nuclear Information System (INIS)

    Tamura, Hiroyuki; Hikita, Shiro

    1985-01-01

    In this paper, we develop an interactive algorithm for identifying multiattribute measurable value functions based on the concept of finite-order independence of structural difference. This concept includes Dyer and Sarin's weak difference independence as special cases. The algorithm developed is composed of four major parts: 1) formulation of the problem 2) assessment of normalized conditional value functions and structural difference functions 3) assessment of corner values 4) assessment of the order of independence of structural difference and selection of the model. A hypothetical numerical example of a trade-off analysis for siting a nuclear power plant is included. (author)

  16. Positive Predictive Values of International Classification of Diseases, 10th Revision Coding Algorithms to Identify Patients With Autosomal Dominant Polycystic Kidney Disease

    Directory of Open Access Journals (Sweden)

    Vinusha Kalatharan

    2016-12-01

    Full Text Available Background: International Classification of Diseases, 10th Revision codes (ICD-10 for autosomal dominant polycystic kidney disease (ADPKD is used within several administrative health care databases. It is unknown whether these codes identify patients who meet strict clinical criteria for ADPKD. Objective: The objective of this study is (1 to determine whether different ICD-10 coding algorithms identify adult patients who meet strict clinical criteria for ADPKD as assessed through medical chart review and (2 to assess the number of patients identified with different ADPKD coding algorithms in Ontario. Design: Validation study of health care database codes, and prevalence. Setting: Ontario, Canada. Patients: For the chart review, 201 adult patients with hospital encounters between April 1, 2002, and March 31, 2014, assigned either ICD-10 codes Q61.2 or Q61.3. Measurements: This study measured positive predictive value of the ICD-10 coding algorithms and the number of Ontarians identified with different coding algorithms. Methods: We manually reviewed a random sample of medical charts in London, Ontario, Canada, and determined whether or not ADPKD was present according to strict clinical criteria. Results: The presence of either ICD-10 code Q61.2 or Q61.3 in a hospital encounter had a positive predictive value of 85% (95% confidence interval [CI], 79%-89% and identified 2981 Ontarians (0.02% of the Ontario adult population. The presence of ICD-10 code Q61.2 in a hospital encounter had a positive predictive value of 97% (95% CI, 86%-100% and identified 394 adults in Ontario (0.003% of the Ontario adult population. Limitations: (1 We could not calculate other measures of validity; (2 the coding algorithms do not identify patients without hospital encounters; and (3 coding practices may differ between hospitals. Conclusions: Most patients with ICD-10 code Q61.2 or Q61.3 assigned during their hospital encounters have ADPKD according to the clinical

  17. Application and validation of case-finding algorithms for identifying individuals with human immunodeficiency virus from administrative data in British Columbia, Canada.

    Directory of Open Access Journals (Sweden)

    Bohdan Nosyk

    Full Text Available To define a population-level cohort of individuals infected with the human immunodeficiency virus (HIV in the province of British Columbia from available registries and administrative datasets using a validated case-finding algorithm.Individuals were identified for possible cohort inclusion from the BC Centre for Excellence in HIV/AIDS (CfE drug treatment program (antiretroviral therapy and laboratory testing datasets (plasma viral load (pVL and CD4 diagnostic test results, the BC Centre for Disease Control (CDC provincial HIV surveillance database (positive HIV tests, as well as databases held by the BC Ministry of Health (MoH; the Discharge Abstract Database (hospitalizations, the Medical Services Plan (physician billing and PharmaNet databases (additional HIV-related medications. A validated case-finding algorithm was applied to distinguish true HIV cases from those likely to have been misclassified. The sensitivity of the algorithms was assessed as the proportion of confirmed cases (those with records in the CfE, CDC and MoH databases positively identified by each algorithm. A priori hypotheses were generated and tested to verify excluded cases.A total of 25,673 individuals were identified as having at least one HIV-related health record. Among 9,454 unconfirmed cases, the selected case-finding algorithm identified 849 individuals believed to be HIV-positive. The sensitivity of this algorithm among confirmed cases was 88%. Those excluded from the cohort were more likely to be female (44.4% vs. 22.5%; p<0.01, had a lower mortality rate (2.18 per 100 person years (100PY vs. 3.14/100PY; p<0.01, and had lower median rates of health service utilization (days of medications dispensed: 9745/100PY vs. 10266/100PY; p<0.01; days of inpatient care: 29/100PY vs. 98/100PY; p<0.01; physician billings: 602/100PY vs. 2,056/100PY; p<0.01.The application of validated case-finding algorithms and subsequent hypothesis testing provided a strong framework for

  18. A Markov random walk under constraint for discovering overlapping communities in complex networks

    International Nuclear Information System (INIS)

    Jin, Di; Yang, Bo; Liu, Dayou; He, Dongxiao; Liu, Jie; Baquero, Carlos

    2011-01-01

    The detection of overlapping communities in complex networks has motivated recent research in relevant fields. Aiming to address this problem, we propose a Markov-dynamics-based algorithm, called UEOC, which means 'unfold and extract overlapping communities'. In UEOC, when identifying each natural community that overlaps, a Markov random walk method combined with a constraint strategy, which is based on the corresponding annealed network (degree conserving random network), is performed to unfold the community. Then, a cutoff criterion with the aid of a local community function, called conductance, which can be thought of as the ratio between the number of edges inside the community and those leaving it, is presented to extract this emerged community from the entire network. The UEOC algorithm depends on only one parameter whose value can be easily set, and it requires no prior knowledge of the hidden community structures. The proposed UEOC has been evaluated both on synthetic benchmarks and on some real-world networks, and has been compared with a set of competing algorithms. The experimental result has shown that UEOC is highly effective and efficient for discovering overlapping communities

  19. Identifying ozone-sensitive communities of (semi-)natural vegetation suitable for mapping exceedance of critical levels

    International Nuclear Information System (INIS)

    Mills, G.; Hayes, F.; Jones, M.L.M.; Cinderby, S.

    2007-01-01

    Using published data on the responses of individual species to ozone, 54 EUNIS (European Nature Information System) level 4 communities with six or more ozone-sensitive species (%OS) and c. 20% or more species tested for ozone sensitivity, were identified as potentially ozone-sensitive. The largest number of these communities (23) was associated with Grasslands, with Heathland, scrub and tundra, and Mires, bogs and fens having the next highest representation at 11 and 8 level 4 communities each respectively. Within the grasslands classification, E4 (Alpine and sub-alpine grasslands), E5 (Woodland fringes and clearings) and E1 (Dry grasslands) were the most sensitive with 68.1, 51.6 and 48.6%OS respectively. It is feasible to map the land-cover for these and other communities at level 2, but it may not be currently possible to map the land-cover for all communities identified to be ozone-sensitive at levels 3 and 4. - Grassland communities such as alpine and sub-alpine grasslands have the highest potential sensitivity ozone, based on the responses of their component species

  20. A Netnographic Study of Entrepreneurial Traits: Evaluating classic typologies using the crowdsourcing algorithm of an online community

    Directory of Open Access Journals (Sweden)

    Marcos Cerqueira Lima

    2014-09-01

    Full Text Available This paper evaluates how the advices of experienced entrepreneurs to young start-up creators in an online community reflect entrepreneurship traits commonly found in conceptual typologies. The overall goal is to contrast and evaluate existing models based on evidence from an online community. This should facilitate future studies to improve current typologies by ranking entrepreneurial traits according to perceived relevance. In order to achieve these objectives, we have conducted a “netnographic study” (i.e., the qualitative analysis of web-based content of 96 answers to the question “What is the best advice for a young, first-time startup CEO?” on Quora.com. Relying on Quora’s ranking algorithm (based on crowdsourcing of votes and community prestige, we focused on the top 50% of answers (which we shall call the “above Quora 50” category considered the most relevant by its 2000+ followers and 120,000+ viewers. We used Nvivo as a Qualitative Data Analysis Software to code all the entries into the literature categories. These codes were then later retrieved using matrix queries to compare the incidence of traits and the perceived relevance of answers. We found that among the 50% highest ranking answers on Quora, the following traits are perceived as the most important for young entrepreneurs to develop: management style, attitude in interpersonal relations, vision, self-concept, leadership style, marketing, market and customer knowledge, innovation, technical knowledge and skills, attitude to growth, ability to adapt, purpose and relations system. These results could lead to improving existing typologies and creating new models capable of better identifying people with the highest potential to succeed in new venture creation.

  1. Identifying Psoriasis and Psoriatic Arthritis Patients in Retrospective Databases When Diagnosis Codes Are Not Available: A Validation Study Comparing Medication/Prescriber Visit-Based Algorithms with Diagnosis Codes.

    Science.gov (United States)

    Dobson-Belaire, Wendy; Goodfield, Jason; Borrelli, Richard; Liu, Fei Fei; Khan, Zeba M

    2018-01-01

    Using diagnosis code-based algorithms is the primary method of identifying patient cohorts for retrospective studies; nevertheless, many databases lack reliable diagnosis code information. To develop precise algorithms based on medication claims/prescriber visits (MCs/PVs) to identify psoriasis (PsO) patients and psoriatic patients with arthritic conditions (PsO-AC), a proxy for psoriatic arthritis, in Canadian databases lacking diagnosis codes. Algorithms were developed using medications with narrow indication profiles in combination with prescriber specialty to define PsO and PsO-AC. For a 3-year study period from July 1, 2009, algorithms were validated using the PharMetrics Plus database, which contains both adjudicated medication claims and diagnosis codes. Positive predictive value (PPV), negative predictive value (NPV), sensitivity, and specificity of the developed algorithms were assessed using diagnosis code as the reference standard. Chosen algorithms were then applied to Canadian drug databases to profile the algorithm-identified PsO and PsO-AC cohorts. In the selected database, 183,328 patients were identified for validation. The highest PPVs for PsO (85%) and PsO-AC (65%) occurred when a predictive algorithm of two or more MCs/PVs was compared with the reference standard of one or more diagnosis codes. NPV and specificity were high (99%-100%), whereas sensitivity was low (≤30%). Reducing the number of MCs/PVs or increasing diagnosis claims decreased the algorithms' PPVs. We have developed an MC/PV-based algorithm to identify PsO patients with a high degree of accuracy, but accuracy for PsO-AC requires further investigation. Such methods allow researchers to conduct retrospective studies in databases in which diagnosis codes are absent. Copyright © 2018 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.

  2. Community detection with consideration of non-topological information

    International Nuclear Information System (INIS)

    Zou Sheng-Rong; Peng Yu-Jing; Liu Ai-Fen; Xu Xiu-Lian; He Da-Ren

    2011-01-01

    In a network described by a graph, only topological structure information is considered to determine how the nodes are connected by edges. Non-topological information denotes that which cannot be determined directly from topological information. This paper shows, by a simple example where scientists in three research groups and one external group form four communities, that in some real world networks non-topological information (in this example, the research group affiliation) dominates community division. If the information has some influence on the network topological structure, the question arises as to how to find a suitable algorithm to identify the communities based only on the network topology. We show that weighted Newman algorithm may be the best choice for this example. We believe that this idea is general for real-world complex networks. (interdisciplinary physics and related areas of science and technology)

  3. Identifying context-specific competencies required by community Australian Football sports trainers.

    Science.gov (United States)

    Donaldson, Alex; Finch, Caroline F

    2012-08-01

    First-aid is a recommended injury prevention and risk management strategy in community sport; however, little is known about the sport-specific competencies required by first-aid providers. To achieve expert consensus on the competencies required by community Australian Football (community-AF) sports trainers. A three-round online Delphi process. Community-AF. 16 Australian sports first-aid and community-AF experts. Rating of competencies as either 'essential', 'expected', 'ideal' or 'not required'. Results After Round 3, 47 of the 77 (61%) competencies were endorsed as 'essential' or 'expected' for a sports trainer to effectively perform the activities required to the standards expected at a community-AF club by ≥75% of experts. These competencies covered: the role of the sports trainer; the responsibilities of the sports trainer; emergency management; injury and illness assessment and immediate management; taping; and injury prevention and risk management. Four competencies (5%) were endorsed as 'ideal' or 'not required' by ≥85% of experts and were excluded from further consideration. The 26 competencies where consensus was not reached were retained as second-tier, optional competencies. Sports trainers are important members of on-field first-aid teams, providing support to both injured players and other sports medicine professionals. The competencies identified in this study provide the basis of a proposed two-tiered community-AF-specific sports trainer education structure that can be implemented by the peak sports body. This includes six mandatory modules, relating to the 'required' competencies, and a further six optional modules covering competencies on which consensus was not reached.

  4. Highly efficient computer algorithm for identifying layer thickness of atomically thin 2D materials

    Science.gov (United States)

    Lee, Jekwan; Cho, Seungwan; Park, Soohyun; Bae, Hyemin; Noh, Minji; Kim, Beom; In, Chihun; Yang, Seunghoon; Lee, Sooun; Seo, Seung Young; Kim, Jehyun; Lee, Chul-Ho; Shim, Woo-Young; Jo, Moon-Ho; Kim, Dohun; Choi, Hyunyong

    2018-03-01

    The fields of layered material research, such as transition-metal dichalcogenides (TMDs), have demonstrated that the optical, electrical and mechanical properties strongly depend on the layer number N. Thus, efficient and accurate determination of N is the most crucial step before the associated device fabrication. An existing experimental technique using an optical microscope is the most widely used one to identify N. However, a critical drawback of this approach is that it relies on extensive laboratory experiences to estimate N; it requires a very time-consuming image-searching task assisted by human eyes and secondary measurements such as atomic force microscopy and Raman spectroscopy, which are necessary to ensure N. In this work, we introduce a computer algorithm based on the image analysis of a quantized optical contrast. We show that our algorithm can apply to a wide variety of layered materials, including graphene, MoS2, and WS2 regardless of substrates. The algorithm largely consists of two parts. First, it sets up an appropriate boundary between target flakes and substrate. Second, to compute N, it automatically calculates the optical contrast using an adaptive RGB estimation process between each target, which results in a matrix with different integer Ns and returns a matrix map of Ns onto the target flake position. Using a conventional desktop computational power, the time taken to display the final N matrix was 1.8 s on average for the image size of 1280 pixels by 960 pixels and obtained a high accuracy of 90% (six estimation errors among 62 samples) when compared to the other methods. To show the effectiveness of our algorithm, we also apply it to TMD flakes transferred on optically transparent c-axis sapphire substrates and obtain a similar result of the accuracy of 94% (two estimation errors among 34 samples).

  5. Scalable Static and Dynamic Community Detection Using Grappolo

    Energy Technology Data Exchange (ETDEWEB)

    Halappanavar, Mahantesh; Lu, Hao; Kalyanaraman, Anantharaman; Tumeo, Antonino

    2017-09-12

    Graph clustering, popularly known as community detection, is a fundamental kernel for several applications of relevance to the Defense Advanced Research Projects Agency’s (DARPA) Hierarchical Identify Verify Exploit (HIVE) Pro- gram. Clusters or communities represent natural divisions within a network that are densely connected within a cluster and sparsely connected to the rest of the network. The need to compute clustering on large scale data necessitates the development of efficient algorithms that can exploit modern architectures that are fundamentally parallel in nature. How- ever, due to their irregular and inherently sequential nature, many of the current algorithms for community detection are challenging to parallelize. In response to the HIVE Graph Challenge, we present several parallelization heuristics for fast community detection using the Louvain method as the serial template. We implement all the heuristics in a software library called Grappolo. Using the inputs from the HIVE Challenge, we demonstrate superior performance and high quality solutions based on four parallelization heuristics. We use Grappolo on static graphs as the first step towards community detection on streaming graphs.

  6. Fast unfolding of communities in large networks

    International Nuclear Information System (INIS)

    Blondel, Vincent D; Guillaume, Jean-Loup; Lambiotte, Renaud; Lefebvre, Etienne

    2008-01-01

    We propose a simple method to extract the community structure of large networks. Our method is a heuristic method that is based on modularity optimization. It is shown to outperform all other known community detection methods in terms of computation time. Moreover, the quality of the communities detected is very good, as measured by the so-called modularity. This is shown first by identifying language communities in a Belgian mobile phone network of 2 million customers and by analysing a web graph of 118 million nodes and more than one billion links. The accuracy of our algorithm is also verified on ad hoc modular networks

  7. GTI: a novel algorithm for identifying outlier gene expression profiles from integrated microarray datasets.

    Directory of Open Access Journals (Sweden)

    John Patrick Mpindi

    Full Text Available BACKGROUND: Meta-analysis of gene expression microarray datasets presents significant challenges for statistical analysis. We developed and validated a new bioinformatic method for the identification of genes upregulated in subsets of samples of a given tumour type ('outlier genes', a hallmark of potential oncogenes. METHODOLOGY: A new statistical method (the gene tissue index, GTI was developed by modifying and adapting algorithms originally developed for statistical problems in economics. We compared the potential of the GTI to detect outlier genes in meta-datasets with four previously defined statistical methods, COPA, the OS statistic, the t-test and ORT, using simulated data. We demonstrated that the GTI performed equally well to existing methods in a single study simulation. Next, we evaluated the performance of the GTI in the analysis of combined Affymetrix gene expression data from several published studies covering 392 normal samples of tissue from the central nervous system, 74 astrocytomas, and 353 glioblastomas. According to the results, the GTI was better able than most of the previous methods to identify known oncogenic outlier genes. In addition, the GTI identified 29 novel outlier genes in glioblastomas, including TYMS and CDKN2A. The over-expression of these genes was validated in vivo by immunohistochemical staining data from clinical glioblastoma samples. Immunohistochemical data were available for 65% (19 of 29 of these genes, and 17 of these 19 genes (90% showed a typical outlier staining pattern. Furthermore, raltitrexed, a specific inhibitor of TYMS used in the therapy of tumour types other than glioblastoma, also effectively blocked cell proliferation in glioblastoma cell lines, thus highlighting this outlier gene candidate as a potential therapeutic target. CONCLUSIONS/SIGNIFICANCE: Taken together, these results support the GTI as a novel approach to identify potential oncogene outliers and drug targets. The algorithm is

  8. Derivation and validation of the automated search algorithms to identify cognitive impairment and dementia in electronic health records.

    Science.gov (United States)

    Amra, Sakusic; O'Horo, John C; Singh, Tarun D; Wilson, Gregory A; Kashyap, Rahul; Petersen, Ronald; Roberts, Rosebud O; Fryer, John D; Rabinstein, Alejandro A; Gajic, Ognjen

    2017-02-01

    Long-term cognitive impairment is a common and important problem in survivors of critical illness. We developed electronic search algorithms to identify cognitive impairment and dementia from the electronic medical records (EMRs) that provide opportunity for big data analysis. Eligible patients met 2 criteria. First, they had a formal cognitive evaluation by The Mayo Clinic Study of Aging. Second, they were hospitalized in intensive care unit at our institution between 2006 and 2014. The "criterion standard" for diagnosis was formal cognitive evaluation supplemented by input from an expert neurologist. Using all available EMR data, we developed and improved our algorithms in the derivation cohort and validated them in the independent validation cohort. Of 993 participants who underwent formal cognitive testing and were hospitalized in intensive care unit, we selected 151 participants at random to form the derivation and validation cohorts. The automated electronic search algorithm for cognitive impairment was 94.3% sensitive and 93.0% specific. The search algorithms for dementia achieved respective sensitivity and specificity of 97% and 99%. EMR search algorithms significantly outperformed International Classification of Diseases codes. Automated EMR data extractions for cognitive impairment and dementia are reliable and accurate and can serve as acceptable and efficient alternatives to time-consuming manual data review. Copyright © 2016 Elsevier Inc. All rights reserved.

  9. Community Detection for Large Graphs

    KAUST Repository

    Peng, Chengbin

    2014-05-04

    Many real world networks have inherent community structures, including social networks, transportation networks, biological networks, etc. For large scale networks with millions or billions of nodes in real-world applications, accelerating current community detection algorithms is in demand, and we present two approaches to tackle this issue -A K-core based framework that can accelerate existing community detection algorithms significantly; -A parallel inference algorithm via stochastic block models that can distribute the workload.

  10. An algorithm for identifying the best current friend in a social network

    Directory of Open Access Journals (Sweden)

    Francisco Javier Moreno

    2015-05-01

    Full Text Available A research field in the area of social networks (SNs is the identification of some types of users and groups. To facilitate this process, a SN is usually represented by a graph. The centrality measures, which identify the most important vertices in a graph according to some criterion, are usual tools to analyze a graph. One of these measures is the PageRank (a measure originally designed to classify web pages. Informally, in the context of a SN, the PageRank of a user i represents the probability that another user of the SN is seeing the page of i after a considerable time of navigation in the SN. In this paper, we define a new type of user in a SN: the best current friend. The idea is to identify, among the friends of a user i, who is the friend k that would generate the highest decrease in the PageRank of i if k stops being his/her friend. This may be useful to identify the users/customers whose friendship/relationship should be a priority to keep. We provide formal definitions, algorithms and some experiments for this subject. Our experiments showed that the best current friend of a user is not necessarily the one who has the highest PageRank in the SN nor the one who has more friends.

  11. Evaluation of ICD-10 algorithms to identify hypopituitary patients in the Danish National Patient Registry

    DEFF Research Database (Denmark)

    Berglund, Agnethe; Olsen, Morten; Andersen, Marianne

    2017-01-01

    : Patients with International Classification of Diseases (10th edition [ICD-10]) diagnoses of hypopituitarism, or other diagnoses of pituitary disorders assumed to be associated with an increased risk of hypopituitarism, recorded in the DNPR during 2000-2012 were identified. Medical records were reviewed...... to confirm or disprove hypopituitarism. RESULTS: Hypopituitarism was confirmed in 911 patients. In a candidate population of 1,661, this yielded an overall positive predictive value (PPV) of 54.8% (95% confidence interval [CI]: 52.4-57.3). Using algorithms searching for patients recorded at least one, three...... or five times with a diagnosis of hypopituitarism (E23.0x) and/or at least once with a diagnosis of postprocedural hypopituitarism (E89.3x), PPVs gradually increased from 73.3% (95% CI: 70.6-75.8) to 83.3% (95% CI: 80.7-85.7). Completeness for the same algorithms, however, decreased from 90.8% (95% CI: 88...

  12. Identifying irregularly shaped crime hot-spots using a multiobjective evolutionary algorithm

    Science.gov (United States)

    Wu, Xiaolan; Grubesic, Tony H.

    2010-12-01

    Spatial cluster detection techniques are widely used in criminology, geography, epidemiology, and other fields. In particular, spatial scan statistics are popular and efficient techniques for detecting areas of elevated crime or disease events. The majority of spatial scan approaches attempt to delineate geographic zones by evaluating the significance of clusters using likelihood ratio statistics tested with the Poisson distribution. While this can be effective, many scan statistics give preference to circular clusters, diminishing their ability to identify elongated and/or irregular shaped clusters. Although adjusting the shape of the scan window can mitigate some of these problems, both the significance of irregular clusters and their spatial structure must be accounted for in a meaningful way. This paper utilizes a multiobjective evolutionary algorithm to find clusters with maximum significance while quantitatively tracking their geographic structure. Crime data for the city of Cincinnati are utilized to demonstrate the advantages of the new approach and highlight its benefits versus more traditional scan statistics.

  13. Detecting community structure using label propagation with consensus weight in complex network

    International Nuclear Information System (INIS)

    Liang Zong-Wen; Li Jian-Ping; Yang Fan; Petropulu Athina

    2014-01-01

    Community detection is a fundamental work to analyse the structural and functional properties of complex networks. The label propagation algorithm (LPA) is a near linear time algorithm to find a good community structure. Despite various subsequent advances, an important issue of this algorithm has not yet been properly addressed. Random update orders within the algorithm severely hamper the stability of the identified community structure. In this paper, we executed the basic label propagation algorithm on networks multiple times, to obtain a set of consensus partitions. Based on these consensus partitions, we created a consensus weighted graph. In this consensus weighted graph, the weight value of the edge was the proportion value that the number of node pairs allocated in the same cluster was divided by the total number of partitions. Then, we introduced consensus weight to indicate the direction of label propagation. In label update steps, by computing the mixing value of consensus weight and label frequency, a node adopted the label which has the maximum mixing value instead of the most frequent one. For extending to different networks, we introduced a proportion parameter to adjust the proportion of consensus weight and label frequency in computing mixing value. Finally, we proposed an approach named the label propagation algorithm with consensus weight (LPAcw), and the experimental results showed that the LPAcw could enhance considerably both the stability and the accuracy of community partitions. (interdisciplinary physics and related areas of science and technology)

  14. REVIEW OF THE GOVERNING EQUATIONS, COMPUTATIONAL ALGORITHMS, AND OTHER COMPONENTS OF THE MODELS-3 COMMUNITY MULTISCALE AIR QUALITY (CMAQ) MODELING SYSTEM

    Science.gov (United States)

    This article describes the governing equations, computational algorithms, and other components entering into the Community Multiscale Air Quality (CMAQ) modeling system. This system has been designed to approach air quality as a whole by including state-of-the-science capabiliti...

  15. Identification of overlapping communities and their hierarchy by locally calculating community-changing resolution levels

    OpenAIRE

    Havemann, Frank; Heinz, Michael; Struck, Alexander; Gläser, Jochen

    2010-01-01

    We propose a new local, deterministic and parameter-free algorithm that detects fuzzy and crisp overlapping communities in a weighted network and simultaneously reveals their hierarchy. Using a local fitness function, the algorithm greedily expands natural communities of seeds until the whole graph is covered. The hierarchy of communities is obtained analytically by calculating resolution levels at which communities grow rather than numerically by testing different resolution levels. This ana...

  16. Tobacco control recommendations identified by LGBT Atlantans in a community-based participatory research project.

    Science.gov (United States)

    Bryant, Lawrence; Damarin, Amanda K; Marshall, Zack

    2014-01-01

    Lesbian, gay, bisexual, and transgender (LGBT) people are increasingly aware that disproportionately high smoking rates severely impact the health of their communities. Motivated to make a change, a group of LGBT community members, policymakers, and researchers from Atlanta carried out a community-based participatory research (CBPR) project. This formative research study sought to identify recommendations for culturally relevant smoking prevention and cessation interventions that could improve the health of Atlanta's LGBT communities. Data presented here come from four focus groups with 36 participants and a community meeting with 30 participants. Among study participants, the most favored interventions were providing LGBT-specific cessation programs, raising awareness about LGBT smoking rates, and getting community venues to go smoke-free. Participants also suggested providing reduced-cost cessation products for low-income individuals, using LGBT "role models" to promote cessation, and ensuring that interventions reach all parts of the community. Findings reinforce insights from community-based research with other marginalized groups. Similarities include the importance of tailoring cessation programs for specific communities, the need to acknowledge differences within communities, and the significance of community spaces in shaping discussions of cessation. Further, this study highlights the need for heightened awareness. The Atlanta LGBT community is largely unaware that high smoking rates affect its health, and is unlikely to take collective action to address this problem until it is understood.

  17. Identifying Social-ecological Linkages to Develop a Community Fire Plan in Mexico

    Directory of Open Access Journals (Sweden)

    Rachel A.S Sheridan

    2015-01-01

    Full Text Available Community forestry in rural Mexico presents a unique opportunity to study the linkages and feedback within coupled social-ecological systems due to the fact that agrarian or indigenous communities control approximately half of the national territory of Mexico. We used social and ecological diagnostic tools to develop a fire management strategy for a communal forest containing an endemic piñón pine species, Pinus cembroides subs. orizabensis, in the state of Tlaxcala, Mexico. The ecological diagnostic was done through fuel inventory, forest structure sampling, and fire behaviour modelling. The social assessment was conducted through household interviews, community workshops, and direct participant observation. The ecological fire hazard was quantified and coupled with the social assessment to develop a fire management plan. Vertical fuel continuity and flashy surface fuels created a high fire hazard. Modelled fire behaviour showed a rapid rate of spread and high flame lengths under multiple scenarios. Relative impunity for starting forest fires, poor community and inter-agency organisation, and lack of project continuity across organisational sectors appear to be the most significant social limiting factors for wildfire management. Combining both social and ecological diagnostic tools provides a comprehensive understanding of the actual risks to forests, and identifies realistic community-supported options for conservation on cooperatively managed lands.

  18. Towards a community effort to identify ethical principles for research in hydrology

    Science.gov (United States)

    Montanari, Alberto

    2010-05-01

    The hydrological community in Europe is growing rapidly in both size and, more importantly, scientific relevance and integrity. The Hydrological Sciences (HS) Division of EGU actively is promoting the above development by identifying research targets, stimulating the involvement of young scientists and managing a scientific open access journal based on a public peer review process. The management of the Division itself and the organisation of the General Assembly are carried out transparently, with the aim to seek an improved involvement of top and young scientists, with a bottom up approach. I believe the HS community is animated by a strong enthusiasm which, however, is not adequately supported by economical funding. In my opinion this is a major problem which HS should consider and discuss. The relevance of the societal and environmental problems dealt with by hydrologists, in a professional way and with exceptional scientific skills, is without doubt and therefore the limited amount of funding is not justified in practice. In my opinion, in order to refine the structure of the HS community, and promote its visibility, we should formally identify HS ethical principles for research in environmental science. The principles should highlight the role of hydrology as well as the ethical and scientific solidity of the HS community. Establishing ethical principles is even more important in view of the transparent approach HS is adopting for reviewing and publishing contributions and in view of the increasing need to transparently prove how public funding for research is administered. Establishing ethical principles for hydrology is not a trivial task. Hydrology is characterised by a relevant uncertainty in data, models and parameters. Hydrology is also relying on a large variety of approaches, ranging from statistical to physically based. The purpose of this poster is to present a collection of ethical principles for scientific research presented by the literature and

  19. Identifying the community structure of the food-trade international multi-network

    Science.gov (United States)

    Torreggiani, S.; Mangioni, G.; Puma, M. J.; Fagiolo, G.

    2018-05-01

    Achieving international food security requires improved understanding of how international trade networks connect countries around the world through the import-export flows of food commodities. The properties of international food trade networks are still poorly documented, especially from a multi-network perspective. In particular, nothing is known about the multi-network’s community structure. Here we find that the individual crop-specific layers of the multi-network have densely connected trading groups, a consistent characteristic over the period 2001–2011. Further, the multi-network is characterized by low variability over this period but with substantial heterogeneity across layers in each year. In particular, the layers are mostly assortative: more-intensively connected countries tend to import from and export to countries that are themselves more connected. We also fit econometric models to identify social, economic and geographic factors explaining the probability that any two countries are co-present in the same community. Our estimates indicate that the probability of country pairs belonging to the same food trade community depends more on geopolitical and economic factors—such as geographical proximity and trade-agreement co-membership—than on country economic size and/or income. These community-structure findings of the multi-network are especially valuable for efforts to understand past and emerging dynamics in the global food system, especially those that examine potential ‘shocks’ to global food trade.

  20. Phylogenetic & Physiological Profiling of Microbial Communities of Contaminated Soils/Sediments: Identifying Microbial consortia...

    Energy Technology Data Exchange (ETDEWEB)

    Terence L. Marsh

    2004-05-26

    The goals of this study were: (1) survey the microbial community in soil samples from a site contaminated with heavy metals using new rapid molecular techniques that are culture-independent; (2) identify phylogenetic signatures of microbial populations that correlate with metal ion contamination; and (3) cultivate these diagnostic strains using traditional as well as novel cultivation techniques in order to identify organisms that may be of value in site evaluation/management or bioremediation.

  1. Identifying and intervening on barriers to healthcare access among members of a small Korean community in the southern USA.

    Science.gov (United States)

    Rhodes, Scott D; Song, Eunyoung; Nam, Sang; Choi, Sarah J; Choi, Seungyong

    2015-04-01

    We used community-based participatory research (CBPR) to explore barriers to healthcare access and utilization and identify potentially effective intervention strategies to increase access among members of the Korean community in North Carolina (NC). Our CBPR partnership conducted 8 focus groups with 63 adult Korean immigrants in northwest NC and 15 individual in-depth interviews and conducted an empowerment-based community forum. We identified 20 themes that we organized into four domains, including practical barriers to health care, negative perceptions about care, contingencies for care, and provider misconceptions about local needs. Forum attendees identified four strategies to improve Korean community health. Despite the implementation of the Patient Protection and Affordable Care Act (ACA), many Korean community members will continue to remain uninsured, and among those who obtain insurance, many barriers will remain. It is imperative to ensure the health of this highly neglected and vulnerable community. Potential strategies include the development of (1) low-literacy materials to educate members of the Korean community about how to access healthcare services, (2) lay health advisor programs to support navigation of service access and utilization, (3) church-based programming, and (4) provider education to reduce misconceptions about Korean community needs. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  2. De-identifying an EHR Database

    DEFF Research Database (Denmark)

    Lauesen, Søren; Pantazos, Kostas; Lippert, Søren

    2011-01-01

    -identified a Danish EHR database with 437,164 patients. The goal was to generate a version with real medical records, but related to artificial persons. We developed a de-identification algorithm that uses lists of named entities, simple language analysis, and special rules. Our algorithm consists of 3 steps: collect...... lists of identifiers from the database and external resources, define a replacement for each identifier, and replace identifiers in structured data and free text. Some patient records could not be safely de-identified, so the de-identified database has 323,122 patient records with an acceptable degree...... of anonymity, readability and correctness (F-measure of 95%). The algorithm has to be adjusted for each culture, language and database....

  3. [Algorithms based on medico-administrative data in the field of endocrine, nutritional and metabolic diseases, especially diabetes].

    Science.gov (United States)

    Fosse-Edorh, S; Rigou, A; Morin, S; Fezeu, L; Mandereau-Bruno, L; Fagot-Campagna, A

    2017-10-01

    Medico-administrative databases represent a very interesting source of information in the field of endocrine, nutritional and metabolic diseases. The objective of this article is to describe the early works of the Redsiam working group in this field. Algorithms developed in France in the field of diabetes, the treatment of dyslipidemia, precocious puberty, and bariatric surgery based on the National Inter-schema Information System on Health Insurance (SNIIRAM) data were identified and described. Three algorithms for identifying people with diabetes are available in France. These algorithms are based either on full insurance coverage for diabetes or on claims of diabetes treatments, or on the combination of these two methods associated with hospitalizations related to diabetes. Each of these algorithms has a different purpose, and the choice should depend on the goal of the study. Algorithms for identifying people treated for dyslipidemia or precocious puberty or who underwent bariatric surgery are also available. Early work from the Redsiam working group in the field of endocrine, nutritional and metabolic diseases produced an inventory of existing algorithms in France, linked with their goals, together with a presentation of their limitations and advantages, providing useful information for the scientific community. This work will continue with discussions about algorithms on the incidence of diabetes in children, thyroidectomy for thyroid nodules, hypothyroidism, hypoparathyroidism, and amyloidosis. Copyright © 2017 Elsevier Masson SAS. All rights reserved.

  4. Identifying deterministic signals in simulated gravitational wave data: algorithmic complexity and the surrogate data method

    International Nuclear Information System (INIS)

    Zhao Yi; Small, Michael; Coward, David; Howell, Eric; Zhao Chunnong; Ju Li; Blair, David

    2006-01-01

    We describe the application of complexity estimation and the surrogate data method to identify deterministic dynamics in simulated gravitational wave (GW) data contaminated with white and coloured noises. The surrogate method uses algorithmic complexity as a discriminating statistic to decide if noisy data contain a statistically significant level of deterministic dynamics (the GW signal). The results illustrate that the complexity method is sensitive to a small amplitude simulated GW background (SNR down to 0.08 for white noise and 0.05 for coloured noise) and is also more robust than commonly used linear methods (autocorrelation or Fourier analysis)

  5. Community occupational therapists' clinical reasoning: identifying tacit knowledge.

    Science.gov (United States)

    Carrier, Annie; Levasseur, Mélanie; Bédard, Denis; Desrosiers, Johanne

    2010-12-01

      Occupational therapy interventions in the community, a fast expanding practice setting, are central to an important social priority, the ability to live at home. These interventions generally involve only a small number of home visits, which aim at maximising the safety and autonomy of community-dwelling clients. Knowing how community occupational therapists determine their interventions, i.e. their clinical reasoning, can improve intervention efficacy. However, occupational therapists are often uninformed about and neglect the importance of clinical reasoning, which could underoptimise their interventions.   To synthesise current knowledge about community occupational therapists' clinical reasoning.   A scoping study of the literature on community occupational therapists' clinical reasoning was undertaken.   Fifteen textbooks and 25 articles, including six focussing on community occupational therapists' clinical reasoning, were reviewed. Community occupational therapists' clinical reasoning is influenced by internal and external factors. Internal factors include past experiences, expertise and perceived complexity of a problem. One of the external factors, practice context (e.g. organisational or cultural imperatives, physical location of intervention), particularly shapes community occupational therapists' clinical reasoning, which is interactive, complex and multidimensional. However, the exact influence of many factors (personal context, organisational and legal aspects of health care, lack of resources and increased number of referrals) remains unclear.   Further studies are needed to understand better the influence of internal and external factors. The extent to which these factors mould the way community occupational therapists think and act could have a direct influence on the services they provide to their clients. © 2010 The Authors. Australian Occupational Therapy Journal © 2010 Australian Association of Occupational Therapists.

  6. Using a Novel Evolutionary Algorithm to More Effectively Apply Community-Driven EcoHealth Interventions in Big Data with Application to Chagas Disease

    Science.gov (United States)

    Rizzo, D. M.; Hanley, J.; Monroy, C.; Rodas, A.; Stevens, L.; Dorn, P.

    2016-12-01

    Chagas disease is a deadly, neglected tropical disease that is endemic to every country in Central and South America. The principal insect vector of Chagas disease in Central America is Triatoma dimidiata. EcoHealth interventions are an environmentally friendly alternative that use local materials to lower household infestation, reduce the risk of infestation, and improve the quality of life. Our collaborators from La Universidad de San Carlos de Guatemala along with Ministry of Health Officials reach out to communities with high infestation and teach the community EcoHealth interventions. The process of identifying which interventions have the potential to be most effective as well as the houses that are most at risk is both expensive and time consuming. In order to better identify the risk factors associated with household infestation of T. dimidiata, a number of studies have conducted socioeconomic and entomologic surveys that contain numerous potential risk factors consisting of both nominal and ordinal data. Univariate logistic regression is one of the more popular methods for determining which risk factors are most closely associated with infestation. However, this tool has limitations, especially with the large amount and type of "Big Data" associated with our study sites (e.g., 5 villages comprise of socioeconomic, demographic, and entomologic data). The infestation of a household with T. dimidiata is a complex problem that is most likely not univariate in nature and is likely to contain higher order epistatic relationships that cannot be discovered using univariate logistic regression. Add to this, the problems raised with using p-values in traditional statistics. Also, our T. dimidiata infestation dataset is too large to exhaustively search. Therefore, we use a novel evolutionary algorithm to efficiently search for higher order interactions in surveys associated with households infested with T. dimidiata. In this study, we use our novel evolutionary

  7. Classification trees for identifying non-use of community-based long-term care services among older adults.

    Science.gov (United States)

    Penkunas, Michael James; Eom, Kirsten Yuna; Chan, Angelique Wei-Ming

    2017-10-01

    Home- and center-based long-term care (LTC) services allow older adults to remain in the community while simultaneously helping caregivers cope with the stresses associated with providing care. Despite these benefits, the uptake of community-based LTC services among older adults remains low. We analyzed data from a longitudinal study in Singapore to identify the characteristics of individuals with referrals to home-based LTC services or day rehabilitation services at the time of hospital discharge. Classification and regression tree analysis was employed to identify combinations of clinical and sociodemographic characteristics of patients and their caregivers for individuals who did not take up their referred services. Patients' level of limitation in activities of daily living (ADL) and caregivers' ethnicity and educational level were the most distinguishing characteristics for identifying older adults who failed to take up their referred home-based services. For day rehabilitation services, patients' level of ADL limitation, home size, age, and possession of a national medical savings account, as well as caregivers' education level, and gender were significant factors influencing service uptake. Identifying subgroups of patients with high rates of non-use can help clinicians target individuals who are need of community-based LTC services but unlikely to engage in formal treatment. Copyright © 2017 Elsevier B.V. All rights reserved.

  8. Tacit knowledge of public health nurses in identifying community health problems and need for new services: a case study.

    Science.gov (United States)

    Yoshioka-Maeda, Kyoko; Murashima, Sachiyo; Asahara, Kiyomi

    2006-09-01

    The purpose of this study was to explore the tacit knowledge of public health nurses in identifying community health problems and developing relevant new projects. Previous research only roughly showed those skills for creating new community health services, such as lobbying. Nine Japanese public health nurses who had created new projects in their municipalities were selected by theoretical sampling and interviewed in 2002-2003. Yin's Case Study Method, especially the multiple-case study design, was used. All 9 public health nurses used similar approaches in identifying community health problems and the need for creating new services, even though their experiences differed and the kinds of projects varied. They identified the difficulties of clients, recognized clients who had the same problems, elucidated the limitations of existing services, and forecasted outcomes from the neglect of the clients' problems. Then they succeeded in creating a new project by examining individual health problems in the context of their community's characteristics, societal factors, and using existing policies to support their clients. This is the first study to explore the skills of public health nurses and their intention to use such skills in creating new projects as well as the exact process. They could identify community health problems that will be the basis for developing new services to provide care for individual clients. This is different from the traditional community assessment approach that requires the collection of a huge amount of information to clarify community health problems. The tacit knowledge of public health nurses will help to create needs-oriented new services more smoothly.

  9. Identifying Effective Methods of Instruction for Adult Emergent Readers through Community-Based Research

    Science.gov (United States)

    Blackmer, Rachel; Hayes-Harb, Rachel

    2016-01-01

    We present a community-based research project aimed at identifying effective methods and materials for teaching English literacy skills to adult English as a second language emergent readers. We conducted a quasi-experimental study whereby we evaluated the efficacy of two approaches, one based on current practices at the English Skills Learning…

  10. Hybrid feature selection algorithm using symmetrical uncertainty and a harmony search algorithm

    Science.gov (United States)

    Salameh Shreem, Salam; Abdullah, Salwani; Nazri, Mohd Zakree Ahmad

    2016-04-01

    Microarray technology can be used as an efficient diagnostic system to recognise diseases such as tumours or to discriminate between different types of cancers in normal tissues. This technology has received increasing attention from the bioinformatics community because of its potential in designing powerful decision-making tools for cancer diagnosis. However, the presence of thousands or tens of thousands of genes affects the predictive accuracy of this technology from the perspective of classification. Thus, a key issue in microarray data is identifying or selecting the smallest possible set of genes from the input data that can achieve good predictive accuracy for classification. In this work, we propose a two-stage selection algorithm for gene selection problems in microarray data-sets called the symmetrical uncertainty filter and harmony search algorithm wrapper (SU-HSA). Experimental results show that the SU-HSA is better than HSA in isolation for all data-sets in terms of the accuracy and achieves a lower number of genes on 6 out of 10 instances. Furthermore, the comparison with state-of-the-art methods shows that our proposed approach is able to obtain 5 (out of 10) new best results in terms of the number of selected genes and competitive results in terms of the classification accuracy.

  11. The Top Ten Algorithms in Data Mining

    CERN Document Server

    Wu, Xindong

    2009-01-01

    From classification and clustering to statistical learning, association analysis, and link mining, this book covers the most important topics in data mining research. It presents the ten most influential algorithms used in the data mining community today. Each chapter provides a detailed description of the algorithm, a discussion of available software implementation, advanced topics, and exercises. With a simple data set, examples illustrate how each algorithm works and highlight the overall performance of each algorithm in a real-world application. Featuring contributions from leading researc

  12. Use of GIS to identify optimal settings for cancer prevention and control in African American communities

    Science.gov (United States)

    Alcaraz, Kassandra I.; Kreuter, Matthew W.; Bryan, Rebecca P.

    2009-01-01

    Objective Rarely have Geographic Information Systems (GIS) been used to inform community-based outreach and intervention planning. This study sought to identify community settings most likely to reach individuals from geographically localized areas. Method An observational study conducted in an urban city in Missouri during 2003–2007 placed computerized breast cancer education kiosks in seven types of community settings: beauty salons, churches, health fairs, neighborhood health centers, Laundromats, public libraries and social service agencies. We used GIS to measure distance between kiosk users’ (n=7,297) home ZIP codes and the location where they used the kiosk. Mean distances were compared across settings. Results Mean distance between individuals’ home ZIP codes and the location where they used the kiosk varied significantly (pLaundromats (2.3 miles) and public libraries (2.8 miles) and greatest among kiosk users at health fairs (7.6 miles). Conclusion Some community settings are more likely than others to reach highly localized populations. A better understanding of how and where to reach specific populations can complement the progress already being made in identifying populations at increased disease risk. PMID:19422844

  13. Network Community Detection on Metric Space

    Directory of Open Access Journals (Sweden)

    Suman Saha

    2015-08-01

    Full Text Available Community detection in a complex network is an important problem of much interest in recent years. In general, a community detection algorithm chooses an objective function and captures the communities of the network by optimizing the objective function, and then, one uses various heuristics to solve the optimization problem to extract the interesting communities for the user. In this article, we demonstrate the procedure to transform a graph into points of a metric space and develop the methods of community detection with the help of a metric defined for a pair of points. We have also studied and analyzed the community structure of the network therein. The results obtained with our approach are very competitive with most of the well-known algorithms in the literature, and this is justified over the large collection of datasets. On the other hand, it can be observed that time taken by our algorithm is quite less compared to other methods and justifies the theoretical findings.

  14. Identifying diabetes-related important protein targets with few interacting partners with the PageRank algorithm.

    Science.gov (United States)

    Grolmusz, Vince I

    2015-04-01

    Diabetes is a growing concern for the developed nations worldwide. New genomic, metagenomic and gene-technologic approaches may yield considerable results in the next several years in its early diagnosis, or in advances in therapy and management. In this work, we highlight some human proteins that may serve as new targets in the early diagnosis and therapy. With the help of a very successful mathematical tool for network analysis that formed the basis of the early successes of Google(TM), Inc., we analyse the human protein-protein interaction network gained from the IntAct database with a mathematical algorithm. The novelty of our approach is that the new protein targets suggested do not have many interacting partners (so, they are not hubs or super-hubs), so their inhibition or promotion probably will not have serious side effects. We have identified numerous possible protein targets for diabetes therapy and/or management; some of these have been well known for a long time (these validate our method), some of them appeared in the literature in the last 12 months (these show the cutting edge of the algorithm), and the remainder are still unknown to be connected with diabetes, witnessing completely new hits of the method.

  15. Do maize models capture the impacts of heat and drought stresses on yield? Using algorithm ensembles to identify successful approaches.

    Science.gov (United States)

    Jin, Zhenong; Zhuang, Qianlai; Tan, Zeli; Dukes, Jeffrey S; Zheng, Bangyou; Melillo, Jerry M

    2016-09-01

    Stresses from heat and drought are expected to increasingly suppress crop yields, but the degree to which current models can represent these effects is uncertain. Here we evaluate the algorithms that determine impacts of heat and drought stress on maize in 16 major maize models by incorporating these algorithms into a standard model, the Agricultural Production Systems sIMulator (APSIM), and running an ensemble of simulations. Although both daily mean temperature and daylight temperature are common choice of forcing heat stress algorithms, current parameterizations in most models favor the use of daylight temperature even though the algorithm was designed for daily mean temperature. Different drought algorithms (i.e., a function of soil water content, of soil water supply to demand ratio, and of actual to potential transpiration ratio) simulated considerably different patterns of water shortage over the growing season, but nonetheless predicted similar decreases in annual yield. Using the selected combination of algorithms, our simulations show that maize yield reduction was more sensitive to drought stress than to heat stress for the US Midwest since the 1980s, and this pattern will continue under future scenarios; the influence of excessive heat will become increasingly prominent by the late 21st century. Our review of algorithms in 16 crop models suggests that the impacts of heat and drought stress on plant yield can be best described by crop models that: (i) incorporate event-based descriptions of heat and drought stress, (ii) consider the effects of nighttime warming, and (iii) coordinate the interactions among multiple stresses. Our study identifies the proficiency with which different model formulations capture the impacts of heat and drought stress on maize biomass and yield production. The framework presented here can be applied to other modeled processes and used to improve yield predictions of other crops with a wide variety of crop models. © 2016 John

  16. Identify Beta-Hairpin Motifs with Quadratic Discriminant Algorithm Based on the Chemical Shifts.

    Directory of Open Access Journals (Sweden)

    Feng YongE

    Full Text Available Successful prediction of the beta-hairpin motif will be helpful for understanding the of the fold recognition. Some algorithms have been proposed for the prediction of beta-hairpin motifs. However, the parameters used by these methods were primarily based on the amino acid sequences. Here, we proposed a novel model for predicting beta-hairpin structure based on the chemical shift. Firstly, we analyzed the statistical distribution of chemical shifts of six nuclei in not beta-hairpin and beta-hairpin motifs. Secondly, we used these chemical shifts as features combined with three algorithms to predict beta-hairpin structure. Finally, we achieved the best prediction, namely sensitivity of 92%, the specificity of 94% with 0.85 of Mathew's correlation coefficient using quadratic discriminant analysis algorithm, which is clearly superior to the same method for the prediction of beta-hairpin structure from 20 amino acid compositions in the three-fold cross-validation. Our finding showed that the chemical shift is an effective parameter for beta-hairpin prediction, suggesting the quadratic discriminant analysis is a powerful algorithm for the prediction of beta-hairpin.

  17. Identifying mechanisms that structure ecological communities by snapping model parameters to empirically observed tradeoffs.

    Science.gov (United States)

    Thomas Clark, Adam; Lehman, Clarence; Tilman, David

    2018-04-01

    Theory predicts that interspecific tradeoffs are primary determinants of coexistence and community composition. Using information from empirically observed tradeoffs to augment the parametrisation of mechanism-based models should therefore improve model predictions, provided that tradeoffs and mechanisms are chosen correctly. We developed and tested such a model for 35 grassland plant species using monoculture measurements of three species characteristics related to nitrogen uptake and retention, which previous experiments indicate as important at our site. Matching classical theoretical expectations, these characteristics defined a distinct tradeoff surface, and models parameterised with these characteristics closely matched observations from experimental multi-species mixtures. Importantly, predictions improved significantly when we incorporated information from tradeoffs by 'snapping' characteristics to the nearest location on the tradeoff surface, suggesting that the tradeoffs and mechanisms we identify are important determinants of local community structure. This 'snapping' method could therefore constitute a broadly applicable test for identifying influential tradeoffs and mechanisms. © 2018 The Authors. Ecology Letters published by CNRS and John Wiley & Sons Ltd.

  18. A Clinical Algorithm to Identify HIV Patients at High Risk for Incident Active Tuberculosis: A Prospective 5-Year Cohort Study.

    Directory of Open Access Journals (Sweden)

    Susan Shin-Jung Lee

    Full Text Available Predicting the risk of tuberculosis (TB in people living with HIV (PLHIV using a single test is currently not possible. We aimed to develop and validate a clinical algorithm, using baseline CD4 cell counts, HIV viral load (pVL, and interferon-gamma release assay (IGRA, to identify PLHIV who are at high risk for incident active TB in low-to-moderate TB burden settings where highly active antiretroviral therapy (HAART is routinely provided.A prospective, 5-year, cohort study of adult PLHIV was conducted from 2006 to 2012 in two hospitals in Taiwan. HAART was initiated based on contemporary guidelines (CD4 count < = 350/μL. Cox regression was used to identify the predictors of active TB and to construct the algorithm. The validation cohorts included 1455 HIV-infected individuals from previous published studies. Area under the receiver operating characteristic (ROC curve was calculated.Seventeen of 772 participants developed active TB during a median follow-up period of 5.21 years. Baseline CD4 < 350/μL or pVL ≥ 100,000/mL was a predictor of active TB (adjusted HR 4.87, 95% CI 1.49-15.90, P = 0.009. A positive baseline IGRA predicted TB in patients with baseline CD4 ≥ 350/μL and pVL < 100,000/mL (adjusted HR 6.09, 95% CI 1.52-24.40, P = 0.01. Compared with an IGRA-alone strategy, the algorithm improved the sensitivity from 37.5% to 76.5%, the negative predictive value from 98.5% to 99.2%. Compared with an untargeted strategy, the algorithm spared 468 (60.6% from unnecessary TB preventive treatment. Area under the ROC curve was 0.692 (95% CI: 0.587-0.798 for the study cohort and 0.792 (95% CI: 0.776-0.808 and 0.766 in the 2 validation cohorts.A validated algorithm incorporating the baseline CD4 cell count, HIV viral load, and IGRA status can be used to guide targeted TB preventive treatment in PLHIV in low-to-moderate TB burden settings where HAART is routinely provided to all PLHIV. The implementation of this algorithm will avoid unnecessary

  19. Effects of multi-state links in network community detection

    International Nuclear Information System (INIS)

    Rocco, Claudio M.; Moronta, José; Ramirez-Marquez, José E.; Barker, Kash

    2017-01-01

    A community is defined as a group of nodes of a network that are densely interconnected with each other but only sparsely connected with the rest of the network. The set of communities (i.e., the network partition) and their inter-community links could be derived using special algorithms account for the topology of the network and, in certain cases, the possible weights associated to the links. In general, the set of weights represents some characteristic as capacity, flow and reliability, among others. The effects of considering weights could be translated to obtain a different partition. In many real situations, particularly when modeling infrastructure systems, networks must be modeled as multi-state networks (e.g., electric power networks). In such networks, each link is characterized by a vector of known random capacities (i.e., the weight on each link could vary according to a known probability distribution). In this paper a simple Monte Carlo approach is proposed to evaluate the effects of multi-state links on community detection as well as on the performance of the network. The approach is illustrated with the topology of an electric power system. - Highlights: • Identify network communities when considering multi-state links. • Identified how effects of considering weights translate to different partition. • Identified importance of Inter-Community Links and changes with respect to community. • Preamble to performing a resilience assessment able to mimic the evolution of the state of each community.

  20. Approximate Computing Techniques for Iterative Graph Algorithms

    Energy Technology Data Exchange (ETDEWEB)

    Panyala, Ajay R.; Subasi, Omer; Halappanavar, Mahantesh; Kalyanaraman, Anantharaman; Chavarria Miranda, Daniel G.; Krishnamoorthy, Sriram

    2017-12-18

    Approximate computing enables processing of large-scale graphs by trading off quality for performance. Approximate computing techniques have become critical not only due to the emergence of parallel architectures but also the availability of large scale datasets enabling data-driven discovery. Using two prototypical graph algorithms, PageRank and community detection, we present several approximate computing heuristics to scale the performance with minimal loss of accuracy. We present several heuristics including loop perforation, data caching, incomplete graph coloring and synchronization, and evaluate their efficiency. We demonstrate performance improvements of up to 83% for PageRank and up to 450x for community detection, with low impact of accuracy for both the algorithms. We expect the proposed approximate techniques will enable scalable graph analytics on data of importance to several applications in science and their subsequent adoption to scale similar graph algorithms.

  1. Source Water Protection Planning for Ontario First Nations Communities: Case Studies Identifying Challenges and Outcomes

    Directory of Open Access Journals (Sweden)

    Leslie Collins

    2017-07-01

    Full Text Available After the Walkerton tragedy in 2000, where drinking water contamination left seven people dead and many suffering from chronic illness, the Province of Ontario, Canada implemented policies to develop Source Water Protection (SWP plans. Under the Clean Water Act (2006, thirty-six regional Conservation Authorities were mandated to develop watershed-based SWP plans under 19 Source Protection Regions. Most First Nations in Ontario are outside of these Source Protection Regions and reserve lands are under Federal jurisdiction. This paper explores how First Nations in Ontario are attempting to address SWP to improve drinking water quality in their communities even though these communities are not part of the Ontario SWP framework. The case studies highlight the gap between the regulatory requirements of the Federal and Provincial governments and the challenges for First Nations in Ontario from lack of funding to implement solutions to address the threats identified in SWP planning. This analysis of different approaches taken by Ontario First Nations shows that the Ontario framework for SWP planning is not an option for the majority of First Nations communities, and does not adequately address threats originating on reserve lands. First Nations attempting to address on-reserve threats to drinking water are using a variety of resources and approaches to develop community SWP plans. However, a common theme of all the cases surveyed is a lack of funding to support implementing solutions for the threats identified by the SWP planning process. Federal government initiatives to address the chronic problem of boil water advisories within Indigenous communities do not recognize SWP planning as a cost-effective tool for improving drinking water quality.

  2. Identifying Health Needs in Peru Through Use of a Community Survey.

    Science.gov (United States)

    Renn, McCartney; Steffen, Lori

    2016-11-01

    Students and faculty from a Midwestern college conducted a neighborhood community needs assessment in an impoverished area of a Peruvian city to identify health needs of residents. Students interviewed residents in their homes, asking about the need for medical, dental, and ophthalmic care and screening for chronic conditions such as diabetes, heart disease, and tuberculosis. The survey provided necessary information to medical mission workers and allowed students to directly observe family living conditions while assessing psychosocial needs of the families interviewed. The challenges of this survey included differing expectations, language barriers, recruiting neighborhood volunteers, safety risks to students, and mistrust by neighborhood residents.

  3. Assessing the Performance of a Machine Learning Algorithm in Identifying Bubbles in Dust Emission

    Science.gov (United States)

    Xu, Duo; Offner, Stella S. R.

    2017-12-01

    Stellar feedback created by radiation and winds from massive stars plays a significant role in both physical and chemical evolution of molecular clouds. This energy and momentum leaves an identifiable signature (“bubbles”) that affects the dynamics and structure of the cloud. Most bubble searches are performed “by eye,” which is usually time-consuming, subjective, and difficult to calibrate. Automatic classifications based on machine learning make it possible to perform systematic, quantifiable, and repeatable searches for bubbles. We employ a previously developed machine learning algorithm, Brut, and quantitatively evaluate its performance in identifying bubbles using synthetic dust observations. We adopt magnetohydrodynamics simulations, which model stellar winds launching within turbulent molecular clouds, as an input to generate synthetic images. We use a publicly available three-dimensional dust continuum Monte Carlo radiative transfer code, HYPERION, to generate synthetic images of bubbles in three Spitzer bands (4.5, 8, and 24 μm). We designate half of our synthetic bubbles as a training set, which we use to train Brut along with citizen-science data from the Milky Way Project (MWP). We then assess Brut’s accuracy using the remaining synthetic observations. We find that Brut’s performance after retraining increases significantly, and it is able to identify yellow bubbles, which are likely associated with B-type stars. Brut continues to perform well on previously identified high-score bubbles, and over 10% of the MWP bubbles are reclassified as high-confidence bubbles, which were previously marginal or ambiguous detections in the MWP data. We also investigate the influence of the size of the training set, dust model, evolutionary stage, and background noise on bubble identification.

  4. Large Scale Community Detection Using a Small World Model

    Directory of Open Access Journals (Sweden)

    Ranjan Kumar Behera

    2017-11-01

    Full Text Available In a social network, small or large communities within the network play a major role in deciding the functionalities of the network. Despite of diverse definitions, communities in the network may be defined as the group of nodes that are more densely connected as compared to nodes outside the group. Revealing such hidden communities is one of the challenging research problems. A real world social network follows small world phenomena, which indicates that any two social entities can be reachable in a small number of steps. In this paper, nodes are mapped into communities based on the random walk in the network. However, uncovering communities in large-scale networks is a challenging task due to its unprecedented growth in the size of social networks. A good number of community detection algorithms based on random walk exist in literature. In addition, when large-scale social networks are being considered, these algorithms are observed to take considerably longer time. In this work, with an objective to improve the efficiency of algorithms, parallel programming framework like Map-Reduce has been considered for uncovering the hidden communities in social network. The proposed approach has been compared with some standard existing community detection algorithms for both synthetic and real-world datasets in order to examine its performance, and it is observed that the proposed algorithm is more efficient than the existing ones.

  5. Finding Community Structures In Social Activity Data

    KAUST Repository

    Peng, Chengbin

    2015-05-19

    Social activity data sets are increasing in number and volume. Finding community structure in such data is valuable in many applications. For example, understand- ing the community structure of social networks may reduce the spread of epidemics or boost advertising revenue; discovering partitions in tra c networks can help to optimize routing and to reduce congestion; finding a group of users with common interests can allow a system to recommend useful items. Among many aspects, qual- ity of inference and e ciency in finding community structures in such data sets are of paramount concern. In this thesis, we propose several approaches to improve com- munity detection in these aspects. The first approach utilizes the concept of K-cores to reduce the size of the problem. The K-core of a graph is the largest subgraph within which each node has at least K connections. We propose a framework that accelerates community detection. It first applies a traditional algorithm that is relatively slow to the K-core, and then uses a fast heuristic to infer community labels for the remaining nodes. The second approach is to scale the algorithm to multi-processor systems. We de- vise a scalable community detection algorithm for large networks based on stochastic block models. It is an alternating iterative algorithm using a maximum likelihood ap- proach. Compared with traditional inference algorithms for stochastic block models, our algorithm can scale to large networks and run on multi-processor systems. The time complexity is linear in the number of edges of the input network. The third approach is to improve the quality. We propose a framework for non- negative matrix factorization that allows the imposition of linear or approximately linear constraints on each factor. An example of the applications is to find community structures in bipartite networks, which is useful in recommender systems. Our algorithms are compared with the results in recent papers and their quality and e

  6. Ensemble method: Community detection based on game theory

    Science.gov (United States)

    Zhang, Xia; Xia, Zhengyou; Xu, Shengwu; Wang, J. D.

    2014-08-01

    Timely and cost-effective analytics over social network has emerged as a key ingredient for success in many businesses and government endeavors. Community detection is an active research area of relevance to analyze online social network. The problem of selecting a particular community detection algorithm is crucial if the aim is to unveil the community structure of a network. The choice of a given methodology could affect the outcome of the experiments because different algorithms have different advantages and depend on tuning specific parameters. In this paper, we propose a community division model based on the notion of game theory, which can combine advantages of previous algorithms effectively to get a better community classification result. By making experiments on some standard dataset, it verifies that our community detection model based on game theory is valid and better.

  7. A new cluster algorithm for graphs

    NARCIS (Netherlands)

    S. van Dongen

    1998-01-01

    textabstractA new cluster algorithm for graphs called the emph{Markov Cluster algorithm ($MCL$ algorithm) is introduced. The graphs may be both weighted (with nonnegative weight) and directed. Let~$G$~be such a graph. The $MCL$ algorithm simulates flow in $G$ by first identifying $G$ in a

  8. Community Detection for Correlation Matrices

    Directory of Open Access Journals (Sweden)

    Mel MacMahon

    2015-04-01

    Full Text Available A challenging problem in the study of complex systems is that of resolving, without prior information, the emergent, mesoscopic organization determined by groups of units whose dynamical activity is more strongly correlated internally than with the rest of the system. The existing techniques to filter correlations are not explicitly oriented towards identifying such modules and can suffer from an unavoidable information loss. A promising alternative is that of employing community detection techniques developed in network theory. Unfortunately, this approach has focused predominantly on replacing network data with correlation matrices, a procedure that we show to be intrinsically biased because of its inconsistency with the null hypotheses underlying the existing algorithms. Here, we introduce, via a consistent redefinition of null models based on random matrix theory, the appropriate correlation-based counterparts of the most popular community detection techniques. Our methods can filter out both unit-specific noise and system-wide dependencies, and the resulting communities are internally correlated and mutually anticorrelated. We also implement multiresolution and multifrequency approaches revealing hierarchically nested subcommunities with “hard” cores and “soft” peripheries. We apply our techniques to several financial time series and identify mesoscopic groups of stocks which are irreducible to a standard, sectorial taxonomy; detect “soft stocks” that alternate between communities; and discuss implications for portfolio optimization and risk management.

  9. Community Detection for Correlation Matrices

    Science.gov (United States)

    MacMahon, Mel; Garlaschelli, Diego

    2015-04-01

    A challenging problem in the study of complex systems is that of resolving, without prior information, the emergent, mesoscopic organization determined by groups of units whose dynamical activity is more strongly correlated internally than with the rest of the system. The existing techniques to filter correlations are not explicitly oriented towards identifying such modules and can suffer from an unavoidable information loss. A promising alternative is that of employing community detection techniques developed in network theory. Unfortunately, this approach has focused predominantly on replacing network data with correlation matrices, a procedure that we show to be intrinsically biased because of its inconsistency with the null hypotheses underlying the existing algorithms. Here, we introduce, via a consistent redefinition of null models based on random matrix theory, the appropriate correlation-based counterparts of the most popular community detection techniques. Our methods can filter out both unit-specific noise and system-wide dependencies, and the resulting communities are internally correlated and mutually anticorrelated. We also implement multiresolution and multifrequency approaches revealing hierarchically nested subcommunities with "hard" cores and "soft" peripheries. We apply our techniques to several financial time series and identify mesoscopic groups of stocks which are irreducible to a standard, sectorial taxonomy; detect "soft stocks" that alternate between communities; and discuss implications for portfolio optimization and risk management.

  10. Computational geometry algorithms and applications

    CERN Document Server

    de Berg, Mark; Overmars, Mark; Schwarzkopf, Otfried

    1997-01-01

    Computational geometry emerged from the field of algorithms design and anal­ ysis in the late 1970s. It has grown into a recognized discipline with its own journals, conferences, and a large community of active researchers. The suc­ cess of the field as a research discipline can on the one hand be explained from the beauty of the problems studied and the solutions obtained, and, on the other hand, by the many application domains--computer graphics, geographic in­ formation systems (GIS), robotics, and others-in which geometric algorithms play a fundamental role. For many geometric problems the early algorithmic solutions were either slow or difficult to understand and implement. In recent years a number of new algorithmic techniques have been developed that improved and simplified many of the previous approaches. In this textbook we have tried to make these modem algorithmic solutions accessible to a large audience. The book has been written as a textbook for a course in computational geometry, but it can ...

  11. Community-Engaged Research to Identify House Parent Perspectives on Support and Risk within the House and Ball Scene

    Science.gov (United States)

    Kubicek, Katrina; Beyer, William H.; McNeeley, Miles; Weiss, George; Omni, Legendary Father Taz Ultra; Kipke, Michele D.

    2012-01-01

    This paper describes a community-engaged study with the Los Angeles House and Ball scene, in which the perspectives of the leaders of these communities are captured to better understand how the House and Ball communities may protect and/or increase its members’ risks for HIV infection. Data were collected through in-depth interviews with House parents (N=26). This study identified key features of both support (e.g., family and support; acceptance; validation and recognition) and risk (e.g., members’ struggle to maintain status in the Ballroom scene; sex work; substance use; danger of becoming too involved in the Ball community; perception and stigma of Ballroom scene within the larger gay community) within these communities. Findings are discussed in relation to framing how to leverage the supportive aspects of the House and Ball communities to design relevant HIV prevention interventions. PMID:22206442

  12. A cooperative game framework for detecting overlapping communities in social networks

    Science.gov (United States)

    Jonnalagadda, Annapurna; Kuppusamy, Lakshmanan

    2018-02-01

    Community detection in social networks is a challenging and complex task, which received much attention from researchers of multiple domains in recent years. The evolution of communities in social networks happens merely due to the self-interest of the nodes. The interesting feature of community structure in social networks is the multi membership of the nodes resulting in overlapping communities. Assuming the nodes of the social network as self-interested players, the dynamics of community formation can be captured in the form of a game. In this paper, we propose a greedy algorithm, namely, Weighted Graph Community Game (WGCG), in order to model the interactions among the self-interested nodes of the social network. The proposed algorithm employs the Shapley value mechanism to discover the inherent communities of the underlying social network. The experimental evaluation on the real-world and synthetic benchmark networks demonstrates that the performance of the proposed algorithm is superior to the state-of-the-art overlapping community detection algorithms.

  13. Community landscapes: an integrative approach to determine overlapping network module hierarchy, identify key nodes and predict network dynamics.

    Directory of Open Access Journals (Sweden)

    István A Kovács

    Full Text Available BACKGROUND: Network communities help the functional organization and evolution of complex networks. However, the development of a method, which is both fast and accurate, provides modular overlaps and partitions of a heterogeneous network, has proven to be rather difficult. METHODOLOGY/PRINCIPAL FINDINGS: Here we introduce the novel concept of ModuLand, an integrative method family determining overlapping network modules as hills of an influence function-based, centrality-type community landscape, and including several widely used modularization methods as special cases. As various adaptations of the method family, we developed several algorithms, which provide an efficient analysis of weighted and directed networks, and (1 determine persvasively overlapping modules with high resolution; (2 uncover a detailed hierarchical network structure allowing an efficient, zoom-in analysis of large networks; (3 allow the determination of key network nodes and (4 help to predict network dynamics. CONCLUSIONS/SIGNIFICANCE: The concept opens a wide range of possibilities to develop new approaches and applications including network routing, classification, comparison and prediction.

  14. Overlapping communities detection based on spectral analysis of line graphs

    Science.gov (United States)

    Gui, Chun; Zhang, Ruisheng; Hu, Rongjing; Huang, Guoming; Wei, Jiaxuan

    2018-05-01

    Community in networks are often overlapping where one vertex belongs to several clusters. Meanwhile, many networks show hierarchical structure such that community is recursively grouped into hierarchical organization. In order to obtain overlapping communities from a global hierarchy of vertices, a new algorithm (named SAoLG) is proposed to build the hierarchical organization along with detecting the overlap of community structure. SAoLG applies the spectral analysis into line graphs to unify the overlap and hierarchical structure of the communities. In order to avoid the limitation of absolute distance such as Euclidean distance, SAoLG employs Angular distance to compute the similarity between vertices. Furthermore, we make a micro-improvement partition density to evaluate the quality of community structure and use it to obtain the more reasonable and sensible community numbers. The proposed SAoLG algorithm achieves a balance between overlap and hierarchy by applying spectral analysis to edge community detection. The experimental results on one standard network and six real-world networks show that the SAoLG algorithm achieves higher modularity and reasonable community number values than those generated by Ahn's algorithm, the classical CPM and GN ones.

  15. A Performance Evaluation of Lightning-NO Algorithms in CMAQ

    Science.gov (United States)

    In the Community Multiscale Air Quality (CMAQv5.2) model, we have implemented two algorithms for lightning NO production; one algorithm is based on the hourly observed cloud-to-ground lightning strike data from National Lightning Detection Network (NLDN) to replace the previous m...

  16. Identifying Natural Alignments Between Ambulatory Surgery Centers and Local Health Systems: Building Broader Communities of Surgical Care.

    Science.gov (United States)

    Funk, Russell J; Owen-Smith, Jason; Landon, Bruce E; Birkmeyer, John D; Hollingsworth, John M

    2017-02-01

    To develop and compare methods for identifying natural alignments between ambulatory surgery centers (ASCs) and hospitals that anchor local health systems. Using all-payer data from Florida's State Ambulatory Surgery and Inpatient Databases (2005-2009), we developed 3 methods for identifying alignments between ASCS and hospitals. The first, a geographic proximity approach, used spatial data to assign an ASC to its nearest hospital neighbor. The second, a predominant affiliation approach, assigned an ASC to the hospital with which it shared a plurality of surgeons. The third, a network community approach, linked an ASC with a larger group of hospitals held together by naturally occurring physician networks. We compared each method in terms of its ability to capture meaningful and stable affiliations and its administrative simplicity. Although the proximity approach was simplest to implement and produced the most durable alignments, ASC surgeon's loyalty to the assigned hospital was low with this method. The predominant affiliation and network community approaches performed better and nearly equivalently on these metrics, capturing more meaningful affiliations between ASCs and hospitals. However, the latter's alignments were least durable, and it was complex to administer. We describe 3 methods for identifying natural alignments between ASCs and hospitals, each with strengths and weaknesses. These methods will help health system managers identify ASCs with which to partner. Moreover, health services researchers and policy analysts can use them to study broader communities of surgical care.

  17. A new algorithm for identifying the flavour of B0 s mesons at LHCb

    NARCIS (Netherlands)

    Aaij, R.; Abellán Beteta, C.; Adeva, B.; Adinolfi, M.; Affolder, A.; Ajaltouni, Z.; Akar, S.; Albrecht, J.; Alessio, F.; Alexander, M.; Ali, S.; Alkhazov, G.; Alvarez Cartelle, P.; Alves, A. A.; Amato, S.; Amerio, S.; Amhis, Y.; Everse, LA; Anderlini, L.; Andreassi, G.; Andreotti, M.; Andrews, J.E.; Appleby, R. B.; Aquines Gutierrez, O.; Archilli, F.; d'Argent, P.; Artamonov, A.; Artuso, M.; Aslanides, E.; Auriemma, G.; Baalouch, M.; Bachmann, S.; Back, J. J.; Badalov, A.; Baesso, C.; Baldini, W.; Barlow, R. J.; Barschel, C.; Barsuk, S.; Barter, W.; Batozskaya, V.; Battista, V.; Bay, A.; Beaucourt, L.; Beddow, J.; Bedeschi, F.; Bediaga, I.; Bel, L. J.; Bellee, V.; Belloli, N.; Belyaev, I.; Ben-Haim, E.; Bencivenni, G.; Benson, S.; Benton, J.; Berezhnoy, A.; Bernet, R.; Bertolin, A.; Betti, F.; Bettler, M-O.; Van Beuzekom, Martin; Bifani, S.; Billoir, P.; Bird, T.D.; Birnkraut, A.; Bizzeti, A.; Blake, T.; Blanc, F.; Blouw, J.; Blusk, S.; Bocci, V.; Bondar, A.; Bondar, N.; Bonivento, W.; Borgheresi, A.; Borghi, S.; Borisyak, M.; Borsato, M.; Bowcock, T. J. V.; Bowen, E.; Bozzi, C.; Braun, S.; Britsch, M.; Britton, T.; Brodzicka, J.; Brook, N. H.; Buchanan, E.; Burr, C.; Bursche, A.; Buytaert, J.; Cadeddu, S.; Calabrese, R.; Calvi, M.; Calvo Gomez, M.; Campana, P.; Campora Perez, D.; Capriotti, L.; Carbone, A.; Carboni, G.; Cardinale, R.; Cardini, A.; Carniti, P.; Carson, L.; Carvalho Akiba, K.; Casse, G.; Cassina, L.; Castillo Garcia, L.; Cattaneo, M.; Cauet, Ch; Cavallero, G.; Cenci, R.; Charles, M.; Charpentier, Ph; Chatzikonstantinidis, G.; Chefdeville, M.; Chen, S.; Cheung, S-F.; Chiapolini, N.; Chrzaszcz, M.; Cid Vidal, X.; Ciezarek, G.; Clarke, P. E. L.; Clemencic, M.; Cliff, H. V.; Closier, J.; Coco, V.; Cogan, J.; Cogneras, E.; Cogoni, V.; Cojocariu, L.; Collazuol, G.; Collins, P.; Comerma-Montells, A.; Contu, A.; Cook, A.; Coombes, M.; Coquereau, S.; Corti, G.; Corvo, M.; Couturier, B.; Cowan, G. A.; Craik, D. C.; Crocombe, A.; Cruz Torres, M.; Cunliffe, S.; Currie, C.R.; D'Ambrosio, C.; Dall'Occo, E.; Dalseno, J.; David, P. N.Y.; Davis, A.; De Aguiar Francisco, O.; De Bruyn, K.; De Capua, S.; De Cian, M.; de Miranda, J. M.; Paula, L.E.; De Simone, P.; Dean, C-T.; Decamp, D.; Deckenhoff, M.; Del Buono, L.; Déléage, N.; Demmer, M.; Derkach, D.; Deschamps, O.; Dettori, F.; Dey, B.; Di Canto, A.; Di Ruscio, F.; Dijkstra, H.; Donleavy, S.; Dordei, F.; Dorigo, M.; Dosil Suárez, A.; Dovbnya, A.; Dreimanis, K.; Dufour, L.; Dujany, G.; Dungs, K.; Durante, P.; Dzhelyadin, R.; Dziurda, A.; Dzyuba, A.; Easo, S.; Egede, U.; Egorychev, V.; Eidelman, S.; Eisenhardt, S.; Eitschberger, U.; Ekelhof, R.; Eklund, L.; El Rifai, I.; Elsasser, Ch.; Ely, S.; Esen, S.; Evans, H. M.; Evans, T. M.; Falabella, A.; Färber, C.; Farley, N.; Farry, S.; Fay, R.; Fazzini, D.; Ferguson, D.; Fernandez Albor, V.; Ferrari, F.; Ferreira Rodrigues, F.; Ferro-Luzzi, M.; Filippov, S.; Fiore, M.; Fiorini, M.; Firlej, M.; Fitzpatrick, C.; Fiutowski, T.; Fleuret, F.; Fohl, K.; Fol, P.; Fontana, Mark; Fontanelli, F.; Forshaw, D. C.; Forty, R.; Frank, M.; Frei, C.; Frosini, M.; Fu, J.; Furfaro, E.; Gallas Torreira, A.; Galli, D.; Gallorini, S.; Gambetta, S.; Gandelman, M.; Gandini, P.; Gao, Y.; García Pardiñas, J.; Garra Tico, J.; Garrido, L.; Gascon, D.; Carvalho-Gaspar, M.; Gavardi, L.; Gazzoni, G.; Gerick, D.; Gersabeck, E.; Gersabeck, M.; Gershon, T. J.; Ghez, Ph; Gian, S.; Gibson, V.; Girard, O. G.; Giubega, L.; Gligorov, V. V.; Göbel, C.; Golubkov, D.; Golutvin, A.; Gomes, A.Q.; Gotti, C.; Grabalosa Gándara, M.; Graciani Diaz, R.; Granado Cardoso, L. A.; Graugés, E.; Graverini, E.; Graziani, G.; Grecu, A.; Griffith, P.; Grillo, L.; Grönberg, O.; Gui, B.; Gushchin, E.; Guz, Yu; Gys, T.; Hadavizadeh, T.; Hadjivasiliou, C.; Haefeli, G.; Haen, C.; Haines, S. C.; Hall, S.; Hamilton, B.; Han, X.; Hansmann-Menzemer, S.; Harnew, N.; Harnew, S. T.; Harrison, J.; He, J.; Head, T.; Heijne, V.; Heister, A.J.G.A.M.; Hennessy, K.; Henrard, P.; Henry, L.; Hernando Morata, J. A.; van Herwijnen, E.; Heß, M.; Hicheur, A.; Hill, D.; Hoballah, M.; Hombach, C.; Hulsbergen, W.; Humair, T.; Hushchyn, M.; Hussain, N.; Hutchcroft, D. E.; Hynds, D.; Idzik, M.; Ilten, P.; Jacobsson, R.; Jaeger, A.; Jalocha, J.; Jans, E.; Jawahery, A.; John, M.; Johnson, D.; Jones, C. R.; Joram, C.; Jost, B.; Jurik, N.; Kandybei, S.; Kanso, W.; Karacson, M.; Karbach, T. M.; Karodia, S.; Kecke, M.; Kelsey, M. H.; Kenyon, I. R.; Kenzie, M.; Ketel, T.; Khairullin, E.; Khanji, B.; Khurewathanakul, C.; Kirn, T.; Klaver, S.M.; Klimaszewski, K.; Kochebina, O.; Kolpin, M.; Komarov, I.; Koopman, R. F.; Koppenburg, P.; Kozeiha, M.; Kravchuk, L.; Kreplin, K.; Kreps, M.; Krocker, G.; Krokovny, P.; Kruse, F.; Krzemien, W.; Kucewicz, W.; Kucharczyk, M.; Kudryavtsev, V.; Kuonen, A. K.; Kurek, K.; Kvaratskheliya, T.; Lacarrere, D.; Lafferty, G. D.; Lai, A.; Lambert, D.M.; Lanfranchi, G.; Langenbruch, C.; Langhans, B.; Latham, T. E.; Lazzeroni, C.; Le Gac, R.; Van Leerdam, J.; Lees, J. P.; Lefèvre, R.; Leflat, A.; Lefrançois, J.; Lemos Cid, E.; Leroy, O.; Lesiak, T.; Leverington, B.; Li, Y.; Likhomanenko, T.; Liles, M.; Lindner, R.; Linn, S.C.; Lionetto, F.; Liu, B.; Liu, X.; Loh, D.; Longstaff, I.; Lopes, J. H.; Lucchesi, D.; Lucio Martinez, M.; Luo, H.; Lupato, A.; Luppi, E.; Lupton, O.; Lusardi, N.; Lusiani, A.; Machefert, F.; Maciuc, F.; Maev, O.; Maguire, K.; Malde, S.; Malinin, A.; Manca, G.; Mancinelli, G.; Manning, P.; Mapelli, A.; Maratas, J.; Marchand, J. F.; Marconi, U.; Marin Benito, C.; Marino, P.; Marks, J.; Martellotti, G.; Martin, M.; Martinelli-Boneschi, F.; Martinez-Santos, D.; Martinez-Vidal, F.; Martins Tostes, D.; Massacrier, L. M.; Massafferri, A.; Matev, R.; Mathad, A.; Mathe, Z.; Matteuzzi, C.; Mauri, A.; Maurin, B.; Mazurov, A.; McCann, M.; McCarthy, J.; Mcnab, A.; McNulty, R.; Meadows, B. T.; Meier, F.; Meissner, M.; Melnychuk, D.; Merk, M.; Merli, A.; Michielin, E.; Milanes, D. A.; Minard, M. N.; Mitzel, D. S.; Molina Rodriguez, J.; Monroy, I. A.; Monteil, S.; Morandin, M.; Morawski, P.; Mordà, A.; Morello, M. J.; Moron, J.; Morris, A. B.; Mountain, R.; Muheim, F.; Möller, D.; Möller, J.; Möller, K.; Möller, V.; Mussini, M.; Muster, B.; Naik, P.; Nakada, T.; Nandakumar, R.; Nandi, A.; Nasteva, I.; Needham, M.; Neri, N.; Neubert, S.; Neufeld, N.; Neuner, M.; Nguyen, A. D.; Nguyen-Mau, C.; Niess, V.; Nieswand, S.; Niet, R.; Nikitin, N.; Nikodem, T.; Novoselov, A.; O'Hanlon, D. P.; Oblakowska-Mucha, A.; Obraztsov, V.; Ogilvy, S.; Okhrimenko, O.; Oldeman, R.; Onderwater, C. J.G.; Osorio Rodrigues, B.; Otalora Goicochea, J. M.; Otto, E.A.; Owen, R.P.; Oyanguren, A.; Palano, A.; Palombo, F.; Palutan, M.; Panman, J.; Papanestis, A.; Pappagallo, M.; Pappalardo, L.L.; Pappenheimer, C.; Parker, W.S; Parkes, C.; Passaleva, G.; Patel, G. D.; Patel, M.; Patrignani, C.; Pearce, D.A.; Pellegrino, A.; Penso, G.; Pepe Altarelli, M.; Perazzini, S.; Perret, P.; Pescatore, L.; Petridis, K.; Petrolini, A.; Petruzzo, M.; Picatoste Olloqui, E.; Pietrzyk, B.; Pikies, M.; Pinci, D.; Pistone, A.; Piucci, A.; Playfer, S.; Plo Casasus, M.; Poikela, T.; Polci, F.; Poluektov, A.; Polyakov, I.; Polycarpo, E.; Popov, A.; Popov, D.; Popovici, B.; Potterat, C.; Price, M. E.; Price, J.D.; Prisciandaro, J.; Pritchard, C.A.; Prouve, C.; Pugatch, V.; Puig Navarro, A.; Punzi, G.; Qian, Y.W.; Quagliani, R.; Rachwal, B.; Rademacker, J. H.; Rama, M.; Ramos Pernas, M.; Rangel, M. S.; Raniuk, I.; Raven, G.; Redi, F.; Reichert, S.; dos Reis, A. C.; Renaudin, V.; Ricciardi, S.; Richards, Jennifer S; Rihl, M.; Rinnert, K.; Rives Molina, V.; Robbe, P.; Rodrigues, A. B.; Rodrigues, L.E.T.; Rodriguez Lopez, J. A.; Rodriguez Perez, P.; Rogozhnikov, A.; Roiser, S.; Romanovsky, V.; Romero Vidal, A.; Ronayne, J. W.; Rotondo, M.; Ruf, T.; Ruiz Valls, P.; Saborido Silva, J. J.; Sagidova, N.; Saitta, B.; Salustino Guimaraes, V.; Sanchez Mayordomo, C.; Sanmartin Sedes, B.; Santacesaria, R.; Santamarina Rios, C.; Santimaria, M.; Santovetti, E.; Sarti, A.; Satriano, C.; Satta, A.; Saunders, D. M.; Savrina, D.; Schael, S.; Schiller, M.; Schindler, R. H.; Schlupp, M.; Schmelling, M.; Schmelzer, T.; Schmidt, B.; Schneider, O.; Schopper, A.; Schubiger, M.; Schune, M. H.; Schwemmer, R.; Sciascia, B.; Sciubba, A.; Semennikov, A.; Serra, N.; Serrano, J.; Sestini, L.; Seyfert, P.; Shapkin, M.; Shapoval, I.; Shcheglov, Y.; Shears, T.; Shekhtman, L.; Shevchenko, V.; Shires, A.; Siddi, B. G.; Silva Coutinho, R.; Silva de Oliveira, L.; Simi, G.; Sirendi, M.; Skidmore, N.; Skwarnicki, T.; Smith, E.; Smith, I. T.; Smith, J; Smith, M.; Snoek, H.; Sokoloff, M. D.; Soler, F. J. P.; Soomro, F.; de Souza, D.K.; Souza De Paula, B.; Spaan, B.; Spradlin, P.; Sridharan, S.; Stagni, F.; Stahl, M.; Stahl, S.; Stefkova, S.; Steinkamp, O.; Stenyakin, O.; Stevenson-Moore, P.; Stoica, S.; Stone, S.; Storaci, B.; Stracka, S.; Straticiuc, M.; Straumann, U.; Sun, L.; Sutcliffe, W.; Swientek, K.; Swientek, S.; Syropoulos, V.; Szczekowski, M.; Szumlak, T.; T'Jampens, S.; Tayduganov, A.; Tekampe, T.; Tellarini, G.; Teubert, F.; Thomas, C.; Thomas, E.; Van Tilburg, J.; Tisserand, V.; Tobin, M. N.; Todd, Jim; Tolk, S.; Tomassetti, L.; Tonelli, D.; Topp-Joergensen, S.; Tournefier, E.; Tourneur, S.; Trabelsi, K.; Traill, M.; Tran, N.T.M.T.; Tresch, M.; Trisovic, A.; Tsaregorodtsev, A.; Tsopelas, P.; Tuning, N.; Ukleja, A.; Ustyuzhanin, A.; Uwer, U.; Vacca, C.; Vagnoni, V.; Valenti, G.; Vallier, A.; Vazquez Gomez, R.; Vazquez Regueiro, P.; Vázquez Sierra, C.; Vecchi, S.; van Veghel-Plandsoen, M.M.; Velthuis, M.J.; Veltri, M.; Veneziano, G.; Vesterinen, M.; Viaud, B.; Vieira, D.; Vieites Diaz, M.; Vilasis-Cardona, X.; Volkov, V.; Vollhardt, A.; Voong, D.; Vorobyev, A.; Vorobyev, V.; Voß, C.; De Vries, J. A.; Waldi, R.; Wallace, C.; Wallace, R.; Walsh, John; Wang, J.; Ward, D. R.; Watson, N. K.; Websdale, D.; Weiden, A.; Whitehead, M.; Wicht, J.; Wilkinson, G.; Wilkinson, M.; Williams, M.; Williams, M.P.; Williams, M.; Williams, T.; Wilson, James F; Wimberley, J.; Wishahi, J.; Wislicki, W.; Witek, M.; Wormser, G.; Wotton, S. A.; Wraight, K.; Wright, S.J.; Wyllie, K.; Xie, Y.; Xu, Z.; Yang, Z.; Yu, J.; Yuan, X.; Yushchenko, O.; Zangoli, M.; Zavertyaev, M.; Zhang, L.; Zhang, Y.; Zhelezov, A.; Zhokhov, A.; Zhong, L.; Zhukov, V.; Zucchelli, S.

    2016-01-01

    A new algorithm for the determination of the initial flavour of B0 s mesons is presented. The algorithm is based on two neural networks and exploits the b hadron production mechanism at a hadron collider. The first network is trained to select charged kaons produced in association with the B0 s

  18. Explaining algorithms using metaphors

    CERN Document Server

    Forišek, Michal

    2013-01-01

    There is a significant difference between designing a new algorithm, proving its correctness, and teaching it to an audience. When teaching algorithms, the teacher's main goal should be to convey the underlying ideas and to help the students form correct mental models related to the algorithm. This process can often be facilitated by using suitable metaphors. This work provides a set of novel metaphors identified and developed as suitable tools for teaching many of the 'classic textbook' algorithms taught in undergraduate courses worldwide. Each chapter provides exercises and didactic notes fo

  19. 76 FR 11433 - Federal Transition To Secure Hash Algorithm (SHA)-256

    Science.gov (United States)

    2011-03-02

    ... ADMINISTRATION [FAR-N-2011-01; Docket No. 2011-0083; Sequence 1] Federal Transition To Secure Hash Algorithm (SHA... acquisition community to transition to Secure Hash Algorithm SHA-256. SHA-256 is a cryptographic hash function... persons attending. Please cite ``Federal Transition to Secure Hash Algorithm SHA-256'' in all...

  20. Identifying individual- and population-level characteristics that influence rates of risky alcohol consumption in regional communities.

    Science.gov (United States)

    Breen, Courtney; Shakeshaft, Anthony; Sanson-Fisher, Rob; D'Este, Catherine; Mattick, Richard P; Gilmour, Stuart

    2014-02-01

    To examine the extent to which individual- and community- level characteristics account for differences in risky alcohol consumption. A cross-sectional postal survey of 2,977 randomly selected individuals from 20 regional communities in NSW, Australia. Individuals drinking at harmful levels on the AUDIT and for risk of harm in the short term and long-term were identified. Multi-level modelling of the correlates of risky alcohol consumption at the individual and community level was conducted. There were differences between communities in alcohol consumption patterns. Being male, unmarried and reporting worse health were significant individual-level correlates for drinking at levels for risk of harm in the long term. The number of GPs (+) and police (-) were significant community characteristics. Being younger (≤25), unmarried, Australian born and with a larger income was associated with drinking at levels for risk of harm in the short term and harmful drinking on the AUDIT. The number of hotels and clubs was positively associated with drinking at levels for risk of harm in the short term. Rates of risky drinking vary significantly between communities and both individual and community characteristics are significantly associated with risky alcohol consumption. A combination of individual- and population-level interventions, tailored to the risk profile of individual communities, is most likely to be optimally effective. © 2014 The Authors. ANZJPH © 2014 Public Health Association of Australia.

  1. Distributed detection of communities in complex networks using synthetic coordinates

    International Nuclear Information System (INIS)

    Papadakis, H; Fragopoulou, P; Panagiotakis, C

    2014-01-01

    Various applications like finding Web communities, detecting the structure of social networks, and even analyzing a graph’s structure to uncover Internet attacks are just some of the applications for which community detection is important. In this paper, we propose an algorithm that finds the entire community structure of a network, on the basis of local interactions between neighboring nodes and an unsupervised distributed hierarchical clustering algorithm. The novelty of the proposed approach, named SCCD (standing for synthetic coordinate community detection), lies in the fact that the algorithm is based on the use of Vivaldi synthetic network coordinates computed by a distributed algorithm. The current paper not only presents an efficient distributed community finding algorithm, but also demonstrates that synthetic network coordinates could be used to derive efficient solutions to a variety of problems. Experimental results and comparisons with other methods from the literature are presented for a variety of benchmark graphs with known community structure, derived from varying a number of graph parameters and real data set graphs. The experimental results and comparisons to existing methods with similar computation cost on real and synthetic data sets demonstrate the high performance and robustness of the proposed scheme. (paper)

  2. Generalization of Risch's algorithm to special functions

    International Nuclear Information System (INIS)

    Raab, Clemens G.

    2013-05-01

    Symbolic integration deals with the evaluation of integrals in closed form. We present an overview of Risch's algorithm including recent developments. The algorithms discussed are suited for both indefinite and definite integration. They can also be used to compute linear relations among integrals and to find identities for special functions given by parameter integrals. The aim of this presentation is twofold: to introduce the reader to some basic ideas of differential algebra in the context of integration and to raise awareness in the physics community of computer algebra algorithms for indefinite and definite integration.

  3. The health informatics cohort enhancement project (HICE: using routinely collected primary care data to identify people with a lifetime diagnosis of psychotic disorder

    Directory of Open Access Journals (Sweden)

    Economou Alexis

    2012-02-01

    Full Text Available Abstract Background We have previously demonstrated that routinely collected primary care data can be used to identify potential participants for trials in depression [1]. Here we demonstrate how patients with psychotic disorders can be identified from primary care records for potential inclusion in a cohort study. We discuss the strengths and limitations of this approach; assess its potential value and report challenges encountered. Methods We designed an algorithm with which we searched for patients with a lifetime diagnosis of psychotic disorders within the Secure Anonymised Information Linkage (SAIL database of routinely collected health data. The algorithm was validated against the "gold standard" of a well established operational criteria checklist for psychotic and affective illness (OPCRIT. Case notes of 100 patients from a community mental health team (CMHT in Swansea were studied of whom 80 had matched GP records. Results The algorithm had favourable test characteristics, with a very good ability to detect patients with psychotic disorders (sensitivity > 0.7 and an excellent ability not to falsely identify patients with psychotic disorders (specificity > 0.9. Conclusions With certain limitations our algorithm can be used to search the general practice data and reliably identify patients with psychotic disorders. This may be useful in identifying candidates for potential inclusion in cohort studies.

  4. Identifying gender-preferred communication styles within online cancer communities: a retrospective, longitudinal analysis.

    Science.gov (United States)

    Durant, Kathleen T; McCray, Alexa T; Safran, Charles

    2012-01-01

    The goal of this research is to determine if different gender-preferred social styles can be observed within the user interactions at an online cancer community. To achieve this goal, we identify and measure variables that pertain to each gender-specific social style. We perform social network and statistical analysis on the communication flow of 8,388 members at six different cancer forums over eight years. Kruskal-Wallis tests were conducted to measure the difference between the number of intimate (and highly intimate) dyads, relationship length, and number of communications. We determine that two patients are more likely to form an intimate bond on a gender-specific cancer forum (ovarian P = communicates with more members than a female patient (Ovarian forum P = 0.0406, Breast forum P = 0.0013). A relationship between two patients is longer on the gender-specific cancer forums than a connection between two members not identified as patients (ovarian forum P = 0.00406, breast forum P = 0.00013, prostate forum P = .0.0003). The high level of interconnectedness among the prostate patients supports the hypothesis that men prefer to socialize in large, interconnected, less-intimate groups. A female patient is more likely to form a highly intimate connection with another female patient; this finding is consistent with the hypothesis that woman prefer fewer, more intimate connections. The relationships of same-gender cancer patients last longer than other relationships; this finding demonstrates homophily within these online communities. Our findings regarding online communication preferences are in agreement with research findings from person-to-person communication preference studies. These findings should be considered when designing online communities as well as designing and evaluating psychosocial and educational interventions for cancer patients.

  5. Semisupervised Community Detection by Voltage Drops

    Directory of Open Access Journals (Sweden)

    Min Ji

    2016-01-01

    Full Text Available Many applications show that semisupervised community detection is one of the important topics and has attracted considerable attention in the study of complex network. In this paper, based on notion of voltage drops and discrete potential theory, a simple and fast semisupervised community detection algorithm is proposed. The label propagation through discrete potential transmission is accomplished by using voltage drops. The complexity of the proposal is OV+E for the sparse network with V vertices and E edges. The obtained voltage value of a vertex can be reflected clearly in the relationship between the vertex and community. The experimental results on four real networks and three benchmarks indicate that the proposed algorithm is effective and flexible. Furthermore, this algorithm is easily applied to graph-based machine learning methods.

  6. Vectorised Spreading Activation algorithm for centrality measurement

    Directory of Open Access Journals (Sweden)

    Alexander Troussov

    2011-01-01

    Full Text Available Spreading Activation is a family of graph-based algorithms widely used in areas such as information retrieval, epidemic models, and recommender systems. In this paper we introduce a novel Spreading Activation (SA method that we call Vectorised Spreading Activation (VSA. VSA algorithms, like “traditional” SA algorithms, iteratively propagate the activation from the initially activated set of nodes to the other nodes in a network through outward links. The level of the node’s activation could be used as a centrality measurement in accordance with dynamic model-based view of centrality that focuses on the outcomes for nodes in a network where something is flowing from node to node across the edges. Representing the activation by vectors allows the use of the information about various dimensionalities of the flow and the dynamic of the flow. In this capacity, VSA algorithms can model multitude of complex multidimensional network flows. We present the results of numerical simulations on small synthetic social networks and multi­dimensional network models of folksonomies which show that the results of VSA propagation are more sensitive to the positions of the initial seed and to the community structure of the network than the results produced by traditional SA algorithms. We tentatively conclude that the VSA methods could be instrumental to develop scalable and computationally efficient algorithms which could achieve synergy between computation of centrality indexes with detection of community structures in networks. Based on our preliminary results and on improvements made over previous studies, we foresee advances and applications in the current state of the art of this family of algorithms and their applications to centrality measurement.

  7. Using the Chandra Source-Finding Algorithm to Automatically Identify Solar X-ray Bright Points

    Science.gov (United States)

    Adams, Mitzi L.; Tennant, A.; Cirtain, J. M.

    2009-01-01

    This poster details a technique of bright point identification that is used to find sources in Chandra X-ray data. The algorithm, part of a program called LEXTRCT, searches for regions of a given size that are above a minimum signal to noise ratio. The algorithm allows selected pixels to be excluded from the source-finding, thus allowing exclusion of saturated pixels (from flares and/or active regions). For Chandra data the noise is determined by photon counting statistics, whereas solar telescopes typically integrate a flux. Thus the calculated signal-to-noise ratio is incorrect, but we find we can scale the number to get reasonable results. For example, Nakakubo and Hara (1998) find 297 bright points in a September 11, 1996 Yohkoh image; with judicious selection of signal-to-noise ratio, our algorithm finds 300 sources. To further assess the efficacy of the algorithm, we analyze a SOHO/EIT image (195 Angstroms) and compare results with those published in the literature (McIntosh and Gurman, 2005). Finally, we analyze three sets of data from Hinode, representing different parts of the decline to minimum of the solar cycle.

  8. Identifying medication-related needs of HIV patients: foundation for community pharmacist-based services

    Directory of Open Access Journals (Sweden)

    Yardlee Kauffman

    2014-01-01

    Full Text Available Background: Patients living with HIV/AIDS have complex medication regimens. Pharmacists within community pharmacy settings can have a role managing patients living with HIV/AIDS. Patients' perspectives surrounding implementation about community pharmacist-based services is needed as limited information is available. Objective: To identify medication-related needs of HIV-infected patients who receive prescriptions from a community pharmacy. To determine patient perspectives and knowledge of community pharmacist-based services. Methods: A qualitative research study involving in-depth, semi-structured interviews with patients was conducted. Inclusion criteria included: HIV positive men and women at least 18 years of age who receive care at a HIV clinic, currently take medication(s and use a community pharmacy for all prescription fills. Patients were recruited from one urban and one rural health center. Patients answered questions about their perceptions and knowledge about the role and value of pharmacy services and completed a demographic survey. The recordings of the interviews were transcribed verbatim and were analyzed using principles of Grounded Theory. Results: Twenty-nine interviews were conducted: 15 participants from the urban site and 14 from the rural site. Five main themes emerged including: patients experience ongoing and varying medication-related needs; patients desire a pharmacist who is caring, knowledgeable and integrated with health care providers; patients expect ready access to drug therapy; patients value an individualized patient encounter, and patients need to be informed that a pharmacist-service exists. Conclusion: Patients with HIV value individualized and personal encounters with pharmacists at time intervals that are convenient for the patient. Patients felt that a one-on-one encounter with a pharmacist would be most valuable when initiating or modifying medication therapy. These patient perspectives can be useful for

  9. A community effort to assess and improve drug sensitivity prediction algorithms.

    Science.gov (United States)

    Costello, James C; Heiser, Laura M; Georgii, Elisabeth; Gönen, Mehmet; Menden, Michael P; Wang, Nicholas J; Bansal, Mukesh; Ammad-ud-din, Muhammad; Hintsanen, Petteri; Khan, Suleiman A; Mpindi, John-Patrick; Kallioniemi, Olli; Honkela, Antti; Aittokallio, Tero; Wennerberg, Krister; Collins, James J; Gallahan, Dan; Singer, Dinah; Saez-Rodriguez, Julio; Kaski, Samuel; Gray, Joe W; Stolovitzky, Gustavo

    2014-12-01

    Predicting the best treatment strategy from genomic information is a core goal of precision medicine. Here we focus on predicting drug response based on a cohort of genomic, epigenomic and proteomic profiling data sets measured in human breast cancer cell lines. Through a collaborative effort between the National Cancer Institute (NCI) and the Dialogue on Reverse Engineering Assessment and Methods (DREAM) project, we analyzed a total of 44 drug sensitivity prediction algorithms. The top-performing approaches modeled nonlinear relationships and incorporated biological pathway information. We found that gene expression microarrays consistently provided the best predictive power of the individual profiling data sets; however, performance was increased by including multiple, independent data sets. We discuss the innovations underlying the top-performing methodology, Bayesian multitask MKL, and we provide detailed descriptions of all methods. This study establishes benchmarks for drug sensitivity prediction and identifies approaches that can be leveraged for the development of new methods.

  10. Algorithms to Identify Statin Intolerance in Medicare Administrative Claim Data.

    Science.gov (United States)

    Colantonio, Lisandro D; Kent, Shia T; Huang, Lei; Chen, Ligong; Monda, Keri L; Serban, Maria-Corina; Manthripragada, Angelika; Kilgore, Meredith L; Rosenson, Robert S; Muntner, Paul

    2016-10-01

    To compare characteristics of patients with possible statin intolerance identified using different claims-based algorithms versus patients with high adherence to statins. We analyzed 134,863 Medicare beneficiaries initiating statins between 2007 and 2011. Statin intolerance and discontinuation, and high adherence to statins, defined by proportion of days covered ≥80 %, were assessed during the 365 days following statin initiation. Definition 1 of statin intolerance included statin down-titration or discontinuation with ezetimibe initiation, having a claim for a rhabdomyolysis or antihyperlipidemic event followed by statin down-titration or discontinuation, or switching between ≥3 types of statins. Definition 2 included beneficiaries who met Definition 1 and those who down-titrated statin intensity. We also analyzed beneficiaries who met Definition 2 of statin intolerance or discontinued statins. The prevalence of statin intolerance was 1.0 % (n = 1320) and 5.2 % (n = 6985) using Definitions 1 and 2, respectively. Overall, 45,266 (33.6 %) beneficiaries had statin intolerance by Definition 2 or discontinued statins and 55,990 (41.5 %) beneficiaries had high adherence to statins. Compared with beneficiaries with high adherence to statins, those with statin intolerance and who had statin intolerance or discontinued statins were more likely to be female versus male, and black, Hispanic or Asian versus white. The multivariable adjusted odds ratio for statin intolerance by Definitions 1 and 2 comparing patients initiating high versus low/moderate intensity statins were 2.82 (95%CI: 2.42-3.29), and 8.58 (8.07-9.12), respectively, and for statin intolerance or statin discontinuation was 2.35 (2.25-2.45). Definitions of statin intolerance presented herein can be applied to analyses using administrative claims data.

  11. The use of concept mapping to identify community-driven intervention strategies for physical and mental health.

    Science.gov (United States)

    Vaughn, Lisa M; Jacquez, Farrah; McLinden, Daniel

    2013-09-01

    Research that partners with youth and community stakeholders increases contextual relevance and community buy-in and therefore maximizes the chance for intervention success. Concept mapping is a mixed-method participatory research process that accesses the input of the community in a collaborative manner. After a school-wide health needs assessment at a low-income, minority/immigrant K-8 school identified bullying and obesity as the most important health issues, concept mapping was used to identify and prioritize specific strategies to address these two areas. Stakeholders including 160 K-8 students, 33 college students working in the school, 35 parents, 20 academic partners, and 22 teachers/staff brainstormed strategies to reduce and prevent obesity and bullying. A smaller group of stakeholders worked individually to complete an unstructured sorting of these strategies into groups of similar ideas, once for obesity and again for bullying. Multidimensional scaling and cluster analysis was applied to the sorting data to produce a series of maps that illustrated the stakeholders' conceptual thinking about obesity and bullying prevention strategies. The maps for both obesity and bullying organized specific strategies into themes that included education, parental role, teacher/school supervision, youth role, expert/professional role, and school structure/support.

  12. Hyperbolic mapping of complex networks based on community information

    Science.gov (United States)

    Wang, Zuxi; Li, Qingguang; Jin, Fengdong; Xiong, Wei; Wu, Yao

    2016-08-01

    To improve the hyperbolic mapping methods both in terms of accuracy and running time, a novel mapping method called Community and Hyperbolic Mapping (CHM) is proposed based on community information in this paper. Firstly, an index called Community Intimacy (CI) is presented to measure the adjacency relationship between the communities, based on which a community ordering algorithm is introduced. According to the proposed Community-Sector hypothesis, which supposes that most nodes of one community gather in a same sector in hyperbolic space, CHM maps the ordered communities into hyperbolic space, and then the angular coordinates of nodes are randomly initialized within the sector that they belong to. Therefore, all the network nodes are so far mapped to hyperbolic space, and then the initialized angular coordinates can be optimized by employing the information of all nodes, which can greatly improve the algorithm precision. By applying the proposed dual-layer angle sampling method in the optimization procedure, CHM reduces the time complexity to O(n2) . The experiments show that our algorithm outperforms the state-of-the-art methods.

  13. Mapping radioactivity in groundwater to identify elevated exposure in remote and rural communities

    Energy Technology Data Exchange (ETDEWEB)

    Kleinschmidt, Ross, E-mail: ross_kleinschmidt@health.qld.gov.a [Queensland University of Technology, Faculty of Science and Technology, Discipline of Physics, 2 George Street, Brisbane, Queensland 4000 (Australia); Health Physics Unit, Queensland Health Forensic and Scientific Services, 39 Kessels Road, Coopers Plains, Queensland 4108 (Australia); Black, Jeffrey [Health Physics Unit, Queensland Health Forensic and Scientific Services, 39 Kessels Road, Coopers Plains, Queensland 4108 (Australia); Akber, Riaz [Queensland University of Technology, Faculty of Science and Technology, Discipline of Physics, 2 George Street, Brisbane, Queensland 4000 (Australia)

    2011-03-15

    A survey of radioactivity in groundwater (110 sites) was conducted as a precursor to providing a baseline of radiation exposure in rural and remote communities in Queensland, Australia, that may be impacted upon by exposure pathways associated with the supply, treatment, use and wastewater treatment of the resource. Radionuclides in groundwater, including {sup 238}U, {sup 226}Ra, {sup 222}Rn, {sup 228}Ra, {sup 224}Ra and {sup 40}K were measured and found to contain activity concentration levels of up to 0.71 BqL{sup -1}, 0.96 BqL{sup -1}, 108 BqL{sup -1}, 2.8 BqL{sup -1}, 0.11 BqL{sup -1} and 0.19 BqL{sup -1} respectively. Activity concentration results were classified by aquifer lithology, showing correlation between increased radium isotope concentration and basic volcanic host rock. The groundwater survey and mapping results were further assessed using an investigation assessment tool to identify seven remote or rural communities that may require additional radiation dose assessment beyond that attributed to ingestion of potable water. - Research highlights: {yields} We studied the concentration of naturally occurring radioactivity in groundwater in Queensland, Australia. {yields} Groundwater radioactivity concentrations were classified by aquifer type, location and magnitude. {yields} Radioactivity concentration in groundwater was used to develop a tool to determine the potential for elevated radiation exposure to rural and remote communities, based on a case study of a reference site. {yields} Of 110 groundwater bores tested, seven were assessed as requiring further community dose assessment.

  14. Marshall Rosenbluth and the Metropolis algorithm

    International Nuclear Information System (INIS)

    Gubernatis, J.E.

    2005-01-01

    The 1953 publication, 'Equation of State Calculations by Very Fast Computing Machines' by N. Metropolis, A. W. Rosenbluth and M. N. Rosenbluth, and M. Teller and E. Teller [J. Chem. Phys. 21, 1087 (1953)] marked the beginning of the use of the Monte Carlo method for solving problems in the physical sciences. The method described in this publication subsequently became known as the Metropolis algorithm, undoubtedly the most famous and most widely used Monte Carlo algorithm ever published. As none of the authors made subsequent use of the algorithm, they became unknown to the large simulation physics community that grew from this publication and their roles in its development became the subject of mystery and legend. At a conference marking the 50th anniversary of the 1953 publication, Marshall Rosenbluth gave his recollections of the algorithm's development. The present paper describes the algorithm, reconstructs the historical context in which it was developed, and summarizes Marshall's recollections

  15. Geospatial techniques to Identify the Location of Farmers Markets and Community Gardens within Food Deserts in Virginia

    Science.gov (United States)

    Sriharan, S.; Meekins, D.; Comar, M.; Bradshaw, S.; Jackson, L.

    2017-12-01

    Specifically, a food desert is defined as an area where populations live more than one mile from a supermarket or large grocery store if in an urban area or more than 10 miles from a supermarket or large grocery store if in a rural area (Ver Ploeg et al. 2012). According to the U.S. Department of Agriculture, a food desert is "an area in the United States with limited access to affordable and nutritious food, particularly such an area composed of predominately lower-income neighborhoods and communities" (110th Congress 2008). Three fourths of these food deserts are urban. In the Commonwealth of Virginia, Petersburg City is among the eight primary localities, where its population is living in a food desert. This project will compare those identified food deserts in Virginia (areas around Virginia State University) with focus to where farmers markets and community gardens are being established. The hypothesis of this study is that these minority groups do not get healthy food due to limited access to grocery stores and superstores. To address this problem, the community development activities should focus on partnering local Petersburg convenience stores with farmers and community gardeners to sell fresh produce. Existing data was collected on convenient stores and community gardens in Petersburg City and Chesterfield County. Rare data was generated for Emporia, Lynchburg and Hopewell. The data was compiled through field work and mapping with ArcGIS where markets and gardens are being established, and create a spatial analysis of their location We have localities that reflect both rural and urban areas. The project provides educational support for students who will find solution to community problems by developing activities to: (a) define and examine characteristics of food deserts, (b) identify causes and consequences of food deserts and determine if their community is a food desert, (c) research closest food desert to their school, and (d) design solutions to help

  16. Identifying Pedophiles "Eligible" for Community Notification under Megan's Law: A Multivariate Model for Actuarially Anchored Decisions.

    Science.gov (United States)

    Pallone, Nathaniel J.; Hennessy, James J.; Voelbel, Gerald T.

    1998-01-01

    A scientifically sound methodology for identifying offenders about whose presence the community should be notified is demonstrated. A stepwise multiple regression was calculated among incarcerated pedophiles (N=52) including both psychological and legal data; a precision-weighted equation produced 90.4% "true positives." This methodology can be…

  17. Analysis and visualization of social user communities

    Directory of Open Access Journals (Sweden)

    Daniel LÓPEZ SÁNCHEZ

    2016-06-01

    Full Text Available In this paper, a novel framework for social user clustering is proposed. Given a current controversial political topic, the Louvain Modularity algorithm is used to detect communities of users sharing the same political preferences. The political alignment of a set of users is labeled manually by a human expert and then the quality of the community detection is evaluated against this gold standard. In the last section, we propose a novel force-directed graph algorithm to generate a visual representation of the detected communities.   

  18. Finding local communities in protein networks.

    Science.gov (United States)

    Voevodski, Konstantin; Teng, Shang-Hua; Xia, Yu

    2009-09-18

    Protein-protein interactions (PPIs) play fundamental roles in nearly all biological processes, and provide major insights into the inner workings of cells. A vast amount of PPI data for various organisms is available from BioGRID and other sources. The identification of communities in PPI networks is of great interest because they often reveal previously unknown functional ties between proteins. A large number of global clustering algorithms have been applied to protein networks, where the entire network is partitioned into clusters. Here we take a different approach by looking for local communities in PPI networks. We develop a tool, named Local Protein Community Finder, which quickly finds a community close to a queried protein in any network available from BioGRID or specified by the user. Our tool uses two new local clustering algorithms Nibble and PageRank-Nibble, which look for a good cluster among the most popular destinations of a short random walk from the queried vertex. The quality of a cluster is determined by proportion of outgoing edges, known as conductance, which is a relative measure particularly useful in undersampled networks. We show that the two local clustering algorithms find communities that not only form excellent clusters, but are also likely to be biologically relevant functional components. We compare the performance of Nibble and PageRank-Nibble to other popular and effective graph partitioning algorithms, and show that they find better clusters in the graph. Moreover, Nibble and PageRank-Nibble find communities that are more functionally coherent. The Local Protein Community Finder, accessible at http://xialab.bu.edu/resources/lpcf, allows the user to quickly find a high-quality community close to a queried protein in any network available from BioGRID or specified by the user. We show that the communities found by our tool form good clusters and are functionally coherent, making our application useful for biologists who wish to

  19. Finding local communities in protein networks

    Directory of Open Access Journals (Sweden)

    Teng Shang-Hua

    2009-09-01

    Full Text Available Abstract Background Protein-protein interactions (PPIs play fundamental roles in nearly all biological processes, and provide major insights into the inner workings of cells. A vast amount of PPI data for various organisms is available from BioGRID and other sources. The identification of communities in PPI networks is of great interest because they often reveal previously unknown functional ties between proteins. A large number of global clustering algorithms have been applied to protein networks, where the entire network is partitioned into clusters. Here we take a different approach by looking for local communities in PPI networks. Results We develop a tool, named Local Protein Community Finder, which quickly finds a community close to a queried protein in any network available from BioGRID or specified by the user. Our tool uses two new local clustering algorithms Nibble and PageRank-Nibble, which look for a good cluster among the most popular destinations of a short random walk from the queried vertex. The quality of a cluster is determined by proportion of outgoing edges, known as conductance, which is a relative measure particularly useful in undersampled networks. We show that the two local clustering algorithms find communities that not only form excellent clusters, but are also likely to be biologically relevant functional components. We compare the performance of Nibble and PageRank-Nibble to other popular and effective graph partitioning algorithms, and show that they find better clusters in the graph. Moreover, Nibble and PageRank-Nibble find communities that are more functionally coherent. Conclusion The Local Protein Community Finder, accessible at http://xialab.bu.edu/resources/lpcf, allows the user to quickly find a high-quality community close to a queried protein in any network available from BioGRID or specified by the user. We show that the communities found by our tool form good clusters and are functionally coherent

  20. Algorithm of actions to identify and reduce risks in the production of milk and plant products

    Directory of Open Access Journals (Sweden)

    L. E. Glagoleva

    2016-01-01

    Full Text Available Foods with a new generation of functional and improved consumer properties, corresponds to the modern concepts of nutrition science and consumer needs. functional food production is a major global trend in food science and the subject of innovation. One of the important trends is the use of plant complexes and plant food systems. Using the plant complexes (PC and plant food systems (PFS provides a number of benefits: improved consumer properties of the product, do not need to change the process, it is possible to control directional rheological properties and consistency of the finished products, reduced the number of risk points in the production cycle. This paper describes the development of an algorithm of action to identify and mitigate risks in the production of milk and plant products. Also conducted a risk analysis, identified and assessed the risks in the process of production, installed capacity of available resources to reduce the level of risk. Established and submitted to the critical control points in production processes, as well as the critical limits for each critical control points, and the procedure for corrective action in case of violations of the past. During the study, measured changes in the quantitative and qualitative composition of microflora of semi-finished and Quantity of Mesophilic Aerobic and Facultative Anaerobic Microorganisms (QMAFAnM. To determine QMAFAnM samples were taken: 1 – cheesecakes (control, 2 – cheesecakes with RPS. Microbiological studies analyzed frozen-conjugated semi-finished products was determined within 90 days. It is clear from the data that the cottage cheese with semi-finished products have a lower RPM 11.7%. Analyzing the data, it is possible to conclude that the physico-chemical, organoleptic and microbiological indicators of products was developed to set standards on cheese semi-finished products. multilevel structure that characterizes the quality indicators has been developed and is

  1. Strength-based well-being indicators for Indigenous children and families: A literature review of Indigenous communities' identified well-being indicators.

    Science.gov (United States)

    Rountree, Jennifer; Smith, Addie

    2016-01-01

    Mainstream child and family well-being indicators frequently are based on measuring health, economic, and social deficits, and do not reflect Indigenous holistic and strength-based definitions of health and well-being. The present article is a review of literature that features Indigenous communities' self-identified strength-based indicators of child and family well-being. The literature search included Indigenous communities from across the world, incorporating findings from American Indians and Alaska Natives, First Nations, Native Hawaiians, Māori, Aboriginal Australians, and Sámi communities. Sorting the identified indicators into the quadrants of the Relational Worldview, an Indigenous framework for well-being based on medicine wheel teachings that views health and well-being as a balance among physical, mental, contextual, and spiritual factors, the authors discuss the findings.

  2. Identifying gender-preferred communication styles within online cancer communities: a retrospective, longitudinal analysis.

    Directory of Open Access Journals (Sweden)

    Kathleen T Durant

    Full Text Available BACKGROUND: The goal of this research is to determine if different gender-preferred social styles can be observed within the user interactions at an online cancer community. To achieve this goal, we identify and measure variables that pertain to each gender-specific social style. METHODS AND FINDINGS: We perform social network and statistical analysis on the communication flow of 8,388 members at six different cancer forums over eight years. Kruskal-Wallis tests were conducted to measure the difference between the number of intimate (and highly intimate dyads, relationship length, and number of communications. We determine that two patients are more likely to form an intimate bond on a gender-specific cancer forum (ovarian P = <0.0001, breast P = 0.0089, prostate P = 0.0021. Two female patients are more likely to form a highly intimate bond on a female-specific cancer forum (Ovarian P<0.0001, Breast P<0.01. Typically a male patient communicates with more members than a female patient (Ovarian forum P = 0.0406, Breast forum P = 0.0013. A relationship between two patients is longer on the gender-specific cancer forums than a connection between two members not identified as patients (ovarian forum P = 0.00406, breast forum P = 0.00013, prostate forum P = .0.0003. CONCLUSION: The high level of interconnectedness among the prostate patients supports the hypothesis that men prefer to socialize in large, interconnected, less-intimate groups. A female patient is more likely to form a highly intimate connection with another female patient; this finding is consistent with the hypothesis that woman prefer fewer, more intimate connections. The relationships of same-gender cancer patients last longer than other relationships; this finding demonstrates homophily within these online communities. Our findings regarding online communication preferences are in agreement with research findings from person-to-person communication

  3. A new algorithm for identifying the flavour of B 0s mesons at LHCb

    International Nuclear Information System (INIS)

    Aaij, R.; Alessio, F.; Beteta, C. Abellán; Adeva, B.; Adinolfi, M.; Affolder, A.; Ajaltouni, Z.; Akar, S.; Albrecht, J.; Alexander, M.; Ali, S.; Alkhazov, G.; Cartelle, P. Alvarez; Jr, A.A. Alves; Amato, S.; Amerio, S.; Amhis, Y.; An, L.; Anderlini, L.; Andreassi, G.

    2016-01-01

    A new algorithm for the determination of the initial flavour of B 0 s mesons is presented. The algorithm is based on two neural networks and exploits the b hadron production mechanism at a hadron collider. The first network is trained to select charged kaons produced in association with the B 0 s meson. The second network combines the kaon charges to assign the B 0 s flavour and estimates the probability of a wrong assignment. The algorithm is calibrated using data corresponding to an integrated luminosity of 3 fb −1 collected by the LHCb experiment in proton-proton collisions at 7 and 8 TeV centre-of-mass energies. The calibration is performed in two ways: by resolving the B 0 s – B-bar 0 s flavour oscillations in B 0 s  →  D − s π + decays, and by analysing flavour-specific B * s2 (5840) 0  →  B + K − decays. The tagging power measured in B 0 s  →  D − s π + decays is found to be (1.80 ± 0.19 (stat) ± 0.18 (syst))%, which is an improvement of about 50% compared to a similar algorithm previously used in the LHCb experiment.

  4. Algorithms to identify colonic ischemia, complications of constipation and irritable bowel syndrome in medical claims data: development and validation.

    Science.gov (United States)

    Sands, Bruce E; Duh, Mei-Sheng; Cali, Clorinda; Ajene, Anuli; Bohn, Rhonda L; Miller, David; Cole, J Alexander; Cook, Suzanne F; Walker, Alexander M

    2006-01-01

    A challenge in the use of insurance claims databases for epidemiologic research is accurate identification and verification of medical conditions. This report describes the development and validation of claims-based algorithms to identify colonic ischemia, hospitalized complications of constipation, and irritable bowel syndrome (IBS). From the research claims databases of a large healthcare company, we selected at random 120 potential cases of IBS and 59 potential cases each of colonic ischemia and hospitalized complications of constipation. We sought the written medical records and were able to abstract 107, 57, and 51 records, respectively. We established a 'true' case status for each subject by applying standard clinical criteria to the available chart data. Comparing the insurance claims histories to the assigned case status, we iteratively developed, tested, and refined claims-based algorithms that would capture the diagnoses obtained from the medical records. We set goals of high specificity for colonic ischemia and hospitalized complications of constipation, and high sensitivity for IBS. The resulting algorithms substantially improved on the accuracy achievable from a naïve acceptance of the diagnostic codes attached to insurance claims. The specificities for colonic ischemia and serious complications of constipation were 87.2 and 92.7%, respectively, and the sensitivity for IBS was 98.9%. U.S. commercial insurance claims data appear to be usable for the study of colonic ischemia, IBS, and serious complications of constipation. (c) 2005 John Wiley & Sons, Ltd.

  5. Sensor and ad-hoc networks theoretical and algorithmic aspects

    CERN Document Server

    Makki, S Kami; Pissinou, Niki; Makki, Shamila; Karimi, Masoumeh; Makki, Kia

    2008-01-01

    This book brings together leading researchers and developers in the field of wireless sensor networks to explain the special problems and challenges of the algorithmic aspects of sensor and ad-hoc networks. The book also fosters communication not only between the different sensor and ad-hoc communities, but also between those communities and the distributed systems and information systems communities. The topics addressed pertain to the sensors and mobile environment.

  6. Evaluation Of Algorithms Of Anti- HIV Antibody Tests

    Directory of Open Access Journals (Sweden)

    Paranjape R.S

    1997-01-01

    Full Text Available Research question: Can alternate algorithms be used in place of conventional algorithm for epidemiological studies of HIV infection with less expenses? Objective: To compare the results of HIV sero- prevalence as determined by test algorithms combining three kits with conventional test algorithm. Study design: Cross â€" sectional. Participants: 282 truck drivers. Statistical analysis: Sensitivity and specificity analysis and predictive values. Results: Three different algorithms that do not include Western Blot (WB were compared with the conventional algorithm, in a truck driver population with 5.6% prevalence of HIV â€"I infection. Algorithms with one EIA (Genetic Systems or Biotest and a rapid test (immunocomb or with two EIAs showed 100% positive predictive value in relation to the conventional algorithm. Using an algorithm with EIA as screening test and a rapid test as a confirmatory test was 50 to 70% less expensive than the conventional algorithm per positive scrum sample. These algorithms obviate the interpretation of indeterminate results and also give differential diagnosis of HIV-2 infection. Alternate algorithms are ideally suited for community based control programme in developing countries. Application of these algorithms in population with low prevalence should also be studied in order to evaluate universal applicability.

  7. Clustering coefficient and community structure of bipartite networks

    Science.gov (United States)

    Zhang, Peng; Wang, Jinliang; Li, Xiaojia; Li, Menghui; Di, Zengru; Fan, Ying

    2008-12-01

    Many real-world networks display natural bipartite structure, where the basic cycle is a square. In this paper, with the similar consideration of standard clustering coefficient in binary networks, a definition of the clustering coefficient for bipartite networks based on the fraction of squares is proposed. In order to detect community structures in bipartite networks, two different edge clustering coefficients LC4 and LC3 of bipartite networks are defined, which are based on squares and triples respectively. With the algorithm of cutting the edge with the least clustering coefficient, communities in artificial and real world networks are identified. The results reveal that investigating bipartite networks based on the original structure can show the detailed properties that is helpful to get deep understanding about the networks.

  8. Comparing Whole Building Energy Implications of Sidelighting Systems with Alternate Manual Blind Control Algorithms

    Directory of Open Access Journals (Sweden)

    Christopher Dyke

    2015-05-01

    Full Text Available Currently, there is no manual blind control guideline used consistently throughout the energy modeling community. This paper identifies and compares five manual blind control algorithms with unique control patterns and reports blind occlusion, rate of change data, and annual building energy consumption. The blind control schemes detailed here represent five reasonable candidates for use in lighting and energy simulation based on difference driving factors. This study was performed on a medium-sized office building using EnergyPlus with the internal daylight harvesting engine. Results show that applying manual blind control algorithms affects the total annual consumption of the building by as much as 12.5% and 11.5% for interior and exterior blinds respectively, compared to the Always Retracted blinds algorithm. Peak demand was also compared showing blind algorithms affected zone load sizing by as much as 9.8%. The alternate algorithms were tested for their impact on American Society of Heating, Refrigeration and Air-Conditioning Engineers (ASHRAE Guideline 14 calibration metrics and all models were found to differ from the original calibrated baseline by more than the recommended ±15% for coefficient of variance of the mean square error (CVRMSE and ±5% for normalized mean bias error (NMBE. The paper recommends that energy modelers use one or more manual blind control algorithms during design stages when making decisions about energy efficiency and other design alternatives.

  9. Overlapping community detection in networks with positive and negative links

    International Nuclear Information System (INIS)

    Chen, Y; Wang, X L; Yuan, B; Tang, B Z

    2014-01-01

    Complex networks considering both positive and negative links have gained considerable attention during the past several years. Community detection is one of the main challenges for complex network analysis. Most of the existing algorithms for community detection in a signed network aim at providing a hard-partition of the network where any node should belong to a community or not. However, they cannot detect overlapping communities where a node is allowed to belong to multiple communities. The overlapping communities widely exist in many real-world networks. In this paper, we propose a signed probabilistic mixture (SPM) model for overlapping community detection in signed networks. Compared with the existing models, the advantages of our methodology are (i) providing soft-partition solutions for signed networks; (ii) providing soft memberships of nodes. Experiments on a number of signed networks show that our SPM model: (i) can identify assortative structures or disassortative structures as the same as other state-of-the-art models; (ii) can detect overlapping communities; (iii) outperforms other state-of-the-art models at shedding light on the community detection in synthetic signed networks. (paper)

  10. On the relationship between Gaussian stochastic blockmodels and label propagation algorithms

    International Nuclear Information System (INIS)

    Zhang, Junhao; Hu, Junfeng; Chen, Tongfei

    2015-01-01

    The problem of community detection has received great attention in recent years. Many methods have been proposed to discover communities in networks. In this paper, we propose a Gaussian stochastic blockmodel that uses Gaussian distributions to fit weight of edges in networks for non-overlapping community detection. The maximum likelihood estimation of this model has the same objective function as general label propagation with node preference. The node preference of a specific vertex turns out to be a value proportional to the intra-community eigenvector centrality (the corresponding entry in principal eigenvector of the adjacency matrix of the subgraph inside that vertex's community) under maximum likelihood estimation. Additionally, the maximum likelihood estimation of a constrained version of our model is highly related to another extension of the label propagation algorithm, namely, the label propagation algorithm under constraint. Experiments show that the proposed Gaussian stochastic blockmodel performs well on various benchmark networks. (paper)

  11. Identifying Patients with Bacteremia in Community-Hospital Emergency Rooms: A Retrospective Cohort Study.

    Directory of Open Access Journals (Sweden)

    Taro Takeshima

    Full Text Available (1 To develop a clinical prediction rule to identify patients with bacteremia, using only information that is readily available in the emergency room (ER of community hospitals, and (2 to test the validity of that rule with a separate, independent set of data.Multicenter retrospective cohort study.To derive the clinical prediction rule we used data from 3 community hospitals in Japan (derivation. We tested the rule using data from one other community hospital (validation, which was not among the three "derivation" hospitals.Adults (age ≥ 16 years old who had undergone blood-culture testing while in the ER between April 2011 and March 2012. For the derivation data, n = 1515 (randomly sampled from 7026 patients, and for the validation data n = 467 (from 823 patients.We analyzed 28 candidate predictors of bacteremia, including demographic data, signs and symptoms, comorbid conditions, and basic laboratory data. Chi-square tests and multiple logistic regression were used to derive an integer risk score (the "ID-BactER" score. Sensitivity, specificity, likelihood ratios, and the area under the receiver operating characteristic curve (i.e., the AUC were computed.There were 241 cases of bacteremia in the derivation data. Eleven candidate predictors were used in the ID-BactER score: age, chills, vomiting, mental status, temperature, systolic blood pressure, abdominal sign, white blood-cell count, platelets, blood urea nitrogen, and C-reactive protein. The AUCs was 0.80 (derivation and 0.74 (validation. For ID-BactER scores ≥ 2, the sensitivities for derivation and validation data were 98% and 97%, and specificities were 20% and 14%, respectively.The ID-BactER score can be computed from information that is readily available in the ERs of community hospitals. Future studies should focus on developing a score with a higher specificity while maintaining the desired sensitivity.

  12. Machine learning for identifying botnet network traffic

    DEFF Research Database (Denmark)

    Stevanovic, Matija; Pedersen, Jens Myrup

    2013-01-01

    . Due to promise of non-invasive and resilient detection, botnet detection based on network traffic analysis has drawn a special attention of the research community. Furthermore, many authors have turned their attention to the use of machine learning algorithms as the mean of inferring botnet......-related knowledge from the monitored traffic. This paper presents a review of contemporary botnet detection methods that use machine learning as a tool of identifying botnet-related traffic. The main goal of the paper is to provide a comprehensive overview on the field by summarizing current scientific efforts....... The contribution of the paper is three-fold. First, the paper provides a detailed insight on the existing detection methods by investigating which bot-related heuristic were assumed by the detection systems and how different machine learning techniques were adapted in order to capture botnet-related knowledge...

  13. HPC-NMF: A High-Performance Parallel Algorithm for Nonnegative Matrix Factorization

    Energy Technology Data Exchange (ETDEWEB)

    2016-08-22

    NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient distributed algorithms to solve the problem for big data sets. We propose a high-performance distributed-memory parallel algorithm that computes the factorization by iteratively solving alternating non-negative least squares (NLS) subproblems for $\\WW$ and $\\HH$. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). As opposed to previous implementation, our algorithm is also flexible: It performs well for both dense and sparse matrices, and allows the user to choose any one of the multiple algorithms for solving the updates to low rank factors $\\WW$ and $\\HH$ within the alternating iterations.

  14. Inclusive Flavour Tagging Algorithm

    International Nuclear Information System (INIS)

    Likhomanenko, Tatiana; Derkach, Denis; Rogozhnikov, Alex

    2016-01-01

    Identifying the flavour of neutral B mesons production is one of the most important components needed in the study of time-dependent CP violation. The harsh environment of the Large Hadron Collider makes it particularly hard to succeed in this task. We present an inclusive flavour-tagging algorithm as an upgrade of the algorithms currently used by the LHCb experiment. Specifically, a probabilistic model which efficiently combines information from reconstructed vertices and tracks using machine learning is proposed. The algorithm does not use information about underlying physics process. It reduces the dependence on the performance of lower level identification capacities and thus increases the overall performance. The proposed inclusive flavour-tagging algorithm is applicable to tag the flavour of B mesons in any proton-proton experiment. (paper)

  15. Data fusion for a vision-aided radiological detection system: Calibration algorithm performance

    Science.gov (United States)

    Stadnikia, Kelsey; Henderson, Kristofer; Martin, Allan; Riley, Phillip; Koppal, Sanjeev; Enqvist, Andreas

    2018-05-01

    In order to improve the ability to detect, locate, track and identify nuclear/radiological threats, the University of Florida nuclear detection community has teamed up with the 3D vision community to collaborate on a low cost data fusion system. The key is to develop an algorithm to fuse the data from multiple radiological and 3D vision sensors as one system. The system under development at the University of Florida is being assessed with various types of radiological detectors and widely available visual sensors. A series of experiments were devised utilizing two EJ-309 liquid organic scintillation detectors (one primary and one secondary), a Microsoft Kinect for Windows v2 sensor and a Velodyne HDL-32E High Definition LiDAR Sensor which is a highly sensitive vision sensor primarily used to generate data for self-driving cars. Each experiment consisted of 27 static measurements of a source arranged in a cube with three different distances in each dimension. The source used was Cf-252. The calibration algorithm developed is utilized to calibrate the relative 3D-location of the two different types of sensors without need to measure it by hand; thus, preventing operator manipulation and human errors. The algorithm can also account for the facility dependent deviation from ideal data fusion correlation. Use of the vision sensor to determine the location of a sensor would also limit the possible locations and it does not allow for room dependence (facility dependent deviation) to generate a detector pseudo-location to be used for data analysis later. Using manually measured source location data, our algorithm-predicted the offset detector location within an average of 20 cm calibration-difference to its actual location. Calibration-difference is the Euclidean distance from the algorithm predicted detector location to the measured detector location. The Kinect vision sensor data produced an average calibration-difference of 35 cm and the HDL-32E produced an average

  16. Efficient Record Linkage Algorithms Using Complete Linkage Clustering.

    Science.gov (United States)

    Mamun, Abdullah-Al; Aseltine, Robert; Rajasekaran, Sanguthevar

    2016-01-01

    Data from different agencies share data of the same individuals. Linking these datasets to identify all the records belonging to the same individuals is a crucial and challenging problem, especially given the large volumes of data. A large number of available algorithms for record linkage are prone to either time inefficiency or low-accuracy in finding matches and non-matches among the records. In this paper we propose efficient as well as reliable sequential and parallel algorithms for the record linkage problem employing hierarchical clustering methods. We employ complete linkage hierarchical clustering algorithms to address this problem. In addition to hierarchical clustering, we also use two other techniques: elimination of duplicate records and blocking. Our algorithms use sorting as a sub-routine to identify identical copies of records. We have tested our algorithms on datasets with millions of synthetic records. Experimental results show that our algorithms achieve nearly 100% accuracy. Parallel implementations achieve almost linear speedups. Time complexities of these algorithms do not exceed those of previous best-known algorithms. Our proposed algorithms outperform previous best-known algorithms in terms of accuracy consuming reasonable run times.

  17. Importance of Natural and Anthropogenic Environmental Factors to Fish Communities of the Fox River in Illinois

    Science.gov (United States)

    Schnier, Spencer; Cai, Ximing; Cao, Yong

    2016-02-01

    The dominant environmental determinants of aquatic communities have been a persistent topic for many years. Interactions between natural and anthropogenic characteristics within the aquatic environment influence fish communities in complex ways that make the effect of a single characteristic difficult to ascertain. Researchers are faced with the question of how to deal with a large number of variables and complex interrelationships. This study utilized multiple approaches to identify key environmental variables to fish communities of the Fox River Basin in Illinois: Pearson and Spearman correlations, an algorithm based on information theory called mutual information, and a measure of variable importance built into the machine learning algorithm Random Forest. The results are based on a dataset developed for this study, which uses a fish index of biological integrity (IBI) and its ten component metrics as response variables and a range of environmental variables describing geomorphology, stream flow statistics, climate, and both reach-scale and watershed-scale land use as independent variables. Agricultural land use and the magnitude and duration of low flow events were ranked by the algorithms as key factors for the study area. Reach-scale characteristics were dominant for native sunfish, and stream flow metrics were rated highly for native suckers. Regression tree analyses of environmental variables on fish IBI identified breakpoints in percent agricultural land in the watershed (~64 %), duration of low flow pulses (~12 days), and 90-day minimum flow (~0.13 cms). The findings should be useful for building predictive models and design of more effective monitoring systems and restoration plans.

  18. Algorithmic detectability threshold of the stochastic block model

    Science.gov (United States)

    Kawamoto, Tatsuro

    2018-03-01

    The assumption that the values of model parameters are known or correctly learned, i.e., the Nishimori condition, is one of the requirements for the detectability analysis of the stochastic block model in statistical inference. In practice, however, there is no example demonstrating that we can know the model parameters beforehand, and there is no guarantee that the model parameters can be learned accurately. In this study, we consider the expectation-maximization (EM) algorithm with belief propagation (BP) and derive its algorithmic detectability threshold. Our analysis is not restricted to the community structure but includes general modular structures. Because the algorithm cannot always learn the planted model parameters correctly, the algorithmic detectability threshold is qualitatively different from the one with the Nishimori condition.

  19. Thresher: an improved algorithm for peak height thresholding of microbial community profiles.

    Science.gov (United States)

    Starke, Verena; Steele, Andrew

    2014-11-15

    This article presents Thresher, an improved technique for finding peak height thresholds for automated rRNA intergenic spacer analysis (ARISA) profiles. We argue that thresholds must be sample dependent, taking community richness into account. In most previous fragment analyses, a common threshold is applied to all samples simultaneously, ignoring richness variations among samples and thereby compromising cross-sample comparison. Our technique solves this problem, and at the same time provides a robust method for outlier rejection, selecting for removal any replicate pairs that are not valid replicates. Thresholds are calculated individually for each replicate in a pair, and separately for each sample. The thresholds are selected to be the ones that minimize the dissimilarity between the replicates after thresholding. If a choice of threshold results in the two replicates in a pair failing a quantitative test of similarity, either that threshold or that sample must be rejected. We compare thresholded ARISA results with sequencing results, and demonstrate that the Thresher algorithm outperforms conventional thresholding techniques. The software is implemented in R, and the code is available at http://verenastarke.wordpress.com or by contacting the author. vstarke@ciw.edu or http://verenastarke.wordpress.com Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  20. Neural network based automated algorithm to identify joint locations on hand/wrist radiographs for arthritis assessment

    International Nuclear Information System (INIS)

    Duryea, J.; Zaim, S.; Wolfe, F.

    2002-01-01

    Arthritis is a significant and costly healthcare problem that requires objective and quantifiable methods to evaluate its progression. Here we describe software that can automatically determine the locations of seven joints in the proximal hand and wrist that demonstrate arthritic changes. These are the five carpometacarpal (CMC1, CMC2, CMC3, CMC4, CMC5), radiocarpal (RC), and the scaphocapitate (SC) joints. The algorithm was based on an artificial neural network (ANN) that was trained using independent sets of digitized hand radiographs and manually identified joint locations. The algorithm used landmarks determined automatically by software developed in our previous work as starting points. Other than requiring user input of the location of nonanatomical structures and the orientation of the hand on the film, the procedure was fully automated. The software was tested on two datasets: 50 digitized hand radiographs from patients participating in a large clinical study, and 60 from subjects participating in arthritis research studies and who had mild to moderate rheumatoid arthritis (RA). It was evaluated by a comparison to joint locations determined by a trained radiologist using manual tracing. The success rate for determining the CMC, RC, and SC joints was 87%-99%, for normal hands and 81%-99% for RA hands. This is a first step in performing an automated computer-aided assessment of wrist joints for arthritis progression. The software provides landmarks that will be used by subsequent image processing routines to analyze each joint individually for structural changes such as erosions and joint space narrowing

  1. miRvestigator: web application to identify miRNAs responsible for co-regulated gene expression patterns discovered through transcriptome profiling.

    Science.gov (United States)

    Plaisier, Christopher L; Bare, J Christopher; Baliga, Nitin S

    2011-07-01

    Transcriptome profiling studies have produced staggering numbers of gene co-expression signatures for a variety of biological systems. A significant fraction of these signatures will be partially or fully explained by miRNA-mediated targeted transcript degradation. miRvestigator takes as input lists of co-expressed genes from Caenorhabditis elegans, Drosophila melanogaster, G. gallus, Homo sapiens, Mus musculus or Rattus norvegicus and identifies the specific miRNAs that are likely to bind to 3' un-translated region (UTR) sequences to mediate the observed co-regulation. The novelty of our approach is the miRvestigator hidden Markov model (HMM) algorithm which systematically computes a similarity P-value for each unique miRNA seed sequence from the miRNA database miRBase to an overrepresented sequence motif identified within the 3'-UTR of the query genes. We have made this miRNA discovery tool accessible to the community by integrating our HMM algorithm with a proven algorithm for de novo discovery of miRNA seed sequences and wrapping these algorithms into a user-friendly interface. Additionally, the miRvestigator web server also produces a list of putative miRNA binding sites within 3'-UTRs of the query transcripts to facilitate the design of validation experiments. The miRvestigator is freely available at http://mirvestigator.systemsbiology.net.

  2. Identifying novel fruit-related genes in Arabidopsis thaliana based on the random walk with restart algorithm.

    Science.gov (United States)

    Zhang, Yunhua; Dai, Li; Liu, Ying; Zhang, YuHang; Wang, ShaoPeng

    2017-01-01

    Fruit is essential for plant reproduction and is responsible for protection and dispersal of seeds. The development and maturation of fruit is tightly regulated by numerous genetic factors that respond to environmental and internal stimulation. In this study, we attempted to identify novel fruit-related genes in a model organism, Arabidopsis thaliana, using a computational method. Based on validated fruit-related genes, the random walk with restart (RWR) algorithm was applied on a protein-protein interaction (PPI) network using these genes as seeds. The identified genes with high probabilities were filtered by the permutation test and linkage tests. In the permutation test, the genes that were selected due to the structure of the PPI network were discarded. In the linkage tests, the importance of each candidate gene was measured from two aspects: (1) its functional associations with validated genes and (2) its similarity with validated genes on gene ontology (GO) terms and KEGG pathways. Finally, 255 inferred genes were obtained, subsequent extensive analysis of important genes revealed that they mainly contribute to ubiquitination (UBQ9, UBQ8, UBQ11, UBQ10), serine hydroxymethyl transfer (SHM7, SHM5, SHM6) or glycol-metabolism (HXKL2_ARATH, CSY5, GAPCP1), suggesting essential roles during the development and maturation of fruit in Arabidopsis thaliana.

  3. Implementation of Winnowing Algorithm Based K-Gram to Identify Plagiarism on File Text-Based Document

    Directory of Open Access Journals (Sweden)

    Nurdiansyah Yanuar

    2018-01-01

    Full Text Available Plagiarism occurs when the students have tasks and pursued by the deadline. Plagiarism is considered as the fastest way to accomplish the tasks. This reason makes the author tried to build a plagiarism detection system with Winnowing algorithm as document similarity search algorithm. The documents that being tested are Indonesian journals with extension .doc, .docx, and/or .txt. Similarity calculation process through two stages, the first is the process of making a document fingerprint using Winnowing algorithm and the second is using Jaccard coefficient similarity. In order to develop this system, the author used iterative waterfall model approach. The main objective of this project is to determine the level of plagiarism. It is expected to prevent plagiarism either intentionally or unintentionally before our journal published by displaying the percentage of similarity in the journals that we make.

  4. BANYAN. XI. The BANYAN Σ Multivariate Bayesian Algorithm to Identify Members of Young Associations with 150 pc

    Science.gov (United States)

    Gagné, Jonathan; Mamajek, Eric E.; Malo, Lison; Riedel, Adric; Rodriguez, David; Lafrenière, David; Faherty, Jacqueline K.; Roy-Loubier, Olivier; Pueyo, Laurent; Robin, Annie C.; Doyon, René

    2018-03-01

    BANYAN Σ is a new Bayesian algorithm to identify members of young stellar associations within 150 pc of the Sun. It includes 27 young associations with ages in the range ∼1–800 Myr, modeled with multivariate Gaussians in six-dimensional (6D) XYZUVW space. It is the first such multi-association classification tool to include the nearest sub-groups of the Sco-Cen OB star-forming region, the IC 2602, IC 2391, Pleiades and Platais 8 clusters, and the ρ Ophiuchi, Corona Australis, and Taurus star formation regions. A model of field stars is built from a mixture of multivariate Gaussians based on the Besançon Galactic model. The algorithm can derive membership probabilities for objects with only sky coordinates and proper motion, but can also include parallax and radial velocity measurements, as well as spectrophotometric distance constraints from sequences in color–magnitude or spectral type–magnitude diagrams. BANYAN Σ benefits from an analytical solution to the Bayesian marginalization integrals over unknown radial velocities and distances that makes it more accurate and significantly faster than its predecessor BANYAN II. A contamination versus hit rate analysis is presented and demonstrates that BANYAN Σ achieves a better classification performance than other moving group tools available in the literature, especially in terms of cross-contamination between young associations. An updated list of bona fide members in the 27 young associations, augmented by the Gaia-DR1 release, as well as all parameters for the 6D multivariate Gaussian models for each association and the Galactic field neighborhood within 300 pc are presented. This new tool will make it possible to analyze large data sets such as the upcoming Gaia-DR2 to identify new young stars. IDL and Python versions of BANYAN Σ are made available with this publication, and a more limited online web tool is available at http://www.exoplanetes.umontreal.ca/banyan/banyansigma.php.

  5. Identifying individual fires from satellite-derived burned area data

    CSIR Research Space (South Africa)

    Archibald, S

    2009-07-01

    Full Text Available An algorithm for identifying individual fires from the Modis burned area data product is introduced for southern Africa. This algorithm gives the date of burning, size of fire, and location of the centroid for all fires identified over 8 years...

  6. MatureBayes: a probabilistic algorithm for identifying the mature miRNA within novel precursors.

    Directory of Open Access Journals (Sweden)

    Katerina Gkirtzou

    Full Text Available BACKGROUND: MicroRNAs (miRNAs are small, single stranded RNAs with a key role in post-transcriptional regulation of thousands of genes across numerous species. While several computational methods are currently available for identifying miRNA genes, accurate prediction of the mature miRNA remains a challenge. Existing approaches fall short in predicting the location of mature miRNAs but also in finding the functional strand(s of miRNA precursors. METHODOLOGY/PRINCIPAL FINDINGS: Here, we present a computational tool that incorporates a Naive Bayes classifier to identify mature miRNA candidates based on sequence and secondary structure information of their miRNA precursors. We take into account both positive (true mature miRNAs and negative (same-size non-mature miRNA sequences examples to optimize sensitivity as well as specificity. Our method can accurately predict the start position of experimentally verified mature miRNAs for both human and mouse, achieving a significantly larger (often double performance accuracy compared with two existing methods. Moreover, the method exhibits a very high generalization performance on miRNAs from two other organisms. More importantly, our method provides direct evidence about the features of miRNA precursors which may determine the location of the mature miRNA. We find that the triplet of positions 7, 8 and 9 from the mature miRNA end towards the closest hairpin have the largest discriminatory power, are relatively conserved in terms of sequence composition (mostly contain a Uracil and are located within or in very close proximity to the hairpin loop, suggesting the existence of a possible recognition site for Dicer and associated proteins. CONCLUSIONS: This work describes a novel algorithm for identifying the start position of mature miRNA(s produced by miRNA precursors. Our tool has significantly better (often double performance than two existing approaches and provides new insights about the potential use

  7. Results of Evolution Supervised by Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    Lorentz JÄNTSCHI

    2010-09-01

    Full Text Available The efficiency of a genetic algorithm is frequently assessed using a series of operators of evolution like crossover operators, mutation operators or other dynamic parameters. The present paper aimed to review the main results of evolution supervised by genetic algorithms used to identify solutions to agricultural and horticultural hard problems and to discuss the results of using a genetic algorithms on structure-activity relationships in terms of behavior of evolution supervised by genetic algorithms. A genetic algorithm had been developed and implemented in order to identify the optimal solution in term of estimation power of a multiple linear regression approach for structure-activity relationships. Three survival and three selection strategies (proportional, deterministic and tournament were investigated in order to identify the best survival-selection strategy able to lead to the model with higher estimation power. The Molecular Descriptors Family for structure characterization of a sample of 206 polychlorinated biphenyls with measured octanol-water partition coefficients was used as case study. Evolution using different selection and survival strategies proved to create populations of genotypes living in the evolution space with different diversity and variability. Under a series of criteria of comparisons these populations proved to be grouped and the groups were showed to be statistically different one to each other. The conclusions about genetic algorithm evolution according to a number of criteria were also highlighted.

  8. A new algorithm for identifying the flavour of $B_s^0$ mesons at LHCb

    CERN Document Server

    Aaij, Roel; Adeva, Bernardo; Adinolfi, Marco; Affolder, Anthony; Ajaltouni, Ziad; Akar, Simon; Albrecht, Johannes; Alessio, Federico; Alexander, Michael; Ali, Suvayu; Alkhazov, Georgy; Alvarez Cartelle, Paula; Alves Jr, Antonio Augusto; Amato, Sandra; Amerio, Silvia; Amhis, Yasmine; An, Liupan; Anderlini, Lucio; Andreassi, Guido; Andreotti, Mirco; Andrews, Jason; Appleby, Robert; Aquines Gutierrez, Osvaldo; Archilli, Flavio; d'Argent, Philippe; Artamonov, Alexander; Artuso, Marina; Aslanides, Elie; Auriemma, Giulio; Baalouch, Marouen; Bachmann, Sebastian; Back, John; Badalov, Alexey; Baesso, Clarissa; Baldini, Wander; Barlow, Roger; Barschel, Colin; Barsuk, Sergey; Barter, William; Batozskaya, Varvara; Battista, Vincenzo; Bay, Aurelio; Beaucourt, Leo; Beddow, John; Bedeschi, Franco; Bediaga, Ignacio; Bel, Lennaert; Bellee, Violaine; Belloli, Nicoletta; Belyaev, Ivan; Ben-Haim, Eli; Bencivenni, Giovanni; Benson, Sean; Benton, Jack; Berezhnoy, Alexander; Bernet, Roland; Bertolin, Alessandro; Betti, Federico; Bettler, Marc-Olivier; van Beuzekom, Martinus; Bifani, Simone; Billoir, Pierre; Bird, Thomas; Birnkraut, Alex; Bizzeti, Andrea; Blake, Thomas; Blanc, Frédéric; Blouw, Johan; Blusk, Steven; Bocci, Valerio; Bondar, Alexander; Bondar, Nikolay; Bonivento, Walter; Borgheresi, Alessio; Borghi, Silvia; Borisyak, Maxim; Borsato, Martino; Bowcock, Themistocles; Bowen, Espen Eie; Bozzi, Concezio; Braun, Svende; Britsch, Markward; Britton, Thomas; Brodzicka, Jolanta; Brook, Nicholas; Buchanan, Emma; Burr, Christopher; Bursche, Albert; Buytaert, Jan; Cadeddu, Sandro; Calabrese, Roberto; Calvi, Marta; Calvo Gomez, Miriam; Campana, Pierluigi; Campora Perez, Daniel; Capriotti, Lorenzo; Carbone, Angelo; Carboni, Giovanni; Cardinale, Roberta; Cardini, Alessandro; Carniti, Paolo; Carson, Laurence; Carvalho Akiba, Kazuyoshi; Casse, Gianluigi; Cassina, Lorenzo; Castillo Garcia, Lucia; Cattaneo, Marco; Cauet, Christophe; Cavallero, Giovanni; Cenci, Riccardo; Charles, Matthew; Charpentier, Philippe; Chatzikonstantinidis, Georgios; Chefdeville, Maximilien; Chen, Shanzhen; Cheung, Shu-Faye; Chiapolini, Nicola; Chrzaszcz, Marcin; Cid Vidal, Xabier; Ciezarek, Gregory; Clarke, Peter; Clemencic, Marco; Cliff, Harry; Closier, Joel; Coco, Victor; Cogan, Julien; Cogneras, Eric; Cogoni, Violetta; Cojocariu, Lucian; Collazuol, Gianmaria; Collins, Paula; Comerma-Montells, Albert; Contu, Andrea; Cook, Andrew; Coombes, Matthew; Coquereau, Samuel; Corti, Gloria; Corvo, Marco; Couturier, Benjamin; Cowan, Greig; Craik, Daniel Charles; Crocombe, Andrew; Cruz Torres, Melissa Maria; Cunliffe, Samuel; Currie, Robert; D'Ambrosio, Carmelo; Dall'Occo, Elena; Dalseno, Jeremy; David, Pieter; Davis, Adam; De Aguiar Francisco, Oscar; De Bruyn, Kristof; De Capua, Stefano; De Cian, Michel; De Miranda, Jussara; De Paula, Leandro; De Simone, Patrizia; Dean, Cameron Thomas; Decamp, Daniel; Deckenhoff, Mirko; Del Buono, Luigi; Déléage, Nicolas; Demmer, Moritz; Derkach, Denis; Deschamps, Olivier; Dettori, Francesco; Dey, Biplab; Di Canto, Angelo; Di Ruscio, Francesco; Dijkstra, Hans; Donleavy, Stephanie; Dordei, Francesca; Dorigo, Mirco; Dosil Suárez, Alvaro; Dovbnya, Anatoliy; Dreimanis, Karlis; Dufour, Laurent; Dujany, Giulio; Dungs, Kevin; Durante, Paolo; Dzhelyadin, Rustem; Dziurda, Agnieszka; Dzyuba, Alexey; Easo, Sajan; Egede, Ulrik; Egorychev, Victor; Eidelman, Semen; Eisenhardt, Stephan; Eitschberger, Ulrich; Ekelhof, Robert; Eklund, Lars; El Rifai, Ibrahim; Elsasser, Christian; Ely, Scott; Esen, Sevda; Evans, Hannah Mary; Evans, Timothy; Falabella, Antonio; Färber, Christian; Farley, Nathanael; Farry, Stephen; Fay, Robert; Fazzini, Davide; Ferguson, Dianne; Fernandez Albor, Victor; Ferrari, Fabio; Ferreira Rodrigues, Fernando; Ferro-Luzzi, Massimiliano; Filippov, Sergey; Fiore, Marco; Fiorini, Massimiliano; Firlej, Miroslaw; Fitzpatrick, Conor; Fiutowski, Tomasz; Fleuret, Frederic; Fohl, Klaus; Fol, Philip; Fontana, Marianna; Fontanelli, Flavio; Forshaw, Dean Charles; Forty, Roger; Frank, Markus; Frei, Christoph; Frosini, Maddalena; Fu, Jinlin; Furfaro, Emiliano; Gallas Torreira, Abraham; Galli, Domenico; Gallorini, Stefano; Gambetta, Silvia; Gandelman, Miriam; Gandini, Paolo; Gao, Yuanning; García Pardiñas, Julián; Garra Tico, Jordi; Garrido, Lluis; Gascon, David; Gaspar, Clara; Gavardi, Laura; Gazzoni, Giulio; Gerick, David; Gersabeck, Evelina; Gersabeck, Marco; Gershon, Timothy; Ghez, Philippe; Gianì, Sebastiana; Gibson, Valerie; Girard, Olivier Göran; Giubega, Lavinia-Helena; Gligorov, V.V.; Göbel, Carla; Golubkov, Dmitry; Golutvin, Andrey; Gomes, Alvaro; Gotti, Claudio; Grabalosa Gándara, Marc; Graciani Diaz, Ricardo; Granado Cardoso, Luis Alberto; Graugés, Eugeni; Graverini, Elena; Graziani, Giacomo; Grecu, Alexandru; Griffith, Peter; Grillo, Lucia; Grünberg, Oliver; Gui, Bin; Gushchin, Evgeny; Guz, Yury; Gys, Thierry; Hadavizadeh, Thomas; Hadjivasiliou, Christos; Haefeli, Guido; Haen, Christophe; Haines, Susan; Hall, Samuel; Hamilton, Brian; Han, Xiaoxue; Hansmann-Menzemer, Stephanie; Harnew, Neville; Harnew, Samuel; Harrison, Jonathan; He, Jibo; Head, Timothy; Heijne, Veerle; Heister, Arno; Hennessy, Karol; Henrard, Pierre; Henry, Louis; Hernando Morata, Jose Angel; van Herwijnen, Eric; Heß, Miriam; Hicheur, Adlène; Hill, Donal; Hoballah, Mostafa; Hombach, Christoph; Hulsbergen, Wouter; Humair, Thibaud; Hushchyn, Mikhail; Hussain, Nazim; Hutchcroft, David; Hynds, Daniel; Idzik, Marek; Ilten, Philip; Jacobsson, Richard; Jaeger, Andreas; Jalocha, Pawel; Jans, Eddy; Jawahery, Abolhassan; John, Malcolm; Johnson, Daniel; Jones, Christopher; Joram, Christian; Jost, Beat; Jurik, Nathan; Kandybei, Sergii; Kanso, Walaa; Karacson, Matthias; Karbach, Moritz; Karodia, Sarah; Kecke, Matthieu; Kelsey, Matthew; Kenyon, Ian; Kenzie, Matthew; Ketel, Tjeerd; Khairullin, Egor; Khanji, Basem; Khurewathanakul, Chitsanu; Kirn, Thomas; Klaver, Suzanne; Klimaszewski, Konrad; Kochebina, Olga; Kolpin, Michael; Komarov, Ilya; Koopman, Rose; Koppenburg, Patrick; Kozeiha, Mohamad; Kravchuk, Leonid; Kreplin, Katharina; Kreps, Michal; Krocker, Georg; Krokovny, Pavel; Kruse, Florian; Krzemien, Wojciech; Kucewicz, Wojciech; Kucharczyk, Marcin; Kudryavtsev, Vasily; Kuonen, Axel Kevin; Kurek, Krzysztof; Kvaratskheliya, Tengiz; Lacarrere, Daniel; Lafferty, George; Lai, Adriano; Lambert, Dean; Lanfranchi, Gaia; Langenbruch, Christoph; Langhans, Benedikt; Latham, Thomas; Lazzeroni, Cristina; Le Gac, Renaud; van Leerdam, Jeroen; Lees, Jean-Pierre; Lefèvre, Regis; Leflat, Alexander; Lefrançois, Jacques; Lemos Cid, Edgar; Leroy, Olivier; Lesiak, Tadeusz; Leverington, Blake; Li, Yiming; Likhomanenko, Tatiana; Liles, Myfanwy; Lindner, Rolf; Linn, Christian; Lionetto, Federica; Liu, Bo; Liu, Xuesong; Loh, David; Longstaff, Iain; Lopes, Jose; Lucchesi, Donatella; Lucio Martinez, Miriam; Luo, Haofei; Lupato, Anna; Luppi, Eleonora; Lupton, Oliver; Lusardi, Nicola; Lusiani, Alberto; Machefert, Frederic; Maciuc, Florin; Maev, Oleg; Maguire, Kevin; Malde, Sneha; Malinin, Alexander; Manca, Giulia; Mancinelli, Giampiero; Manning, Peter Michael; Mapelli, Alessandro; Maratas, Jan; Marchand, Jean François; Marconi, Umberto; Marin Benito, Carla; Marino, Pietro; Marks, Jörg; Martellotti, Giuseppe; Martin, Morgan; Martinelli, Maurizio; Martinez Santos, Diego; Martinez Vidal, Fernando; Martins Tostes, Danielle; Massacrier, Laure Marie; Massafferri, André; Matev, Rosen; Mathad, Abhijit; Mathe, Zoltan; Matteuzzi, Clara; Mauri, Andrea; Maurin, Brice; Mazurov, Alexander; McCann, Michael; McCarthy, James; McNab, Andrew; McNulty, Ronan; Meadows, Brian; Meier, Frank; Meissner, Marco; Melnychuk, Dmytro; Merk, Marcel; Merli, Andrea; Michielin, Emanuele; Milanes, Diego Alejandro; Minard, Marie-Noelle; Mitzel, Dominik Stefan; Molina Rodriguez, Josue; Monroy, Ignacio Alberto; Monteil, Stephane; Morandin, Mauro; Morawski, Piotr; Mordà, Alessandro; Morello, Michael Joseph; Moron, Jakub; Morris, Adam Benjamin; Mountain, Raymond; Muheim, Franz; Müller, Dominik; Müller, Janine; Müller, Katharina; Müller, Vanessa; Mussini, Manuel; Muster, Bastien; Naik, Paras; Nakada, Tatsuya; Nandakumar, Raja; Nandi, Anita; Nasteva, Irina; Needham, Matthew; Neri, Nicola; Neubert, Sebastian; Neufeld, Niko; Neuner, Max; Nguyen, Anh Duc; Nguyen-Mau, Chung; Niess, Valentin; Nieswand, Simon; Niet, Ramon; Nikitin, Nikolay; Nikodem, Thomas; Novoselov, Alexey; O'Hanlon, Daniel Patrick; Oblakowska-Mucha, Agnieszka; Obraztsov, Vladimir; Ogilvy, Stephen; Okhrimenko, Oleksandr; Oldeman, Rudolf; Onderwater, Gerco; Osorio Rodrigues, Bruno; Otalora Goicochea, Juan Martin; Otto, Adam; Owen, Patrick; Oyanguren, Maria Aranzazu; Palano, Antimo; Palombo, Fernando; Palutan, Matteo; Panman, Jacob; Papanestis, Antonios; Pappagallo, Marco; Pappalardo, Luciano; Pappenheimer, Cheryl; Parker, William; Parkes, Christopher; Passaleva, Giovanni; Patel, Girish; Patel, Mitesh; Patrignani, Claudia; Pearce, Alex; Pellegrino, Antonio; Penso, Gianni; Pepe Altarelli, Monica; Perazzini, Stefano; Perret, Pascal; Pescatore, Luca; Petridis, Konstantinos; Petrolini, Alessandro; Petruzzo, Marco; Picatoste Olloqui, Eduardo; Pietrzyk, Boleslaw; Pikies, Malgorzata; Pinci, Davide; Pistone, Alessandro; Piucci, Alessio; Playfer, Stephen; Plo Casasus, Maximo; Poikela, Tuomas; Polci, Francesco; Poluektov, Anton; Polyakov, Ivan; Polycarpo, Erica; Popov, Alexander; Popov, Dmitry; Popovici, Bogdan; Potterat, Cédric; Price, Eugenia; Price, Joseph David; Prisciandaro, Jessica; Pritchard, Adrian; Prouve, Claire; Pugatch, Valery; Puig Navarro, Albert; Punzi, Giovanni; Qian, Wenbin; Quagliani, Renato; Rachwal, Bartolomiej; Rademacker, Jonas; Rama, Matteo; Ramos Pernas, Miguel; Rangel, Murilo; Raniuk, Iurii; Raven, Gerhard; Redi, Federico; Reichert, Stefanie; dos Reis, Alberto; Renaudin, Victor; Ricciardi, Stefania; Richards, Sophie; Rihl, Mariana; Rinnert, Kurt; Rives Molina, Vincente; Robbe, Patrick; Rodrigues, Ana Barbara; Rodrigues, Eduardo; Rodriguez Lopez, Jairo Alexis; Rodriguez Perez, Pablo; Rogozhnikov, Alexey; Roiser, Stefan; Romanovsky, Vladimir; Romero Vidal, Antonio; Ronayne, John William; Rotondo, Marcello; Ruf, Thomas; Ruiz Valls, Pablo; Saborido Silva, Juan Jose; Sagidova, Naylya; Saitta, Biagio; Salustino Guimaraes, Valdir; Sanchez Mayordomo, Carlos; Sanmartin Sedes, Brais; Santacesaria, Roberta; Santamarina Rios, Cibran; Santimaria, Marco; Santovetti, Emanuele; Sarti, Alessio; Satriano, Celestina; Satta, Alessia; Saunders, Daniel Martin; Savrina, Darya; Schael, Stefan; Schiller, Manuel; Schindler, Heinrich; Schlupp, Maximilian; Schmelling, Michael; Schmelzer, Timon; Schmidt, Burkhard; Schneider, Olivier; Schopper, Andreas; Schubiger, Maxime; Schune, Marie Helene; Schwemmer, Rainer; Sciascia, Barbara; Sciubba, Adalberto; Semennikov, Alexander; Serra, Nicola; Serrano, Justine; Sestini, Lorenzo; Seyfert, Paul; Shapkin, Mikhail; Shapoval, Illya; Shcheglov, Yury; Shears, Tara; Shekhtman, Lev; Shevchenko, Vladimir; Shires, Alexander; Siddi, Benedetto Gianluca; Silva Coutinho, Rafael; Silva de Oliveira, Luiz Gustavo; Simi, Gabriele; Sirendi, Marek; Skidmore, Nicola; Skwarnicki, Tomasz; Smith, Eluned; Smith, Iwan Thomas; Smith, Jackson; Smith, Mark; Snoek, Hella; Sokoloff, Michael; Soler, Paul; Soomro, Fatima; Souza, Daniel; Souza De Paula, Bruno; Spaan, Bernhard; Spradlin, Patrick; Sridharan, Srikanth; Stagni, Federico; Stahl, Marian; Stahl, Sascha; Stefkova, Slavomira; Steinkamp, Olaf; Stenyakin, Oleg; Stevenson, Scott; Stoica, Sabin; Stone, Sheldon; Storaci, Barbara; Stracka, Simone; Straticiuc, Mihai; Straumann, Ulrich; Sun, Liang; Sutcliffe, William; Swientek, Krzysztof; Swientek, Stefan; Syropoulos, Vasileios; Szczekowski, Marek; Szumlak, Tomasz; T'Jampens, Stephane; Tayduganov, Andrey; Tekampe, Tobias; Tellarini, Giulia; Teubert, Frederic; Thomas, Christopher; Thomas, Eric; van Tilburg, Jeroen; Tisserand, Vincent; Tobin, Mark; Todd, Jacob; Tolk, Siim; Tomassetti, Luca; Tonelli, Diego; Topp-Joergensen, Stig; Tournefier, Edwige; Tourneur, Stephane; Trabelsi, Karim; Traill, Murdo; Tran, Minh Tâm; Tresch, Marco; Trisovic, Ana; Tsaregorodtsev, Andrei; Tsopelas, Panagiotis; Tuning, Niels; Ukleja, Artur; Ustyuzhanin, Andrey; Uwer, Ulrich; Vacca, Claudia; Vagnoni, Vincenzo; Valenti, Giovanni; Vallier, Alexis; Vazquez Gomez, Ricardo; Vazquez Regueiro, Pablo; Vázquez Sierra, Carlos; Vecchi, Stefania; van Veghel, Maarten; Velthuis, Jaap; Veltri, Michele; Veneziano, Giovanni; Vesterinen, Mika; Viaud, Benoit; Vieira, Daniel; Vieites Diaz, Maria; Vilasis-Cardona, Xavier; Volkov, Vladimir; Vollhardt, Achim; Voong, David; Vorobyev, Alexey; Vorobyev, Vitaly; Voß, Christian; de Vries, Jacco; Waldi, Roland; Wallace, Charlotte; Wallace, Ronan; Walsh, John; Wang, Jianchun; Ward, David; Watson, Nigel; Websdale, David; Weiden, Andreas; Whitehead, Mark; Wicht, Jean; Wilkinson, Guy; Wilkinson, Michael; Williams, Mark Richard James; Williams, Matthew; Williams, Mike; Williams, Timothy; Wilson, Fergus; Wimberley, Jack; Wishahi, Julian; Wislicki, Wojciech; Witek, Mariusz; Wormser, Guy; Wotton, Stephen; Wraight, Kenneth; Wright, Simon; Wyllie, Kenneth; Xie, Yuehong; Xu, Zhirui; Yang, Zhenwei; Yu, Jiesheng; Yuan, Xuhao; Yushchenko, Oleg; Zangoli, Maria; Zavertyaev, Mikhail; Zhang, Liming; Zhang, Yanxi; Zhelezov, Alexey; Zhokhov, Anatoly; Zhong, Liang; Zhukov, Valery; Zucchelli, Stefano

    2016-05-17

    A new algorithm for the determination of the initial flavour of $B_s^0$ mesons is presented. The algorithm is based on two neural networks and exploits the $b$ hadron production mechanism at a hadron collider. The first network is trained to select charged kaons produced in association with the $B_s^0$ meson. The second network combines the kaon charges to assign the $B_s^0$ flavour and estimates the probability of a wrong assignment. The algorithm is calibrated using data corresponding to an integrated luminosity of 3 fb$^{-1}$ collected by the LHCb experiment in proton-proton collisions at 7 and 8 TeV centre-of-mass energies. The calibration is performed in two ways: by resolving the $B_s^0$-$\\bar{B}_s^0$ flavour oscillations in $B_s^0 \\to D_s^- \\pi^+$ decays, and by analysing flavour-specific $B_{s 2}^{*}(5840)^0 \\to B^+ K^-$ decays. The tagging power measured in $B_s^0 \\to D_s^- \\pi^+$ decays is found to be $(1.80 \\pm 0.19({\\rm stat}) \\pm 0.18({\\rm syst}))$\\%, which is an improvement of about 50\\% compare...

  9. 2nd International Conference on Harmony Search Algorithm

    CERN Document Server

    Geem, Zong

    2016-01-01

    The Harmony Search Algorithm (HSA) is one of the most well-known techniques in the field of soft computing, an important paradigm in the science and engineering community.  This volume, the proceedings of the 2nd International Conference on Harmony Search Algorithm 2015 (ICHSA 2015), brings together contributions describing the latest developments in the field of soft computing with a special focus on HSA techniques. It includes coverage of new methods that have potentially immense application in various fields. Contributed articles cover aspects of the following topics related to the Harmony Search Algorithm: analytical studies; improved, hybrid and multi-objective variants; parameter tuning; and large-scale applications.  The book also contains papers discussing recent advances on the following topics: genetic algorithms; evolutionary strategies; the firefly algorithm and cuckoo search; particle swarm optimization and ant colony optimization; simulated annealing; and local search techniques.   This book ...

  10. Identifying Value Indicators and Social Capital in Community Health Partnerships

    Science.gov (United States)

    Hausman, Alice J.; Becker, Julie; Brawer, Rickie

    2005-01-01

    Increasingly, public health practice is turning to the application of community collaborative models to improve population health status. Despite the growth of these activities, however, evaluations of the national demonstrations have indicated that community health partnerships fail to achieve measurable results and struggle to maintain integrity…

  11. An Algorithmic Diversity Diet?

    DEFF Research Database (Denmark)

    Sørensen, Jannick Kirk; Schmidt, Jan-Hinrik

    2016-01-01

    With the growing influence of personalized algorithmic recommender systems on the exposure of media content to users, the relevance of discussing the diversity of recommendations increases, particularly as far as public service media (PSM) is concerned. An imagined implementation of a diversity...... diet system however triggers not only the classic discussion of the reach – distinctiveness balance for PSM, but also shows that ‘diversity’ is understood very differently in algorithmic recommender system communities than it is editorially and politically in the context of PSM. The design...... of a diversity diet system generates questions not just about editorial power, personal freedom and techno-paternalism, but also about the embedded politics of recommender systems as well as the human skills affiliated with PSM editorial work and the nature of PSM content....

  12. A rapid two-step algorithm detects and identifies clinical macrolide and beta-lactam antibiotic resistance in clinical bacterial isolates.

    Science.gov (United States)

    Lu, Xuedong; Nie, Shuping; Xia, Chengjing; Huang, Lie; He, Ying; Wu, Runxiang; Zhang, Li

    2014-07-01

    Aiming to identify macrolide and beta-lactam resistance in clinical bacterial isolates rapidly and accurately, a two-step algorithm was developed based on detection of eight antibiotic resistance genes. Targeting at genes linked to bacterial macrolide (msrA, ermA, ermB, and ermC) and beta-lactam (blaTEM, blaSHV, blaCTX-M-1, blaCTX-M-9) antibiotic resistances, this method includes a multiplex real-time PCR, a melting temperature profile analysis as well as a liquid bead microarray assay. Liquid bead microarray assay is applied only when indistinguishable Tm profile is observed. The clinical validity of this method was assessed on clinical bacterial isolates. Among the total 580 isolates that were determined by our diagnostic method, 75% of them were identified by the multiplex real-time PCR with melting temperature analysis alone, while the remaining 25% required both multiplex real-time PCR with melting temperature analysis and liquid bead microarray assay for identification. Compared with the traditional phenotypic antibiotic susceptibility test, an overall agreement of 81.2% (kappa=0.614, 95% CI=0.550-0.679) was observed, with a sensitivity and specificity of 87.7% and 73% respectively. Besides, the average test turnaround time is 3.9h, which is much shorter in comparison with more than 24h for the traditional phenotypic tests. Having the advantages of the shorter operating time and comparable high sensitivity and specificity with the traditional phenotypic test, our two-step algorithm provides an efficient tool for rapid determination of macrolide and beta-lactam antibiotic resistances in clinical bacterial isolates. Copyright © 2014 Elsevier B.V. All rights reserved.

  13. Identifying a borderline personality disorder prodrome: Implications for community screening.

    Science.gov (United States)

    Stepp, Stephanie D; Lazarus, Sophie A

    2017-08-01

    Elucidating early signs and symptoms of borderline personality disorder (BPD) has important implications for screening and identifying youth appropriate for early intervention. The purpose of this study was to identify dimensions of child temperament and psychopathology symptom severity that predict conversion to a positive screen for BPD over a 14-year follow-up period in a large, urban community sample of girls (n = 2 450). Parent and teacher reports of child temperament and psychopathology symptom severity assessed when girls were ages 5-8 years were examined as predictors of new-onset BPD cases when girls were ages 14-22 years. In the final model, parent and teacher ratings of emotionality remained significant predictors of new-onset BPD. Additionally, parent ratings of hyperactivity/impulsivity and depression severity, as well as teacher ratings of inattention severity, were also predictive. Results also revealed that elevations in these dimensions pose a notable increase in risk for conversion to BPD over the follow-up period. Supplementary analyses revealed that with the exception of parent-reported depression severity, these same predictors were associated with increases in BPD symptom severity over the follow-up period. These findings suggest BPD onset in adolescence and early adulthood can be detected from parent and teacher reports of temperament and symptom severity dimensions assessed in childhood. The identification of this prodrome holds promise for advancing early detection of children at risk prior to the development of the full-blown disorder. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.

  14. Learning from nature: Nature-inspired algorithms

    DEFF Research Database (Denmark)

    Albeanu, Grigore; Madsen, Henrik; Popentiu-Vladicescu, Florin

    2016-01-01

    .), genetic and evolutionary strategies, artificial immune systems etc. Well-known examples of applications include: aircraft wing design, wind turbine design, bionic car, bullet train, optimal decisions related to traffic, appropriate strategies to survive under a well-adapted immune system etc. Based......During last decade, the nature has inspired researchers to develop new algorithms. The largest collection of nature-inspired algorithms is biology-inspired: swarm intelligence (particle swarm optimization, ant colony optimization, cuckoo search, bees' algorithm, bat algorithm, firefly algorithm etc...... on collective social behaviour of organisms, researchers have developed optimization strategies taking into account not only the individuals, but also groups and environment. However, learning from nature, new classes of approaches can be identified, tested and compared against already available algorithms...

  15. Spatial correlation analysis of urban traffic state under a perspective of community detection

    Science.gov (United States)

    Yang, Yanfang; Cao, Jiandong; Qin, Yong; Jia, Limin; Dong, Honghui; Zhang, Aomuhan

    2018-05-01

    Understanding the spatial correlation of urban traffic state is essential for identifying the evolution patterns of urban traffic state. However, the distribution of traffic state always has characteristics of large spatial span and heterogeneity. This paper adapts the concept of community detection to the correlation network of urban traffic state and proposes a new perspective to identify the spatial correlation patterns of traffic state. In the proposed urban traffic network, the nodes represent road segments, and an edge between a pair of nodes is added depending on the result of significance test for the corresponding correlation of traffic state. Further, the process of community detection in the urban traffic network (named GWPA-K-means) is applied to analyze the spatial dependency of traffic state. The proposed method extends the traditional K-means algorithm in two steps: (i) redefines the initial cluster centers by two properties of nodes (the GWPA value and the minimum shortest path length); (ii) utilizes the weight signal propagation process to transfer the topological information of the urban traffic network into a node similarity matrix. Finally, numerical experiments are conducted on a simple network and a real urban road network in Beijing. The results show that GWPA-K-means algorithm is valid in spatial correlation analysis of traffic state. The network science and community structure analysis perform well in describing the spatial heterogeneity of traffic state on a large spatial scale.

  16. A Comparative Evaluation of Unsupervised Anomaly Detection Algorithms for Multivariate Data

    Science.gov (United States)

    Goldstein, Markus; Uchida, Seiichi

    2016-01-01

    Anomaly detection is the process of identifying unexpected items or events in datasets, which differ from the norm. In contrast to standard classification tasks, anomaly detection is often applied on unlabeled data, taking only the internal structure of the dataset into account. This challenge is known as unsupervised anomaly detection and is addressed in many practical applications, for example in network intrusion detection, fraud detection as well as in the life science and medical domain. Dozens of algorithms have been proposed in this area, but unfortunately the research community still lacks a comparative universal evaluation as well as common publicly available datasets. These shortcomings are addressed in this study, where 19 different unsupervised anomaly detection algorithms are evaluated on 10 different datasets from multiple application domains. By publishing the source code and the datasets, this paper aims to be a new well-funded basis for unsupervised anomaly detection research. Additionally, this evaluation reveals the strengths and weaknesses of the different approaches for the first time. Besides the anomaly detection performance, computational effort, the impact of parameter settings as well as the global/local anomaly detection behavior is outlined. As a conclusion, we give an advise on algorithm selection for typical real-world tasks. PMID:27093601

  17. Categorizing segmentation quality using a quantitative quality assurance algorithm

    International Nuclear Information System (INIS)

    Rodrigues, George; Louie, Alexander; Best, Lara

    2012-01-01

    Obtaining high levels of contouring consistency is a major limiting step in optimizing the radiotherapeutic ratio. We describe a novel quantitative methodology for the quality assurance (QA) of contour compliance referenced against a community set of contouring experts. Two clinical tumour site scenarios (10 lung cases and one prostate case) were used with QA algorithm. For each case, multiple physicians (lung: n = 6, prostate: n = 25) segmented various target/organ at risk (OAR) structures to define a set of community reference contours. For each set of community contours, a consensus contour (Simultaneous Truth and Performance Level Estimation (STAPLE)) was created. Differences between each individual community contour versus the group consensus contour were quantified by consensus-based contouring penalty metric (PM) scores. New observers segmented these same cases to calculate individual PM scores (for each unique target/OAR) for each new observer–STAPLE pair for comparison against the community and consensus contours. Four physicians contoured the 10 lung cases for a total of 72 contours for quality assurance evaluation against the previously derived community consensus contours. A total of 16 outlier contours were identified by the QA system of which 11 outliers were due to over-contouring discrepancies, three were due to over-/under-contouring discrepancies, and two were due to missing/incorrect nodal contours. In the prostate scenario involving six physicians, the QA system detected a missing penile bulb contour, systematic inner-bladder contouring, and under-contouring of the upper/anterior rectum. A practical methodology for QA has been demonstrated with future clinical trial credentialing, medical education and auto-contouring assessment applications.

  18. Finding needles in a haystack: a methodology for identifying and sampling community-based youth smoking cessation programs.

    Science.gov (United States)

    Emery, Sherry; Lee, Jungwha; Curry, Susan J; Johnson, Tim; Sporer, Amy K; Mermelstein, Robin; Flay, Brian; Warnecke, Richard

    2010-02-01

    Surveys of community-based programs are difficult to conduct when there is virtually no information about the number or locations of the programs of interest. This article describes the methodology used by the Helping Young Smokers Quit (HYSQ) initiative to identify and profile community-based youth smoking cessation programs in the absence of a defined sample frame. We developed a two-stage sampling design, with counties as the first-stage probability sampling units. The second stage used snowball sampling to saturation, to identify individuals who administered youth smoking cessation programs across three economic sectors in each county. Multivariate analyses modeled the relationship between program screening, eligibility, and response rates and economic sector and stratification criteria. Cumulative logit models analyzed the relationship between the number of contacts in a county and the number of programs screened, eligible, or profiled in a county. The snowball process yielded 9,983 unique and traceable contacts. Urban and high-income counties yielded significantly more screened program administrators; urban counties produced significantly more eligible programs, but there was no significant association between the county characteristics and program response rate. There is a positive relationship between the number of informants initially located and the number of programs screened, eligible, and profiled in a county. Our strategy to identify youth tobacco cessation programs could be used to create a sample frame for other nonprofit organizations that are difficult to identify due to a lack of existing directories, lists, or other traditional sample frames.

  19. Automated Means of Identifying Landslide Deposits using LiDAR Data using the Contour Connection Method Algorithm

    Science.gov (United States)

    Olsen, M. J.; Leshchinsky, B. A.; Tanyu, B. F.

    2014-12-01

    Landslides are a global natural hazard, resulting in severe economic, environmental and social impacts every year. Often, landslides occur in areas of repeated slope instability, but despite these trends, significant residential developments and critical infrastructure are built in the shadow of past landslide deposits and marginally stable slopes. These hazards, despite their sometimes enormous scale and regional propensity, however, are difficult to detect on the ground, often due to vegetative cover. However, new developments in remote sensing technology, specifically Light Detection and Ranging mapping (LiDAR) are providing a new means of viewing our landscape. Airborne LiDAR, combined with a level of post-processing, enable the creation of spatial data representative of the earth beneath the vegetation, highlighting the scars of unstable slopes of the past. This tool presents a revolutionary technique to mapping landslide deposits and their associated regions of risk; yet, their inventorying is often done manually, an approach that can be tedious, time-consuming and subjective. However, the associated LiDAR bare earth data present the opportunity to use this remote sensing technology and typical landslide geometry to create an automated algorithm that can detect and inventory deposits on a landscape scale. This algorithm, called the Contour Connection Method (CCM), functions by first detecting steep gradients, often associated with the headscarp of a failed hillslope, and initiating a search, highlighting deposits downslope of the failure. Based on input of search gradients, CCM can assist in highlighting regions identified as landslides consistently on a landscape scale, capable of mapping more than 14,000 hectares rapidly (help better define these regions of risk.

  20. The LabelHash algorithm for substructure matching

    Directory of Open Access Journals (Sweden)

    Bryant Drew H

    2010-11-01

    Full Text Available Abstract Background There is an increasing number of proteins with known structure but unknown function. Determining their function would have a significant impact on understanding diseases and designing new therapeutics. However, experimental protein function determination is expensive and very time-consuming. Computational methods can facilitate function determination by identifying proteins that have high structural and chemical similarity. Results We present LabelHash, a novel algorithm for matching substructural motifs to large collections of protein structures. The algorithm consists of two phases. In the first phase the proteins are preprocessed in a fashion that allows for instant lookup of partial matches to any motif. In the second phase, partial matches for a given motif are expanded to complete matches. The general applicability of the algorithm is demonstrated with three different case studies. First, we show that we can accurately identify members of the enolase superfamily with a single motif. Next, we demonstrate how LabelHash can complement SOIPPA, an algorithm for motif identification and pairwise substructure alignment. Finally, a large collection of Catalytic Site Atlas motifs is used to benchmark the performance of the algorithm. LabelHash runs very efficiently in parallel; matching a motif against all proteins in the 95% sequence identity filtered non-redundant Protein Data Bank typically takes no more than a few minutes. The LabelHash algorithm is available through a web server and as a suite of standalone programs at http://labelhash.kavrakilab.org. The output of the LabelHash algorithm can be further analyzed with Chimera through a plugin that we developed for this purpose. Conclusions LabelHash is an efficient, versatile algorithm for large-scale substructure matching. When LabelHash is running in parallel, motifs can typically be matched against the entire PDB on the order of minutes. The algorithm is able to identify

  1. Identifying the Entrepreneurship Characteristics of the Oil Palm Community Plantation Farmers in the Riau Area

    OpenAIRE

    Brilliant Asmit; Deddy P. Koesrindartoto

    2015-01-01

    Oil palm is an essential and strategic commodity in the Riau area because of its considerable role in supporting the peoples’ economy, especially for plantation farmers. Oil palm plantation activities have brought economic impacts to society there, both for the people who are directly involved with the plantations and for their surrounding communities. This regional advantage is a facility for farmers to be able to develop their farms as plantations. The aims of this research are to identify ...

  2. Efficient motif finding algorithms for large-alphabet inputs

    Directory of Open Access Journals (Sweden)

    Pavlovic Vladimir

    2010-10-01

    Full Text Available Abstract Background We consider the problem of identifying motifs, recurring or conserved patterns, in the biological sequence data sets. To solve this task, we present a new deterministic algorithm for finding patterns that are embedded as exact or inexact instances in all or most of the input strings. Results The proposed algorithm (1 improves search efficiency compared to existing algorithms, and (2 scales well with the size of alphabet. On a synthetic planted DNA motif finding problem our algorithm is over 10× more efficient than MITRA, PMSPrune, and RISOTTO for long motifs. Improvements are orders of magnitude higher in the same setting with large alphabets. On benchmark TF-binding site problems (FNP, CRP, LexA we observed reduction in running time of over 12×, with high detection accuracy. The algorithm was also successful in rapidly identifying protein motifs in Lipocalin, Zinc metallopeptidase, and supersecondary structure motifs for Cadherin and Immunoglobin families. Conclusions Our algorithm reduces computational complexity of the current motif finding algorithms and demonstrate strong running time improvements over existing exact algorithms, especially in important and difficult cases of large-alphabet sequences.

  3. Multisensor data fusion algorithm development

    Energy Technology Data Exchange (ETDEWEB)

    Yocky, D.A.; Chadwick, M.D.; Goudy, S.P.; Johnson, D.K.

    1995-12-01

    This report presents a two-year LDRD research effort into multisensor data fusion. We approached the problem by addressing the available types of data, preprocessing that data, and developing fusion algorithms using that data. The report reflects these three distinct areas. First, the possible data sets for fusion are identified. Second, automated registration techniques for imagery data are analyzed. Third, two fusion techniques are presented. The first fusion algorithm is based on the two-dimensional discrete wavelet transform. Using test images, the wavelet algorithm is compared against intensity modulation and intensity-hue-saturation image fusion algorithms that are available in commercial software. The wavelet approach outperforms the other two fusion techniques by preserving spectral/spatial information more precisely. The wavelet fusion algorithm was also applied to Landsat Thematic Mapper and SPOT panchromatic imagery data. The second algorithm is based on a linear-regression technique. We analyzed the technique using the same Landsat and SPOT data.

  4. Robotics, vision and control fundamental algorithms in Matlab

    CERN Document Server

    Corke, Peter

    2017-01-01

    Robotic vision, the combination of robotics and computer vision, involves the application of computer algorithms to data acquired from sensors. The research community has developed a large body of such algorithms but for a newcomer to the field this can be quite daunting. For over 20 years the author has maintained two open-source MATLAB® Toolboxes, one for robotics and one for vision. They provide implementations of many important algorithms and allow users to work with real problems, not just trivial examples. This book makes the fundamental algorithms of robotics, vision and control accessible to all. It weaves together theory, algorithms and examples in a narrative that covers robotics and computer vision separately and together. Using the latest versions of the Toolboxes the author shows how complex problems can be decomposed and solved using just a few simple lines of code. The topics covered are guided by real problems observed by the author over many years as a practitioner of both robotics and compu...

  5. Malicious Cognitive User Identification Algorithm in Centralized Spectrum Sensing System

    Directory of Open Access Journals (Sweden)

    Jingbo Zhang

    2017-11-01

    Full Text Available Collaborative spectral sensing can fuse the perceived results of multiple cognitive users, and thus will improve the accuracy of perceived results. However, the multi-source features of the perceived results result in security problems in the system. When there is a high probability of a malicious user attack, the traditional algorithm can correctly identify the malicious users. However, when the probability of attack by malicious users is reduced, it is almost impossible to use the traditional algorithm to correctly distinguish between honest users and malicious users, which greatly reduces the perceived performance. To address the problem above, based on the β function and the feedback iteration mathematical method, this paper proposes a malicious user identification algorithm under multi-channel cooperative conditions (β-MIAMC, which involves comprehensively assessing the cognitive user’s performance on multiple sub-channels to identify the malicious user. Simulation results show under the same attack probability, compared with the traditional algorithm, the β-MIAMC algorithm can more accurately identify the malicious users, reducing the false alarm probability of malicious users by more than 20%. When the attack probability is greater than 7%, the proposed algorithm can identify the malicious users with 100% certainty.

  6. Bouc–Wen hysteresis model identification using Modified Firefly Algorithm

    International Nuclear Information System (INIS)

    Zaman, Mohammad Asif; Sikder, Urmita

    2015-01-01

    The parameters of Bouc–Wen hysteresis model are identified using a Modified Firefly Algorithm. The proposed algorithm uses dynamic process control parameters to improve its performance. The algorithm is used to find the model parameter values that results in the least amount of error between a set of given data points and points obtained from the Bouc–Wen model. The performance of the algorithm is compared with the performance of conventional Firefly Algorithm, Genetic Algorithm and Differential Evolution algorithm in terms of convergence rate and accuracy. Compared to the other three optimization algorithms, the proposed algorithm is found to have good convergence rate with high degree of accuracy in identifying Bouc–Wen model parameters. Finally, the proposed method is used to find the Bouc–Wen model parameters from experimental data. The obtained model is found to be in good agreement with measured data. - Highlights: • We describe a new method to find the Bouc–Wen hysteresis model parameters. • We propose a Modified Firefly Algorithm. • We compare our method with existing methods to find that the proposed method performs better. • We use our model to fit experimental results. Good agreement is found

  7. Bouc–Wen hysteresis model identification using Modified Firefly Algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Zaman, Mohammad Asif, E-mail: zaman@stanford.edu [Department of Electrical Engineering, Stanford University (United States); Sikder, Urmita [Department of Electrical Engineering and Computer Sciences, University of California, Berkeley (United States)

    2015-12-01

    The parameters of Bouc–Wen hysteresis model are identified using a Modified Firefly Algorithm. The proposed algorithm uses dynamic process control parameters to improve its performance. The algorithm is used to find the model parameter values that results in the least amount of error between a set of given data points and points obtained from the Bouc–Wen model. The performance of the algorithm is compared with the performance of conventional Firefly Algorithm, Genetic Algorithm and Differential Evolution algorithm in terms of convergence rate and accuracy. Compared to the other three optimization algorithms, the proposed algorithm is found to have good convergence rate with high degree of accuracy in identifying Bouc–Wen model parameters. Finally, the proposed method is used to find the Bouc–Wen model parameters from experimental data. The obtained model is found to be in good agreement with measured data. - Highlights: • We describe a new method to find the Bouc–Wen hysteresis model parameters. • We propose a Modified Firefly Algorithm. • We compare our method with existing methods to find that the proposed method performs better. • We use our model to fit experimental results. Good agreement is found.

  8. Named Entity Linking Algorithm

    Directory of Open Access Journals (Sweden)

    M. F. Panteleev

    2017-01-01

    Full Text Available In the tasks of processing text in natural language, Named Entity Linking (NEL represents the task to define and link some entity, which is found in the text, with some entity in the knowledge base (for example, Dbpedia. Currently, there is a diversity of approaches to solve this problem, but two main classes can be identified: graph-based approaches and machine learning-based ones. Graph and Machine Learning approaches-based algorithm is proposed accordingly to the stated assumptions about the interrelations of named entities in a sentence and in general.In the case of graph-based approaches, it is necessary to solve the problem of identifying an optimal set of the related entities according to some metric that characterizes the distance between these entities in a graph built on some knowledge base. Due to limitations in processing power, to solve this task directly is impossible. Therefore, its modification is proposed. Based on the algorithms of machine learning, an independent solution cannot be built due to small volumes of training datasets relevant to NEL task. However, their use can contribute to improving the quality of the algorithm. The adaptation of the Latent Dirichlet Allocation model is proposed in order to obtain a measure of the compatibility of attributes of various entities encountered in one context.The efficiency of the proposed algorithm was experimentally tested. A test dataset was independently generated. On its basis the performance of the model was compared using the proposed algorithm with the open source product DBpedia Spotlight, which solves the NEL problem.The mockup, based on the proposed algorithm, showed a low speed as compared to DBpedia Spotlight. However, the fact that it has shown higher accuracy, stipulates the prospects for work in this direction.The main directions of development were proposed in order to increase the accuracy of the system and its productivity.

  9. Heat transfer analysis of unsteady graphene oxide nanofluid flow using a fuzzy identifier evolved by genetically encoded mutable smart bee algorithm

    Directory of Open Access Journals (Sweden)

    Mohammadreza Azimi

    2015-03-01

    Full Text Available In the current research, the unsteady two dimensional Graphene Oxide water based nanofluid heat transfer between two moving parallel plates is analyzed using an intelligent black-box identifier. The developed intelligent tool is known as evolvable evolutionary fuzzy inference system (EE-FIS which is based on the integration of low-level fuzzy programming and hyper-level evolutionary computing concepts. Here, the authors propose the use of a modified evolutionary algorithm (EA which is called hybrid genetic mutable smart bee algorithm (HGMSBA. The proposed HGMSBA is used to evolve both antecedent and consequent parts of fuzzy rule base. Besides, it tries to prune the rule base of fuzzy inference system (FIS to decrease its computational complexity and increase its interpretability. By considering the prediction error of the fuzzy identifier as the objective function of HGMSBA, an automatic soft interpolation machine is developed which can intuitively increase the robustness and accuracy of the final model. Here, HGMSBA-FIS is used to provide a nonlinear map between inputs, i.e. nanoparticles solid volume fraction (ϕ, Eckert number (Ec and a moving parameter which describes the movements of plates (S, and output, i.e. Nusselt number (Nu. Prior to proceeding with the modeling process, a comprehensive numerical comparative study is performed to investigate the potentials of the proposed model for nonlinear system identification. After demonstrating the efficacy of HGMSBA for training the FIS, the system is applied to the considered problem. Based on the obtained results, it can be inferred that the developed HGMSBA-FIS black-box identifier can be used as a very authentic tool with respect to accuracy and robustness. Besides, as the proposed black-box is not a physics-based identifier, it frees experts from the cumbersome mathematical formulations, and can be used for advanced real-time applications such as model-based control. The simulations

  10. IMPLANT-ASSOCIATED PATHOLOGY: AN ALGORITHM FOR IDENTIFYING PARTICLES IN HISTOPATHOLOGIC SYNOVIALIS/SLIM DIAGNOSTICS

    Directory of Open Access Journals (Sweden)

    V. Krenn

    2014-01-01

    Full Text Available In histopathologic SLIM diagnostic (synovial-like interface membrane, SLIM apart from diagnosing periprosthetic infection particle identification has an important role to play. The differences in particle pathogenesis and variability of materials in endoprosthetics explain the particle heterogeneity that hampers the diagnostic identification of particles. For this reason, a histopathological particle algorithm has been developed. With minimal methodical complexity this histopathological particle algorithm offers a guide to prosthesis material-particle identification. Light microscopic-morphological as well as enzyme-histochemical characteristics and polarization-optical proporties have set and particles are defined by size (microparticles, macroparticles and supra- macroparticles and definitely characterized in accordance with a dichotomous principle. Based on these criteria, identification and validation of the particles was carried out in 120 joint endoprosthesis pathological cases. A histopathological particle score (HPS is proposed that summarizes the most important information for the orthopedist, material scientist and histopathologist concerning particle identification in the SLIM.

  11. Use of genetic algorithm to identify thermophysical properties of deposited fouling in heat exchanger tubes

    International Nuclear Information System (INIS)

    Adili, Ali; Ben Salah, Mohieddine; Kerkeni, Chekib; Ben Nasrallah, Sassi

    2009-01-01

    At high temperature, the circulation of fluid in heat exchangers provides a tendency for fouling accumulation to take place on the internal surface of tubes. This paper shows an experimental process of thermophysical properties estimation of the fouling deposited on internal surface of a heat exchanger tube using genetic algorithms (GAs). The genetic algorithm is used to minimize an objective function containing calculated and measured temperatures. The experimental bench using a photothermal method with a finite width pulse heat excitation is used and the estimated parameters are obtained with high accuracy

  12. Implementation of the LandTrendr Algorithm on Google Earth Engine

    Directory of Open Access Journals (Sweden)

    Robert E Kennedy

    2018-05-01

    Full Text Available The LandTrendr (LT algorithm has been used widely for analysis of change in Landsat spectral time series data, but requires significant pre-processing, data management, and computational resources, and is only accessible to the community in a proprietary programming language (IDL. Here, we introduce LT for the Google Earth Engine (GEE platform. The GEE platform simplifies pre-processing steps, allowing focus on the translation of the core temporal segmentation algorithm. Temporal segmentation involved a series of repeated random access calls to each pixel’s time series, resulting in a set of breakpoints (“vertices” that bound straight-line segments. The translation of the algorithm into GEE included both transliteration and code analysis, resulting in improvement and logic error fixes. At six study areas representing diverse land cover types across the U.S., we conducted a direct comparison of the new LT-GEE code against the heritage code (LT-IDL. The algorithms agreed in most cases, and where disagreements occurred, they were largely attributable to logic error fixes in the code translation process. The practical impact of these changes is minimal, as shown by an example of forest disturbance mapping. We conclude that the LT-GEE algorithm represents a faithful translation of the LT code into a platform easily accessible by the broader user community.

  13. FRAMEWORK FOR COMPARING SEGMENTATION ALGORITHMS

    Directory of Open Access Journals (Sweden)

    G. Sithole

    2015-05-01

    Full Text Available The notion of a ‘Best’ segmentation does not exist. A segmentation algorithm is chosen based on the features it yields, the properties of the segments (point sets it generates, and the complexity of its algorithm. The segmentation is then assessed based on a variety of metrics such as homogeneity, heterogeneity, fragmentation, etc. Even after an algorithm is chosen its performance is still uncertain because the landscape/scenarios represented in a point cloud have a strong influence on the eventual segmentation. Thus selecting an appropriate segmentation algorithm is a process of trial and error. Automating the selection of segmentation algorithms and their parameters first requires methods to evaluate segmentations. Three common approaches for evaluating segmentation algorithms are ‘goodness methods’, ‘discrepancy methods’ and ‘benchmarks’. Benchmarks are considered the most comprehensive method of evaluation. This paper shortcomings in current benchmark methods are identified and a framework is proposed that permits both a visual and numerical evaluation of segmentations for different algorithms, algorithm parameters and evaluation metrics. The concept of the framework is demonstrated on a real point cloud. Current results are promising and suggest that it can be used to predict the performance of segmentation algorithms.

  14. Fast compact algorithms and software for spline smoothing

    CERN Document Server

    Weinert, Howard L

    2012-01-01

    Fast Compact Algorithms and Software for Spline Smoothing investigates algorithmic alternatives for computing cubic smoothing splines when the amount of smoothing is determined automatically by minimizing the generalized cross-validation score. These algorithms are based on Cholesky factorization, QR factorization, or the fast Fourier transform. All algorithms are implemented in MATLAB and are compared based on speed, memory use, and accuracy. An overall best algorithm is identified, which allows very large data sets to be processed quickly on a personal computer.

  15. Assessment of a Novel Approach to Identify Trichiasis Cases Using Community Treatment Assistants in Tanzania.

    Science.gov (United States)

    Greene, Gregory S; West, Sheila K; Mkocha, Harran; Munoz, Beatriz; Merbs, Shannath L

    2015-12-01

    Simple surgical intervention advocated by the World Health Organization can alleviate trachomatous trichiasis (TT) and prevent subsequent blindness. A large backlog of TT cases remain unidentified and untreated. To increase identification and referral of TT cases, a novel approach using standard screening questions, a card, and simple training for Community Treatment Assistants (CTAs) to use during Mass Drug Administration (MDA) was developed and evaluated in Kongwa District, a trachoma-endemic area of central Tanzania. A community randomized trial was conducted in 36 communities during MDA. CTAs in intervention villages received an additional half-day of training and a TT screening card in addition to the training received by CTAs in villages assigned to usual care. All MDA participants 15 years and older were screened for TT, and senior TT graders confirmed case status by evaluating all screened-positive cases. A random sample of those screened negative for TT and those who did not present at MDA were also evaluated by the master graders. Intervention CTAs identified 5.6 times as many cases (n = 50) as those assigned to usual care (n = 9, p card significantly increased the ability of CTAs to recognize and refer TT cases during MDA; however, further efforts are needed to improve case detection and reduce the number of false positive cases.

  16. Evidence-based algorithm for heparin dosing before cardiopulmonary bypass. Part 1: Development of the algorithm.

    Science.gov (United States)

    McKinney, Mark C; Riley, Jeffrey B

    2007-12-01

    The incidence of heparin resistance during adult cardiac surgery with cardiopulmonary bypass has been reported at 15%-20%. The consistent use of a clinical decision-making algorithm may increase the consistency of patient care and likely reduce the total required heparin dose and other problems associated with heparin dosing. After a directed survey of practicing perfusionists regarding treatment of heparin resistance and a literature search for high-level evidence regarding the diagnosis and treatment of heparin resistance, an evidence-based decision-making algorithm was constructed. The face validity of the algorithm decisive steps and logic was confirmed by a second survey of practicing perfusionists. The algorithm begins with review of the patient history to identify predictors for heparin resistance. The definition for heparin resistance contained in the algorithm is an activated clotting time 450 IU/kg heparin loading dose. Based on the literature, the treatment for heparin resistance used in the algorithm is anti-thrombin III supplement. The algorithm seems to be valid and is supported by high-level evidence and clinician opinion. The next step is a human randomized clinical trial to test the clinical procedure guideline algorithm vs. current standard clinical practice.

  17. Homophyly/kinship hypothesis: Natural communities, and predicting in networks

    Science.gov (United States)

    Li, Angsheng; Li, Jiankou; Pan, Yicheng

    2015-02-01

    It has been a longstanding challenge to understand natural communities in real world networks. We proposed a community finding algorithm based on fitness of networks, two algorithms for prediction, accurate prediction and confirmation of keywords for papers in the citation network Arxiv HEP-TH (high energy physics theory), and the measures of internal centrality, external de-centrality, internal and external slopes to characterize the structures of communities. We implemented our algorithms on 2 citation and 5 cooperation graphs. Our experiments explored and validated a homophyly/kinship principle of real world networks. The homophyly/kinship principle includes: (1) homophyly is the natural selection in real world networks, similar to Darwin's kinship selection in nature, (2) real world networks consist of natural communities generated by the natural selection of homophyly, (3) most individuals in a natural community share a short list of common attributes, (4) natural communities have an internal centrality (or internal heterogeneity) that a natural community has a few nodes dominating most of the individuals in the community, (5) natural communities have an external de-centrality (or external homogeneity) that external links of a natural community homogeneously distributed in different communities, and (6) natural communities of a given network have typical structures determined by the internal slopes, and have typical patterns of outgoing links determined by external slopes, etc. Our homophyly/kinship principle perfectly matches Darwin's observation that animals from ants to people form social groups in which most individuals work for the common good, and that kinship could encourage altruistic behavior. Our homophyly/kinship principle is the network version of Darwinian theory, and builds a bridge between Darwinian evolution and network science.

  18. Factor Analysis of Therapist-Identified Treatment Targets in Community-Based Children's Mental Health.

    Science.gov (United States)

    Love, Allison R; Okado, Izumi; Orimoto, Trina E; Mueller, Charles W

    2018-01-01

    The present study used exploratory and confirmatory factor analyses to identify underlying latent factors affecting variation in community therapists' endorsement of treatment targets. As part of a statewide practice management program, therapist completed monthly reports of treatment targets (up to 10 per month) for a sample of youth (n = 790) receiving intensive in-home therapy. Nearly 75 % of youth were diagnosed with multiple co-occurring disorders. Five factors emerged: Disinhibition, Societal Rules Evasion, Social Engagement Deficits, Emotional Distress, and Management of Biodevelopmental Outcomes. Using logistic regression, primary diagnosis predicted therapist selection of Disinhibition and Emotional Distress targets. Client age predicted endorsement of Societal Rules Evasion targets. Practice-to-research implications are discussed.

  19. Reconfiguration of Cortical Networks in MDD Uncovered by Multiscale Community Detection with fMRI.

    Science.gov (United States)

    He, Ye; Lim, Sol; Fortunato, Santo; Sporns, Olaf; Zhang, Lei; Qiu, Jiang; Xie, Peng; Zuo, Xi-Nian

    2018-04-01

    Major depressive disorder (MDD) is known to be associated with altered interactions between distributed brain regions. How these regional changes relate to the reorganization of cortical functional systems, and their modulation by antidepressant medication, is relatively unexplored. To identify changes in the community structure of cortical functional networks in MDD, we performed a multiscale community detection algorithm on resting-state functional connectivity networks of unmedicated MDD (uMDD) patients (n = 46), medicated MDD (mMDD) patients (n = 38), and healthy controls (n = 50), which yielded a spectrum of multiscale community partitions. we selected an optimal resolution level by identifying the most stable community partition for each group. uMDD and mMDD groups exhibited a similar reconfiguration of the community structure of the visual association and the default mode systems but showed different reconfiguration profiles in the frontoparietal control (FPC) subsystems. Furthermore, the central system (somatomotor/salience) and 3 frontoparietal subsystems showed strengthened connectivity with other communities in uMDD but, with the exception of 1 frontoparietal subsystem, returned to control levels in mMDD. These findings provide evidence for reconfiguration of specific cortical functional systems associated with MDD, as well as potential effects of medication in restoring disease-related network alterations, especially those of the FPC system.

  20. Enhancing Community Detection By Affinity-based Edge Weighting Scheme

    Energy Technology Data Exchange (ETDEWEB)

    Yoo, Andy [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Sanders, Geoffrey [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Henson, Van [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Vassilevski, Panayot [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2015-10-05

    Community detection refers to an important graph analytics problem of finding a set of densely-connected subgraphs in a graph and has gained a great deal of interest recently. The performance of current community detection algorithms is limited by an inherent constraint of unweighted graphs that offer very little information on their internal community structures. In this paper, we propose a new scheme to address this issue that weights the edges in a given graph based on recently proposed vertex affinity. The vertex affinity quantifies the proximity between two vertices in terms of their clustering strength, and therefore, it is ideal for graph analytics applications such as community detection. We also demonstrate that the affinity-based edge weighting scheme can improve the performance of community detection algorithms significantly.

  1. A new hierarchical method to find community structure in networks

    Science.gov (United States)

    Saoud, Bilal; Moussaoui, Abdelouahab

    2018-04-01

    Community structure is very important to understand a network which represents a context. Many community detection methods have been proposed like hierarchical methods. In our study, we propose a new hierarchical method for community detection in networks based on genetic algorithm. In this method we use genetic algorithm to split a network into two networks which maximize the modularity. Each new network represents a cluster (community). Then we repeat the splitting process until we get one node at each cluster. We use the modularity function to measure the strength of the community structure found by our method, which gives us an objective metric for choosing the number of communities into which a network should be divided. We demonstrate that our method are highly effective at discovering community structure in both computer-generated and real-world network data.

  2. Quantum algorithms for testing Boolean functions

    Directory of Open Access Journals (Sweden)

    Erika Andersson

    2010-06-01

    Full Text Available We discuss quantum algorithms, based on the Bernstein-Vazirani algorithm, for finding which variables a Boolean function depends on. There are 2^n possible linear Boolean functions of n variables; given a linear Boolean function, the Bernstein-Vazirani quantum algorithm can deterministically identify which one of these Boolean functions we are given using just one single function query. The same quantum algorithm can also be used to learn which input variables other types of Boolean functions depend on, with a success probability that depends on the form of the Boolean function that is tested, but does not depend on the total number of input variables. We also outline a procedure to futher amplify the success probability, based on another quantum algorithm, the Grover search.

  3. The application of particle swarm optimization to identify gamma spectrum with neural network

    International Nuclear Information System (INIS)

    Shi Dongsheng; Di Yuming; Zhou Chunlin

    2006-01-01

    Aiming at the shortcomings that BP algorithm is usually trapped to a local optimum and it has a low speed of convergence in the application of neural network to identify gamma spectrum, according to the advantage of the globe optimal searching of particle swarm optimization, this paper put forward a new algorithm for neural network training by combining BP algorithm and Particle Swarm Optimization-mixed PSO-BP algorithm. In the application to identify gamma spectrum, the new algorithm overcomes the shortcoming that BP algorithm is usually trapped to a local optimum and the neural network trained by it has a high ability of generalization with identification result of one hundred percent correct. Practical example shows that the mixed PSO-BP algorithm can effectively and reliably be used to identify gamma spectrum. (authors)

  4. An effective trust-based recommendation method using a novel graph clustering algorithm

    Science.gov (United States)

    Moradi, Parham; Ahmadian, Sajad; Akhlaghian, Fardin

    2015-10-01

    Recommender systems are programs that aim to provide personalized recommendations to users for specific items (e.g. music, books) in online sharing communities or on e-commerce sites. Collaborative filtering methods are important and widely accepted types of recommender systems that generate recommendations based on the ratings of like-minded users. On the other hand, these systems confront several inherent issues such as data sparsity and cold start problems, caused by fewer ratings against the unknowns that need to be predicted. Incorporating trust information into the collaborative filtering systems is an attractive approach to resolve these problems. In this paper, we present a model-based collaborative filtering method by applying a novel graph clustering algorithm and also considering trust statements. In the proposed method first of all, the problem space is represented as a graph and then a sparsest subgraph finding algorithm is applied on the graph to find the initial cluster centers. Then, the proposed graph clustering algorithm is performed to obtain the appropriate users/items clusters. Finally, the identified clusters are used as a set of neighbors to recommend unseen items to the current active user. Experimental results based on three real-world datasets demonstrate that the proposed method outperforms several state-of-the-art recommender system methods.

  5. Methods for identifying 30 chronic conditions: application to administrative data.

    Science.gov (United States)

    Tonelli, Marcello; Wiebe, Natasha; Fortin, Martin; Guthrie, Bruce; Hemmelgarn, Brenda R; James, Matthew T; Klarenbach, Scott W; Lewanczuk, Richard; Manns, Braden J; Ronksley, Paul; Sargious, Peter; Straus, Sharon; Quan, Hude

    2015-04-17

    Multimorbidity is common and associated with poor clinical outcomes and high health care costs. Administrative data are a promising tool for studying the epidemiology of multimorbidity. Our goal was to derive and apply a new scheme for using administrative data to identify the presence of chronic conditions and multimorbidity. We identified validated algorithms that use ICD-9 CM/ICD-10 data to ascertain the presence or absence of 40 morbidities. Algorithms with both positive predictive value and sensitivity ≥70% were graded as "high validity"; those with positive predictive value ≥70% and sensitivity <70% were graded as "moderate validity". To show proof of concept, we applied identified algorithms with high to moderate validity to inpatient and outpatient claims and utilization data from 574,409 people residing in Edmonton, Canada during the 2008/2009 fiscal year. Of the 40 morbidities, we identified 30 that could be identified with high to moderate validity. Approximately one quarter of participants had identified multimorbidity (2 or more conditions), one quarter had a single identified morbidity and the remaining participants were not identified as having any of the 30 morbidities. We identified a panel of 30 chronic conditions that can be identified from administrative data using validated algorithms, facilitating the study and surveillance of multimorbidity. We encourage other groups to use this scheme, to facilitate comparisons between settings and jurisdictions.

  6. Algorithms and Data Structures (lecture 1)

    CERN Multimedia

    CERN. Geneva

    2018-01-01

    Algorithms have existed, in one form or another, for as long as humanity has. During the second half of the 20th century, the field was revolutionised with the introduction of ever faster computers. In these lectures we discuss how algorithms are designed, how to evaluate their speed, and how to identify areas of improvement in existing algorithms. An algorithm consists of more than just a series of instructions; almost as important is the memory structure of the data on which it operates. A part of the lectures will be dedicated to a discussion of the various ways one can store data in memory, and their advantages and disadvantages.

  7. Algorithms and Data Structures (lecture 2)

    CERN Multimedia

    CERN. Geneva

    2018-01-01

    Algorithms have existed, in one form or another, for as long as humanity has. During the second half of the 20th century, the field was revolutionised with the introduction of ever faster computers. In these lectures we discuss how algorithms are designed, how to evaluate their speed, and how to identify areas of improvement in existing algorithms. An algorithm consists of more than just a series of instructions; almost as important is the memory structure of the data on which it operates. A part of the lectures will be dedicated to a discussion of the various ways one can store data in memory, and their advantages and disadvantages.

  8. Planning of community heating systems modernization and development

    International Nuclear Information System (INIS)

    Mroz, Tomasz M.

    2008-01-01

    New approach to community heating systems modernization and development planning process has been proposed. It is based on the general decision making aid algorithm. The proposed algorithm takes into account both demand and supply side of community heat market modernization and development. The first step of algorithm - analytical step, refers to data base creation, which is needed for the description of community heating system energy, ecology and economic characteristics. Analysis of those characteristics allows for the identification of heating system market modernization and development potential scenarios. The second algorithm step - decision step, allows for the identification of the most compromise scenarios of system modernization and development. To make the planning process more transparent and to increase the influence of decision makers on the planning process the ELECTRE III method was chosen as the tool of decision aid. The ELECTRE III method is based on the construction of outranking relation and definition of pseudo-criterion. The iteration mode of method application allows the decision maker and analyst for the investigation of the sensitivity of final solution to the changing preference model. One of the methods of statistics - the creditability of mean range method was used for the determination of initial definition of pseudo-criterion. Proposed algorithm and decision aid method were employed for the case study analysis referring to the choice of the heating system for new developing urban area. (author)

  9. Structure-Based Algorithms for Microvessel Classification

    KAUST Repository

    Smith, Amy F.

    2015-02-01

    © 2014 The Authors. Microcirculation published by John Wiley & Sons Ltd. Objective: Recent developments in high-resolution imaging techniques have enabled digital reconstruction of three-dimensional sections of microvascular networks down to the capillary scale. To better interpret these large data sets, our goal is to distinguish branching trees of arterioles and venules from capillaries. Methods: Two novel algorithms are presented for classifying vessels in microvascular anatomical data sets without requiring flow information. The algorithms are compared with a classification based on observed flow directions (considered the gold standard), and with an existing resistance-based method that relies only on structural data. Results: The first algorithm, developed for networks with one arteriolar and one venular tree, performs well in identifying arterioles and venules and is robust to parameter changes, but incorrectly labels a significant number of capillaries as arterioles or venules. The second algorithm, developed for networks with multiple inlets and outlets, correctly identifies more arterioles and venules, but is more sensitive to parameter changes. Conclusions: The algorithms presented here can be used to classify microvessels in large microvascular data sets lacking flow information. This provides a basis for analyzing the distinct geometrical properties and modelling the functional behavior of arterioles, capillaries, and venules.

  10. Using network properties to evaluate targeted immunization algorithms

    Directory of Open Access Journals (Sweden)

    Bita Shams

    2014-09-01

    Full Text Available Immunization of complex network with minimal or limited budget is a challenging issue for research community. In spite of much literature in network immunization, no comprehensive research has been conducted for evaluation and comparison of immunization algorithms. In this paper, we propose an evaluation framework for immunization algorithms regarding available amount of vaccination resources, goal of immunization program, and time complexity. The evaluation framework is designed based on network topological metrics which is extensible to all epidemic spreading model. Exploiting evaluation framework on well-known targeted immunization algorithms shows that in general, immunization based on PageRank centrality outperforms other targeting strategies in various types of networks, whereas, closeness and eigenvector centrality exhibit the worst case performance.

  11. Fast stochastic algorithm for simulating evolutionary population dynamics

    Science.gov (United States)

    Tsimring, Lev; Hasty, Jeff; Mather, William

    2012-02-01

    Evolution and co-evolution of ecological communities are stochastic processes often characterized by vastly different rates of reproduction and mutation and a coexistence of very large and very small sub-populations of co-evolving species. This creates serious difficulties for accurate statistical modeling of evolutionary dynamics. In this talk, we introduce a new exact algorithm for fast fully stochastic simulations of birth/death/mutation processes. It produces a significant speedup compared to the direct stochastic simulation algorithm in a typical case when the total population size is large and the mutation rates are much smaller than birth/death rates. We illustrate the performance of the algorithm on several representative examples: evolution on a smooth fitness landscape, NK model, and stochastic predator-prey system.

  12. Alternative confidence measure for local matching stereo algorithms

    CSIR Research Space (South Africa)

    Ndhlovu, T

    2009-11-01

    Full Text Available The authors present a confidence measure applied to individual disparity estimates in local matching stereo correspondence algorithms. It aims at identifying textureless areas, where most local matching algorithms fail. The confidence measure works...

  13. Soil Fungal Community Associated with Peat in Sarawak Identified Using 18S rDNA Marker

    International Nuclear Information System (INIS)

    Siti Ramlah Ahmad Ali; Sakinah Safari; Mohd Shawal Thakib; Shamsilawani Ahamed Bakeri; Nur Aziemah Ab Ghani

    2016-01-01

    Fungi are principal decomposing microorganisms in acidic environment of peat lands. A useful tool for molecular screening of soil fungal communities using the 18S ribosomal DNA primer has been proven capable of identifying a broad range of fungi species within Ascomycota, Basidiomycota, Zygomycota and Chytridiomycota. Currently, very little information is available on fungal communities in deep peat of Sarawak, Malaysia. In this study, we have isolated the fungi from soil samples taken in deep peat forests and oil palm cultivated areas. The fungal identity was undertaken using 18S ribosomal DNA primer which is EF4-F/ fung5-R. The microscopic structures were conducted to confirm the identity of the isolates. Based on this study, the fungal division most commonly found in deep peat is the Ascomycota. Aspergillus fumigatus was the most common species and more dominant in oil palm cultivated areas and logged-over forest than in primary forest. In the primary forest, the dominant species was the A. flavus, while Hypocrea atroviridis was commonly associated with oil palm cultivated areas and logged-over forest. Other species of fungi isolated in peat primary forests were Penicillium chrysogenum, Trichoderma sp., Phanerochaete sp., Mortierella chlamydospora, A. niger, A. alliaceus, etc. The in-depth difference in the fungal communities for the different sites will be further investigated using the next generation sequencing technology. (author)

  14. Fast image matching algorithm based on projection characteristics

    Science.gov (United States)

    Zhou, Lijuan; Yue, Xiaobo; Zhou, Lijun

    2011-06-01

    Based on analyzing the traditional template matching algorithm, this paper identified the key factors restricting the speed of matching and put forward a brand new fast matching algorithm based on projection. Projecting the grayscale image, this algorithm converts the two-dimensional information of the image into one-dimensional one, and then matches and identifies through one-dimensional correlation, meanwhile, because of normalization has been done, when the image brightness or signal amplitude increasing in proportion, it could also perform correct matching. Experimental results show that the projection characteristics based image registration method proposed in this article could greatly improve the matching speed, which ensuring the matching accuracy as well.

  15. Cardiovascular Disease Population Risk Tool (CVDPoRT): predictive algorithm for assessing CVD risk in the community setting. A study protocol.

    Science.gov (United States)

    Taljaard, Monica; Tuna, Meltem; Bennett, Carol; Perez, Richard; Rosella, Laura; Tu, Jack V; Sanmartin, Claudia; Hennessy, Deirdre; Tanuseputro, Peter; Lebenbaum, Michael; Manuel, Douglas G

    2014-10-23

    Recent publications have called for substantial improvements in the design, conduct, analysis and reporting of prediction models. Publication of study protocols, with prespecification of key aspects of the analysis plan, can help to improve transparency, increase quality and protect against increased type I error. Valid population-based risk algorithms are essential for population health planning and policy decision-making. The purpose of this study is to develop, evaluate and apply cardiovascular disease (CVD) risk algorithms for the population setting. The Ontario sample of the Canadian Community Health Survey (2001, 2003, 2005; 77,251 respondents) will be used to assess risk factors focusing on health behaviours (physical activity, diet, smoking and alcohol use). Incident CVD outcomes will be assessed through linkage to administrative healthcare databases (619,886 person-years of follow-up until 31 December 2011). Sociodemographic factors (age, sex, immigrant status, education) and mediating factors such as presence of diabetes and hypertension will be included as predictors. Algorithms will be developed using competing risks survival analysis. The analysis plan adheres to published recommendations for the development of valid prediction models to limit the risk of overfitting and improve the quality of predictions. Key considerations are fully prespecifying the predictor variables; appropriate handling of missing data; use of flexible functions for continuous predictors; and avoiding data-driven variable selection procedures. The 2007 and 2009 surveys (approximately 50,000 respondents) will be used for validation. Calibration will be assessed overall and in predefined subgroups of importance to clinicians and policymakers. This study has been approved by the Ottawa Health Science Network Research Ethics Board. The findings will be disseminated through professional and scientific conferences, and in peer-reviewed journals. The algorithm will be accessible

  16. Epidemic spreading on complex networks with overlapping and non-overlapping community structure

    Science.gov (United States)

    Shang, Jiaxing; Liu, Lianchen; Li, Xin; Xie, Feng; Wu, Cheng

    2015-02-01

    Many real-world networks exhibit community structure where vertices belong to one or more communities. Recent studies show that community structure plays an import role in epidemic spreading. In this paper, we investigate how the extent of overlap among communities affects epidemics. In order to experiment on the characteristic of overlapping communities, we propose a rewiring algorithm that can change the community structure from overlapping to non-overlapping while maintaining the degree distribution of the network. We simulate the Susceptible-Infected-Susceptible (SIS) epidemic process on synthetic scale-free networks and real-world networks by applying our rewiring algorithm. Experiments show that epidemics spread faster on networks with higher level of overlapping communities. Furthermore, overlapping communities' effect interacts with the average degree's effect. Our work further illustrates the important role of overlapping communities in the process of epidemic spreading.

  17. Multi-User Identification-Based Eye-Tracking Algorithm Using Position Estimation

    Directory of Open Access Journals (Sweden)

    Suk-Ju Kang

    2016-12-01

    Full Text Available This paper proposes a new multi-user eye-tracking algorithm using position estimation. Conventional eye-tracking algorithms are typically suitable only for a single user, and thereby cannot be used for a multi-user system. Even though they can be used to track the eyes of multiple users, their detection accuracy is low and they cannot identify multiple users individually. The proposed algorithm solves these problems and enhances the detection accuracy. Specifically, the proposed algorithm adopts a classifier to detect faces for the red, green, and blue (RGB and depth images. Then, it calculates features based on the histogram of the oriented gradient for the detected facial region to identify multiple users, and selects the template that best matches the users from a pre-determined face database. Finally, the proposed algorithm extracts the final eye positions based on anatomical proportions. Simulation results show that the proposed algorithm improved the average F1 score by up to 0.490, compared with benchmark algorithms.

  18. GASS-WEB: a web server for identifying enzyme active sites based on genetic algorithms.

    Science.gov (United States)

    Moraes, João P A; Pappa, Gisele L; Pires, Douglas E V; Izidoro, Sandro C

    2017-07-03

    Enzyme active sites are important and conserved functional regions of proteins whose identification can be an invaluable step toward protein function prediction. Most of the existing methods for this task are based on active site similarity and present limitations including performing only exact matches on template residues, template size restraints, despite not being capable of finding inter-domain active sites. To fill this gap, we proposed GASS-WEB, a user-friendly web server that uses GASS (Genetic Active Site Search), a method based on an evolutionary algorithm to search for similar active sites in proteins. GASS-WEB can be used under two different scenarios: (i) given a protein of interest, to match a set of specific active site templates; or (ii) given an active site template, looking for it in a database of protein structures. The method has shown to be very effective on a range of experiments and was able to correctly identify >90% of the catalogued active sites from the Catalytic Site Atlas. It also managed to achieve a Matthew correlation coefficient of 0.63 using the Critical Assessment of protein Structure Prediction (CASP 10) dataset. In our analysis, GASS was ranking fourth among 18 methods. GASS-WEB is freely available at http://gass.unifei.edu.br/. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  19. Comparison of Clustering Algorithms for the Identification of Topics on Twitter

    Directory of Open Access Journals (Sweden)

    Marjori N. M. Klinczak

    2016-05-01

    Full Text Available Topic Identification in Social Networks has become an important task when dealing with event detection, particularly when global communities are affected. In order to attack this problem, text processing techniques and machine learning algorithms have been extensively used. In this paper we compare four clustering algorithms – k-means, k-medoids, DBSCAN and NMF (Non-negative Matrix Factorization – in order to detect topics related to textual messages obtained from Twitter. The algorithms were applied to a database initially composed by tweets having hashtags related to the recent Nepal earthquake as initial context. Obtained results suggest that the NMF clustering algorithm presents superior results, providing simpler clusters that are also easier to interpret.

  20. Identification of chaotic systems by neural network with hybrid learning algorithm

    International Nuclear Information System (INIS)

    Pan, S.-T.; Lai, C.-C.

    2008-01-01

    Based on the genetic algorithm (GA) and steepest descent method (SDM), this paper proposes a hybrid algorithm for the learning of neural networks to identify chaotic systems. The systems in question are the logistic map and the Duffing equation. Different identification schemes are used to identify both the logistic map and the Duffing equation, respectively. Simulation results show that our hybrid algorithm is more efficient than that of other methods

  1. [Development and validation of an algorithm to identify cancer recurrences from hospital data bases].

    Science.gov (United States)

    Manzanares-Laya, S; Burón, A; Murta-Nascimento, C; Servitja, S; Castells, X; Macià, F

    2014-01-01

    Hospital cancer registries and hospital databases are valuable and efficient sources of information for research into cancer recurrences. The aim of this study was to develop and validate algorithms for the detection of breast cancer recurrence. A retrospective observational study was conducted on breast cancer cases from the cancer registry of a third level university hospital diagnosed between 2003 and 2009. Different probable cancer recurrence algorithms were obtained by linking the hospital databases and the construction of several operational definitions, with their corresponding sensitivity, specificity, positive predictive value and negative predictive value. A total of 1,523 patients were diagnosed of breast cancer between 2003 and 2009. A request for bone gammagraphy after 6 months from the first oncological treatment showed the highest sensitivity (53.8%) and negative predictive value (93.8%), and a pathology test after 6 months after the diagnosis showed the highest specificity (93.8%) and negative predictive value (92.6%). The combination of different definitions increased the specificity and the positive predictive value, but decreased the sensitivity. Several diagnostic algorithms were obtained, and the different definitions could be useful depending on the interest and resources of the researcher. A higher positive predictive value could be interesting for a quick estimation of the number of cases, and a higher negative predictive value for a more exact estimation if more resources are available. It is a versatile and adaptable tool for other types of tumors, as well as for the needs of the researcher. Copyright © 2014 SECA. Published by Elsevier Espana. All rights reserved.

  2. Online Community Transition Detection

    DEFF Research Database (Denmark)

    Tan, Biying; Zhu, Feida; Qu, Qiang

    2014-01-01

    communities over time. How to automatically detect the online community transitions of individual users is a research problem of immense practical value yet with great technical challenges. In this paper, we propose an algorithm based on the Minimum Description Length (MDL) principle to trace the evolution......Mining user behavior patterns in social networks is of great importance in user behavior analysis, targeted marketing, churn prediction and other applications. However, less effort has been made to study the evolution of user behavior in social communities. In particular, users join and leave...... of community transition of individual users, adaptive to the noisy behavior. Experiments on real data sets demonstrate the efficiency and effectiveness of our proposed method....

  3. Identifying the Entrepreneurship Characteristics of the Oil Palm Community Plantation Farmers in the Riau Area

    Directory of Open Access Journals (Sweden)

    Brilliant Asmit

    2015-12-01

    Full Text Available Oil palm is an essential and strategic commodity in the Riau area because of its considerable role in supporting the peoples’ economy, especially for plantation farmers. Oil palm plantation activities have brought economic impacts to society there, both for the people who are directly involved with the plantations and for their surrounding communities. This regional advantage is a facility for farmers to be able to develop their farms as plantations. The aims of this research are to identify the entrepreneurship characteristics of the oil palm farmers, and also to identify the entrepreneurship characteristics that differentiate the farmers, as seen from their business’ achievements. The research used a grounded theory approach to identify the characteristics of oil palm farmers systematically. The sampling method used for the research was theoretical sampling, which is data gathering driven by the concepts derived from the theory of previous entrepreneurship characteristics studies. The research object is the oil palm farmers in Riau, Indonesia. The results of the analysis identified the entrepreneurship characteristics of the oil palm farmers, they are growth oriented, risk-taking, innovative, with a sense of personal control, self confident, and cooperative. But, among the characteristics, only the characteristic of their cooperation did not differentiate the oil palm farmers in the achievement of their business activities.

  4. Sentiment analysis enhancement with target variable in Kumar’s Algorithm

    Science.gov (United States)

    Arman, A. A.; Kawi, A. B.; Hurriyati, R.

    2016-04-01

    Sentiment analysis (also known as opinion mining) refers to the use of text analysis and computational linguistics to identify and extract subjective information in source materials. Sentiment analysis is widely applied to reviews discussion that is being talked in social media for many purposes, ranging from marketing, customer service, or public opinion of public policy. One of the popular algorithm for Sentiment Analysis implementation is Kumar algorithm that developed by Kumar and Sebastian. Kumar algorithm can identify the sentiment score of the statement, sentence or tweet, but cannot determine the relationship of the object or target related to the sentiment being analysed. This research proposed solution for that challenge by adding additional component that represent object or target to the existing algorithm (Kumar algorithm). The result of this research is a modified algorithm that can give sentiment score based on a given object or target.

  5. Genetic algorithms with memory- and elitism-based immigrants in dynamic environments

    OpenAIRE

    Yang, S

    2008-01-01

    Copyright @ 2008 by the Massachusetts Institute of Technology In recent years the genetic algorithm community has shown a growing interest in studying dynamic optimization problems. Several approaches have been devised. The random immigrants and memory schemes are two major ones. The random immigrants scheme addresses dynamic environments by maintaining the population diversity while the memory scheme aims to adapt genetic algorithms quickly to new environments by reusing historical inform...

  6. Referral outcomes of individuals identified at high risk of cardiovascular disease by community health workers in Bangladesh, Guatemala, Mexico, and South Africa.

    Science.gov (United States)

    Levitt, Naomi S; Puoane, Thandi; Denman, Catalina A; Abrahams-Gessel, Shafika; Surka, Sam; Mendoza, Carlos; Khanam, Masuma; Alam, Sartaj; Gaziano, Thomas A

    2015-01-01

    We have found that community health workers (CHWs) with appropriate training are able to accurately identify people at high cardiovascular disease (CVD) risk in the community who would benefit from the introduction of preventative management, in Bangladesh, Guatemala, Mexico, and South Africa. This paper examines the attendance pattern for those individuals who were so identified and referred to a health care facility for further assessment and management. Patient records from the health centres in each site were reviewed for data on diagnoses made and treatment commenced. Reasons for non-attendance were sought from participants who had not attended after being referred. Qualitative data were collected from study coordinators regarding their experiences in obtaining the records and conducting the record reviews. The perspectives of CHWs and community members, who were screened, were also obtained. Thirty-seven percent (96/263) of those referred attended follow-up: 36 of 52 (69%) were urgent and 60 of 211 (28.4%) were non-urgent referrals. A diagnosis of hypertension (HTN) was made in 69% of urgent referrals and 37% of non-urgent referrals with treatment instituted in all cases. Reasons for non-attendance included limited self-perception of risk, associated costs, health system obstacles, and lack of trust in CHWs to conduct CVD risk assessments and to refer community members into the health system. The existing barriers to referral in the health care systems negatively impact the gains to be had through screening by training CHWs in the use of a simple risk assessment tool. The new diagnoses of HTN and commencement on treatment in those that attended referrals underscores the value of having persons at the highest risk identified in the community setting and referred to a clinic for further evaluation and treatment.

  7. Referral outcomes of individuals identified at high risk of cardiovascular disease by community health workers in Bangladesh, Guatemala, Mexico, and South Africa

    Science.gov (United States)

    Levitt, Naomi S.; Puoane, Thandi; Denman, Catalina A.; Abrahams-Gessel, Shafika; Surka, Sam; Mendoza, Carlos; Khanam, Masuma; Alam, Sartaj; Gaziano, Thomas A.

    2015-01-01

    Background We have found that community health workers (CHWs) with appropriate training are able to accurately identify people at high cardiovascular disease (CVD) risk in the community who would benefit from the introduction of preventative management, in Bangladesh, Guatemala, Mexico, and South Africa. This paper examines the attendance pattern for those individuals who were so identified and referred to a health care facility for further assessment and management. Design Patient records from the health centres in each site were reviewed for data on diagnoses made and treatment commenced. Reasons for non-attendance were sought from participants who had not attended after being referred. Qualitative data were collected from study coordinators regarding their experiences in obtaining the records and conducting the record reviews. The perspectives of CHWs and community members, who were screened, were also obtained. Results Thirty-seven percent (96/263) of those referred attended follow-up: 36 of 52 (69%) were urgent and 60 of 211 (28.4%) were non-urgent referrals. A diagnosis of hypertension (HTN) was made in 69% of urgent referrals and 37% of non-urgent referrals with treatment instituted in all cases. Reasons for non-attendance included limited self-perception of risk, associated costs, health system obstacles, and lack of trust in CHWs to conduct CVD risk assessments and to refer community members into the health system. Conclusions The existing barriers to referral in the health care systems negatively impact the gains to be had through screening by training CHWs in the use of a simple risk assessment tool. The new diagnoses of HTN and commencement on treatment in those that attended referrals underscores the value of having persons at the highest risk identified in the community setting and referred to a clinic for further evaluation and treatment. PMID:25854780

  8. Applying Kitaev's algorithm in an ion trap quantum computer

    International Nuclear Information System (INIS)

    Travaglione, B.; Milburn, G.J.

    2000-01-01

    Full text: Kitaev's algorithm is a method of estimating eigenvalues associated with an operator. Shor's factoring algorithm, which enables a quantum computer to crack RSA encryption codes, is a specific example of Kitaev's algorithm. It has been proposed that the algorithm can also be used to generate eigenstates. We extend this proposal for small quantum systems, identifying the conditions under which the algorithm can successfully generate eigenstates. We then propose an implementation scheme based on an ion trap quantum computer. This scheme allows us to illustrate a simple example, in which the algorithm effectively generates eigenstates

  9. Using the community pharmacy to identify patients at risk of poor asthma control and factors which contribute to this poor control.

    Science.gov (United States)

    Armour, Carol L; Lemay, Kate; Saini, Bandana; Reddel, Helen K; Bosnic-Anticevich, Sinthia Z; Smith, Lorraine D; Burton, Deborah; Song, Yun Ju Christine; Alles, Marie Chehani; Stewart, Kay; Emmerton, Lynne; Krass, Ines

    2011-11-01

    Although asthma can be well controlled by appropriate medication delivered in an appropriate way at an appropriate time, there is evidence that management is often suboptimal. This results in poor asthma control, poor quality of life, and significant morbidity. The objective of this study was to describe a population recruited in community pharmacy identified by trained community pharmacists as being at risk for poor asthma outcomes and to identify factors associated with poor asthma control. It used a cross-sectional design in 96 pharmacies in metropolitan and regional New South Wales, Victoria, Queensland, and Australian Capital Territory in Australia. Community pharmacists with specialized asthma training enrolled 570 patients aged ≥18 years with doctor-diagnosed asthma who were considered at risk of poor asthma outcomes and then conducted a comprehensive asthma assessment. In this assessment, asthma control was classified using a symptom and activity tool based on self-reported frequency of symptoms during the previous month and categorized as poor, fair, or good. Asthma history was discussed, and lung function and inhaler technique were also assessed by the pharmacist. Medication use/adherence was recorded from both pharmacy records and the Brief Medication Questionnaire (BMQ). The symptom and activity tool identified that 437 (77%) recruited patients had poor asthma control. Of the 570 patients, 117 (21%) smoked, 108 (19%) had an action plan, 372 (69%) used combination of inhaled corticosteroid (ICS)/long-acting β(2)-agonist (LABA) medications, and only 17-28% (depending on device) used their inhaler device correctly. In terms of adherence, 90% had their ICS or ICS/LABA dispensed <6 times in the previous 6 months, which is inconsistent with regular use; this low adherence was confirmed from the BMQ scores. A logistic regression model showed that patients who smoked had incorrect inhaler technique or low adherence (assessed by either dispensing history or

  10. An evolving network model with community structure

    International Nuclear Information System (INIS)

    Li Chunguang; Maini, Philip K

    2005-01-01

    Many social and biological networks consist of communities-groups of nodes within which connections are dense, but between which connections are sparser. Recently, there has been considerable interest in designing algorithms for detecting community structures in real-world complex networks. In this paper, we propose an evolving network model which exhibits community structure. The network model is based on the inner-community preferential attachment and inter-community preferential attachment mechanisms. The degree distributions of this network model are analysed based on a mean-field method. Theoretical results and numerical simulations indicate that this network model has community structure and scale-free properties

  11. Two-pass imputation algorithm for missing value estimation in gene expression time series.

    Science.gov (United States)

    Tsiporkova, Elena; Boeva, Veselka

    2007-10-01

    Gene expression microarray experiments frequently generate datasets with multiple values missing. However, most of the analysis, mining, and classification methods for gene expression data require a complete matrix of gene array values. Therefore, the accurate estimation of missing values in such datasets has been recognized as an important issue, and several imputation algorithms have already been proposed to the biological community. Most of these approaches, however, are not particularly suitable for time series expression profiles. In view of this, we propose a novel imputation algorithm, which is specially suited for the estimation of missing values in gene expression time series data. The algorithm utilizes Dynamic Time Warping (DTW) distance in order to measure the similarity between time expression profiles, and subsequently selects for each gene expression profile with missing values a dedicated set of candidate profiles for estimation. Three different DTW-based imputation (DTWimpute) algorithms have been considered: position-wise, neighborhood-wise, and two-pass imputation. These have initially been prototyped in Perl, and their accuracy has been evaluated on yeast expression time series data using several different parameter settings. The experiments have shown that the two-pass algorithm consistently outperforms, in particular for datasets with a higher level of missing entries, the neighborhood-wise and the position-wise algorithms. The performance of the two-pass DTWimpute algorithm has further been benchmarked against the weighted K-Nearest Neighbors algorithm, which is widely used in the biological community; the former algorithm has appeared superior to the latter one. Motivated by these findings, indicating clearly the added value of the DTW techniques for missing value estimation in time series data, we have built an optimized C++ implementation of the two-pass DTWimpute algorithm. The software also provides for a choice between three different

  12. Bio-inspired algorithms applied to molecular docking simulations.

    Science.gov (United States)

    Heberlé, G; de Azevedo, W F

    2011-01-01

    Nature as a source of inspiration has been shown to have a great beneficial impact on the development of new computational methodologies. In this scenario, analyses of the interactions between a protein target and a ligand can be simulated by biologically inspired algorithms (BIAs). These algorithms mimic biological systems to create new paradigms for computation, such as neural networks, evolutionary computing, and swarm intelligence. This review provides a description of the main concepts behind BIAs applied to molecular docking simulations. Special attention is devoted to evolutionary algorithms, guided-directed evolutionary algorithms, and Lamarckian genetic algorithms. Recent applications of these methodologies to protein targets identified in the Mycobacterium tuberculosis genome are described.

  13. Into the Bowels of Depression: Unravelling Medical Symptoms Associated with Depression by Applying Machine-Learning Techniques to a Community Based Population Sample

    Science.gov (United States)

    Dipnall, Joanna F.

    2016-01-01

    Background Depression is commonly comorbid with many other somatic diseases and symptoms. Identification of individuals in clusters with comorbid symptoms may reveal new pathophysiological mechanisms and treatment targets. The aim of this research was to combine machine-learning (ML) algorithms with traditional regression techniques by utilising self-reported medical symptoms to identify and describe clusters of individuals with increased rates of depression from a large cross-sectional community based population epidemiological study. Methods A multi-staged methodology utilising ML and traditional statistical techniques was performed using the community based population National Health and Nutrition Examination Study (2009–2010) (N = 3,922). A Self-organised Mapping (SOM) ML algorithm, combined with hierarchical clustering, was performed to create participant clusters based on 68 medical symptoms. Binary logistic regression, controlling for sociodemographic confounders, was used to then identify the key clusters of participants with higher levels of depression (PHQ-9≥10, n = 377). Finally, a Multiple Additive Regression Tree boosted ML algorithm was run to identify the important medical symptoms for each key cluster within 17 broad categories: heart, liver, thyroid, respiratory, diabetes, arthritis, fractures and osteoporosis, skeletal pain, blood pressure, blood transfusion, cholesterol, vision, hearing, psoriasis, weight, bowels and urinary. Results Five clusters of participants, based on medical symptoms, were identified to have significantly increased rates of depression compared to the cluster with the lowest rate: odds ratios ranged from 2.24 (95% CI 1.56, 3.24) to 6.33 (95% CI 1.67, 24.02). The ML boosted regression algorithm identified three key medical condition categories as being significantly more common in these clusters: bowel, pain and urinary symptoms. Bowel-related symptoms was found to dominate the relative importance of symptoms within the

  14. Into the Bowels of Depression: Unravelling Medical Symptoms Associated with Depression by Applying Machine-Learning Techniques to a Community Based Population Sample.

    Science.gov (United States)

    Dipnall, Joanna F; Pasco, Julie A; Berk, Michael; Williams, Lana J; Dodd, Seetal; Jacka, Felice N; Meyer, Denny

    2016-01-01

    Depression is commonly comorbid with many other somatic diseases and symptoms. Identification of individuals in clusters with comorbid symptoms may reveal new pathophysiological mechanisms and treatment targets. The aim of this research was to combine machine-learning (ML) algorithms with traditional regression techniques by utilising self-reported medical symptoms to identify and describe clusters of individuals with increased rates of depression from a large cross-sectional community based population epidemiological study. A multi-staged methodology utilising ML and traditional statistical techniques was performed using the community based population National Health and Nutrition Examination Study (2009-2010) (N = 3,922). A Self-organised Mapping (SOM) ML algorithm, combined with hierarchical clustering, was performed to create participant clusters based on 68 medical symptoms. Binary logistic regression, controlling for sociodemographic confounders, was used to then identify the key clusters of participants with higher levels of depression (PHQ-9≥10, n = 377). Finally, a Multiple Additive Regression Tree boosted ML algorithm was run to identify the important medical symptoms for each key cluster within 17 broad categories: heart, liver, thyroid, respiratory, diabetes, arthritis, fractures and osteoporosis, skeletal pain, blood pressure, blood transfusion, cholesterol, vision, hearing, psoriasis, weight, bowels and urinary. Five clusters of participants, based on medical symptoms, were identified to have significantly increased rates of depression compared to the cluster with the lowest rate: odds ratios ranged from 2.24 (95% CI 1.56, 3.24) to 6.33 (95% CI 1.67, 24.02). The ML boosted regression algorithm identified three key medical condition categories as being significantly more common in these clusters: bowel, pain and urinary symptoms. Bowel-related symptoms was found to dominate the relative importance of symptoms within the five key clusters. This

  15. A novel baiting microcosm approach used to identify the bacterial community associated with Penicillium bilaii hyphae in soil.

    Directory of Open Access Journals (Sweden)

    Behnoushsadat Ghodsalavi

    Full Text Available It is important to identify and recover bacteria associating with fungi under natural soil conditions to enable eco-physiological studies, and to facilitate the use of bacterial-fungal consortia in environmental biotechnology. We have developed a novel type of baiting microcosm, where fungal hyphae interact with bacteria under close-to-natural soil conditions; an advantage compared to model systems that determine fungal influences on bacterial communities in laboratory media. In the current approach, the hyphae are placed on a solid support, which enables the recovery of hyphae with associated bacteria in contrast to model systems that compare bulk soil and mycosphere soil. We used the baiting microcosm approach to determine, for the first time, the composition of the bacterial community associating in the soil with hyphae of the phosphate-solubilizer, Penicillium bilaii. By applying a cultivation-independent 16S rRNA gene-targeted amplicon sequencing approach, we found a hypha-associated bacterial community with low diversity compared to the bulk soil community and exhibiting massive dominance of Burkholderia OTUs. Burkholderia is known be abundant in soil environments affected by fungi, but the discovery of this massive dominance among bacteria firmly associating with hyphae in soil is novel and made possible by the current bait approach.

  16. Identifying participation needs of people with acquired brain injury in the development of a collective community smart home.

    Science.gov (United States)

    Levasseur, Mélanie; Pigot, Hélène; Couture, Mélanie; Bier, Nathalie; Swaine, Bonnie; Therriault, Pierre-Yves; Giroux, Sylvain

    2016-11-01

    This study explored the personalized and collective participation needs of people with acquired brain injury (ABI) living in a future shared community smart home. An action research study was conducted with 16 persons, seven with ABI, four caregivers and five rehabilitation or smart home healthcare providers. Twelve interviews and two focus groups were conducted, audiotaped, transcribed and analyzed for content. Seventy personalized and 18 collective participation needs were reported related to daily and social activities. Personalized needs concerned interpersonal relationships, general organization of activities, leisure, housing, fitness and nutrition. Collective needs related mainly to housing, general organization of activities and nutrition. Personalized and collective participation needs of people with ABI planning to live in a community smart home are diverse and concern daily as well as social activities. Implications for Rehabilitation To meet participation needs of people with ABI, the design of smart homes must consider all categories of daily and social activities. Considering personalized and collective needs allowed identifying exclusive examples of each. As some persons with ABI had difficulty identifying their needs as well as accepting their limitations and the assistance required, rehabilitation professionals must be involved in needs identification.

  17. Diametrical clustering for identifying anti-correlated gene clusters.

    Science.gov (United States)

    Dhillon, Inderjit S; Marcotte, Edward M; Roshan, Usman

    2003-09-01

    Clustering genes based upon their expression patterns allows us to predict gene function. Most existing clustering algorithms cluster genes together when their expression patterns show high positive correlation. However, it has been observed that genes whose expression patterns are strongly anti-correlated can also be functionally similar. Biologically, this is not unintuitive-genes responding to the same stimuli, regardless of the nature of the response, are more likely to operate in the same pathways. We present a new diametrical clustering algorithm that explicitly identifies anti-correlated clusters of genes. Our algorithm proceeds by iteratively (i). re-partitioning the genes and (ii). computing the dominant singular vector of each gene cluster; each singular vector serving as the prototype of a 'diametric' cluster. We empirically show the effectiveness of the algorithm in identifying diametrical or anti-correlated clusters. Testing the algorithm on yeast cell cycle data, fibroblast gene expression data, and DNA microarray data from yeast mutants reveals that opposed cellular pathways can be discovered with this method. We present systems whose mRNA expression patterns, and likely their functions, oppose the yeast ribosome and proteosome, along with evidence for the inverse transcriptional regulation of a number of cellular systems.

  18. 3rd International Conference on Harmony Search Algorithm

    CERN Document Server

    2017-01-01

    This book presents state-of-the-art technical contributions based around one of the most successful evolutionary optimization algorithms published to date: Harmony Search. Contributions span from novel technical derivations of this algorithm to applications in the broad fields of civil engineering, energy, transportation & mobility and health, among many others and focus not only on its cross-domain applicability, but also on its core evolutionary operators, including elements inspired from other meta-heuristics. The global scientific community is witnessing an upsurge in groundbreaking, new advances in all areas of computational intelligence, with a particular flurry of research focusing on evolutionary computation and bio-inspired optimization. Observed processes in nature and sociology have provided the basis for innovative algorithmic developments aimed at leveraging the inherent capability to adapt characterized by various animals, including ants, fireflies, wolves and humans. However, it is the beha...

  19. Ad Hoc Access Gateway Selection Algorithm

    Science.gov (United States)

    Jie, Liu

    With the continuous development of mobile communication technology, Ad Hoc access network has become a hot research, Ad Hoc access network nodes can be used to expand capacity of multi-hop communication range of mobile communication system, even business adjacent to the community, improve edge data rates. For mobile nodes in Ad Hoc network to internet, internet communications in the peer nodes must be achieved through the gateway. Therefore, the key Ad Hoc Access Networks will focus on the discovery gateway, as well as gateway selection in the case of multi-gateway and handover problems between different gateways. This paper considers the mobile node and the gateway, based on the average number of hops from an average access time and the stability of routes, improved gateway selection algorithm were proposed. An improved gateway selection algorithm, which mainly considers the algorithm can improve the access time of Ad Hoc nodes and the continuity of communication between the gateways, were proposed. This can improve the quality of communication across the network.

  20. Is Telephone Screening Feasible? Accuracy and Cost-Effectiveness of Identifying People Medically Eligible for Home- And Community-Based Services.

    Science.gov (United States)

    Fries, Brant E.; James, Mary; Hammer, Susan S.; Shugarman, Lisa R.; Morris, John N.

    2004-01-01

    Purpose: To determine the accuracy of a telephone-screening system to identify persons eligible for home- and community-based long-term care. Design and Methods: Data from Michigan telephone screens were compared to data from in-person assessments using the Minimum Data Set for Home Care (MDS-HC). Weighted kappa statistics measured the level of…

  1. SigTree: A Microbial Community Analysis Tool to Identify and Visualize Significantly Responsive Branches in a Phylogenetic Tree

    OpenAIRE

    Stevens, John R.; Jones, Todd R.; Lefevre, Michael; Ganesan, Balasubramanian; Weimer, Bart C.

    2017-01-01

    Microbial community analysis experiments to assess the effect of a treatment intervention (or environmental change) on the relative abundance levels of multiple related microbial species (or operational taxonomic units) simultaneously using high throughput genomics are becoming increasingly common. Within the framework of the evolutionary phylogeny of all species considered in the experiment, this translates to a statistical need to identify the phylogenetic branches that exhibit a significan...

  2. A simple algorithm for the identification of clinical COPD phenotypes

    DEFF Research Database (Denmark)

    Burgel, Pierre-Régis; Paillasseur, Jean-Louis; Janssens, Wim

    2017-01-01

    This study aimed to identify simple rules for allocating chronic obstructive pulmonary disease (COPD) patients to clinical phenotypes identified by cluster analyses. Data from 2409 COPD patients of French/Belgian COPD cohorts were analysed using cluster analysis resulting in the identification...... of subgroups, for which clinical relevance was determined by comparing 3-year all-cause mortality. Classification and regression trees (CARTs) were used to develop an algorithm for allocating patients to these subgroups. This algorithm was tested in 3651 patients from the COPD Cohorts Collaborative...... International Assessment (3CIA) initiative. Cluster analysis identified five subgroups of COPD patients with different clinical characteristics (especially regarding severity of respiratory disease and the presence of cardiovascular comorbidities and diabetes). The CART-based algorithm indicated...

  3. Greedy Algorithms for Nonnegativity-Constrained Simultaneous Sparse Recovery

    Science.gov (United States)

    Kim, Daeun; Haldar, Justin P.

    2016-01-01

    This work proposes a family of greedy algorithms to jointly reconstruct a set of vectors that are (i) nonnegative and (ii) simultaneously sparse with a shared support set. The proposed algorithms generalize previous approaches that were designed to impose these constraints individually. Similar to previous greedy algorithms for sparse recovery, the proposed algorithms iteratively identify promising support indices. In contrast to previous approaches, the support index selection procedure has been adapted to prioritize indices that are consistent with both the nonnegativity and shared support constraints. Empirical results demonstrate for the first time that the combined use of simultaneous sparsity and nonnegativity constraints can substantially improve recovery performance relative to existing greedy algorithms that impose less signal structure. PMID:26973368

  4. The 10/66 Dementia Research Group's fully operationalised DSM-IV dementia computerized diagnostic algorithm, compared with the 10/66 dementia algorithm and a clinician diagnosis: a population validation study

    Directory of Open Access Journals (Sweden)

    Krishnamoorthy ES

    2008-06-01

    Full Text Available Abstract Background The criterion for dementia implicit in DSM-IV is widely used in research but not fully operationalised. The 10/66 Dementia Research Group sought to do this using assessments from their one phase dementia diagnostic research interview, and to validate the resulting algorithm in a population-based study in Cuba. Methods The criterion was operationalised as a computerised algorithm, applying clinical principles, based upon the 10/66 cognitive tests, clinical interview and informant reports; the Community Screening Instrument for Dementia, the CERAD 10 word list learning and animal naming tests, the Geriatric Mental State, and the History and Aetiology Schedule – Dementia Diagnosis and Subtype. This was validated in Cuba against a local clinician DSM-IV diagnosis and the 10/66 dementia diagnosis (originally calibrated probabilistically against clinician DSM-IV diagnoses in the 10/66 pilot study. Results The DSM-IV sub-criteria were plausibly distributed among clinically diagnosed dementia cases and controls. The clinician diagnoses agreed better with 10/66 dementia diagnosis than with the more conservative computerized DSM-IV algorithm. The DSM-IV algorithm was particularly likely to miss less severe dementia cases. Those with a 10/66 dementia diagnosis who did not meet the DSM-IV criterion were less cognitively and functionally impaired compared with the DSMIV confirmed cases, but still grossly impaired compared with those free of dementia. Conclusion The DSM-IV criterion, strictly applied, defines a narrow category of unambiguous dementia characterized by marked impairment. It may be specific but incompletely sensitive to clinically relevant cases. The 10/66 dementia diagnosis defines a broader category that may be more sensitive, identifying genuine cases beyond those defined by our DSM-IV algorithm, with relevance to the estimation of the population burden of this disorder.

  5. Detection of algorithmic trading

    Science.gov (United States)

    Bogoev, Dimitar; Karam, Arzé

    2017-10-01

    We develop a new approach to reflect the behavior of algorithmic traders. Specifically, we provide an analytical and tractable way to infer patterns of quote volatility and price momentum consistent with different types of strategies employed by algorithmic traders, and we propose two ratios to quantify these patterns. Quote volatility ratio is based on the rate of oscillation of the best ask and best bid quotes over an extremely short period of time; whereas price momentum ratio is based on identifying patterns of rapid upward or downward movement in prices. The two ratios are evaluated across several asset classes. We further run a two-stage Artificial Neural Network experiment on the quote volatility ratio; the first stage is used to detect the quote volatility patterns resulting from algorithmic activity, while the second is used to validate the quality of signal detection provided by our measure.

  6. A Record Linkage Protocol for a Diabetes Registry at Ethnically Diverse Community Health Centers

    OpenAIRE

    Maizlish, Neil A.; Herrera, Linda

    2005-01-01

    Community health centers serve ethnically diverse populations that may pose challenges for record linkage based on name and date of birth. The objective was to identify an optimal deterministic algorithm to link patient encounters and laboratory results for hemoglobin A1c testing and examine its variability by health center site, patient ethnicity, and other variables. Based on data elements of last name, first name, date of birth, gender, and health center site, matches with ≥50% to < 100% o...

  7. A scalable community detection algorithm for large graphs using stochastic block models

    KAUST Repository

    Peng, Chengbin

    2017-11-24

    Community detection in graphs is widely used in social and biological networks, and the stochastic block model is a powerful probabilistic tool for describing graphs with community structures. However, in the era of

  8. A scalable community detection algorithm for large graphs using stochastic block models

    KAUST Repository

    Peng, Chengbin; Zhang, Zhihua; Wong, Ka-Chun; Zhang, Xiangliang; Keyes, David E.

    2017-01-01

    Community detection in graphs is widely used in social and biological networks, and the stochastic block model is a powerful probabilistic tool for describing graphs with community structures. However, in the era of

  9. Identifying Overlapping Language Communities: The Case of Chiriquí and Panamanian Signed Languages

    Science.gov (United States)

    Parks, Elizabeth S.

    2016-01-01

    In this paper, I use a holographic metaphor to explain the identification of overlapping sign language communities in Panama. By visualizing Panama's complex signing communities as emitting community "hotspots" through social drama on multiple stages, I employ ethnographic methods to explore overlapping contours of Panama's sign language…

  10. Comparison of Firefly algorithm and Artificial Immune System algorithm for lot streaming in -machine flow shop scheduling

    Directory of Open Access Journals (Sweden)

    G. Vijay Chakaravarthy

    2012-11-01

    Full Text Available Lot streaming is a technique used to split the processing of lots into several sublots (transfer batches to allow the overlapping of operations in a multistage manufacturing systems thereby shortening the production time (makespan. The objective of this paper is to minimize the makespan and total flow time of -job, -machine lot streaming problem in a flow shop with equal and variable size sublots and also to determine the optimal sublot size. In recent times researchers are concentrating and applying intelligent heuristics to solve flow shop problems with lot streaming. In this research, Firefly Algorithm (FA and Artificial Immune System (AIS algorithms are used to solve the problem. The results obtained by the proposed algorithms are also compared with the performance of other worked out traditional heuristics. The computational results shows that the identified algorithms are more efficient, effective and better than the algorithms already tested for this problem.

  11. Identifying significant temporal variation in time course microarray data without replicates

    Directory of Open Access Journals (Sweden)

    Porter Weston

    2009-03-01

    Full Text Available Abstract Background An important component of time course microarray studies is the identification of genes that demonstrate significant time-dependent variation in their expression levels. Until recently, available methods for performing such significance tests required replicates of individual time points. This paper describes a replicate-free method that was developed as part of a study of the estrous cycle in the rat mammary gland in which no replicate data was collected. Results A temporal test statistic is proposed that is based on the degree to which data are smoothed when fit by a spline function. An algorithm is presented that uses this test statistic together with a false discovery rate method to identify genes whose expression profiles exhibit significant temporal variation. The algorithm is tested on simulated data, and is compared with another recently published replicate-free method. The simulated data consists both of genes with known temporal dependencies, and genes from a null distribution. The proposed algorithm identifies a larger percentage of the time-dependent genes for a given false discovery rate. Use of the algorithm in a study of the estrous cycle in the rat mammary gland resulted in the identification of genes exhibiting distinct circadian variation. These results were confirmed in follow-up laboratory experiments. Conclusion The proposed algorithm provides a new approach for identifying expression profiles with significant temporal variation without relying on replicates. When compared with a recently published algorithm on simulated data, the proposed algorithm appears to identify a larger percentage of time-dependent genes for a given false discovery rate. The development of the algorithm was instrumental in revealing the presence of circadian variation in the virgin rat mammary gland during the estrous cycle.

  12. New calibration algorithms for dielectric-based microwave moisture sensors

    Science.gov (United States)

    New calibration algorithms for determining moisture content in granular and particulate materials from measurement of the dielectric properties at a single microwave frequency are proposed. The algorithms are based on identifying empirically correlations between the dielectric properties and the par...

  13. Algebraic dynamics algorithm: Numerical comparison with Runge-Kutta algorithm and symplectic geometric algorithm

    Institute of Scientific and Technical Information of China (English)

    WANG ShunJin; ZHANG Hua

    2007-01-01

    Based on the exact analytical solution of ordinary differential equations,a truncation of the Taylor series of the exact solution to the Nth order leads to the Nth order algebraic dynamics algorithm.A detailed numerical comparison is presented with Runge-Kutta algorithm and symplectic geometric algorithm for 12 test models.The results show that the algebraic dynamics algorithm can better preserve both geometrical and dynamical fidelity of a dynamical system at a controllable precision,and it can solve the problem of algorithm-induced dissipation for the Runge-Kutta algorithm and the problem of algorithm-induced phase shift for the symplectic geometric algorithm.

  14. Algebraic dynamics algorithm:Numerical comparison with Runge-Kutta algorithm and symplectic geometric algorithm

    Institute of Scientific and Technical Information of China (English)

    2007-01-01

    Based on the exact analytical solution of ordinary differential equations, a truncation of the Taylor series of the exact solution to the Nth order leads to the Nth order algebraic dynamics algorithm. A detailed numerical comparison is presented with Runge-Kutta algorithm and symplectic geometric algorithm for 12 test models. The results show that the algebraic dynamics algorithm can better preserve both geometrical and dynamical fidelity of a dynamical system at a controllable precision, and it can solve the problem of algorithm-induced dissipation for the Runge-Kutta algorithm and the problem of algorithm-induced phase shift for the symplectic geometric algorithm.

  15. A novel baiting microcosm approach used to identify the bacterial community associated with Penicillium bilaii hyphae in soil

    DEFF Research Database (Denmark)

    Ghodsalavi, Behnoushsadat; Svenningsen, Nanna Bygvraa; Hao, Xiuli

    2017-01-01

    It is important to identify and recover bacteria associating with fungi under natural soil conditions to enable eco-physiological studies, and to facilitate the use of bacterial-fungal consortia in environmental biotechnology. We have developed a novel type of baiting microcosm, where fungal hyphae...... interact with bacteria under close-to-natural soil conditions; an advantage compared to model systems that determine fungal influences on bacterial communities in laboratory media. In the current approach, the hyphae are placed on a solid support, which enables the recovery of hyphae with associated...... bacteria in contrast to model systems that compare bulk soil and mycosphere soil. We used the baiting microcosm approach to determine, for the first time, the composition of the bacterial community associating in the soil with hyphae of the phosphate-solubilizer, Penicillium bilaii. By applying...

  16. IDENTIFYING CONCERNS OF POSTGRADUATES IN COMMUNITY MEDICINE USING A QUALITATIVE RESEARCH METHOD- VISUALISATION IN PARTICIPATORY PROGRAMMES (VIPP

    Directory of Open Access Journals (Sweden)

    Vinay Babu Koganti

    2017-06-01

    Full Text Available BACKGROUND Postgraduation in Community Medicine finds few takers and those who do take it up as a career option have many concerns regarding the course. To understand the issues involved, a qualitative method called VIPP was used, which is a people centered approach to identify issues from the perspectives of those involved. This study is set to identify the problems faced by postgraduate students in Community Medicine regarding their course of study. MATERIALS AND METHODS This study was conducted during a regional postgraduate CME of the NTR University of Health Sciences, Andhra Pradesh. Postgraduates and junior faculty from 5 medical colleges in the region were involved in the exercise after taking their informed consent. Visualisation in Participatory Programmes (VIPP, a qualitative method was used as a means of obtaining information followed by a discussion with visual display of all the mentioned items. RESULTS The themes that emerged are problems faced due to the student’s felt inadequacies, faculty shortcomings, issues regarding the department/college management and lacunae in the course structure and implementation. CONCLUSION In VIPP, sensitive issues are visually displayed for all to see and contemplate. Many of the student’s issues were actually brought on by poor curriculum planning and implementation. This was also undermining students’ self-esteem and causing anxiety about future career prospects.

  17. Identifying role of perceived quality and satisfaction on the utilization status of the community clinic services; Bangladesh context.

    Science.gov (United States)

    Karim, Rizwanul M; Abdullah, Mamun S; Rahman, Anisur M; Alam, Ashraful M

    2016-06-24

    Bangladesh is one among the few countries of the world that provides free medical services at the community level through various public health facilities. It is now evident that, clients' perceived quality of services and their expectations of service standards affect health service utilization to a great extent. The aim of the study was to develop and validate the measures for perception and satisfaction of primary health care quality in Bangladesh context and to identify their aspects on the utilization status of the Community Clinic services. This mixed method cross sectional survey was conducted from January to June 2012, in the catchment area of 12 community clinics. Since most of the outcome indicators focus mainly on women and children, women having children less than 2 years of age were randomly assigned and interviewed for the study purpose. Data were collected through FGD, Key informants interview and a pretested semi- structured questionnaire. About 95 % of the respondents were Muslims and 5 % were Hindus. The average age of the respondents was 23.38 (SD 4.15) and almost all of them are home makers. The average monthly expenditure of their family was 95US $ (SD 32US$). At the beginning of the study, two psychometric research instruments; 24 items perceived quality of primary care services PQPCS scale (chronbach's α = .89) and 22 items community clinic service satisfaction CCSS scale (chronbach's α = .97), were constructed and validated. This study showed less educated, poor, landless mothers utilized the community clinic services more than their educated and wealthier counterpart. Women who lived in their own residence used the community clinic services more frequently than those who lived in a rental house. Perceptions concerning skill and competence of the health care provider and satisfaction indicating interpersonal communication and attitude of the care provider were important predictors for community clinic service utilization

  18. Extracting weights from edge directions to find communities in directed networks

    International Nuclear Information System (INIS)

    Lai, Darong; Lu, Hongtao; Nardini, Christine

    2010-01-01

    Community structures are found to exist ubiquitously in real-world complex networks. We address here the problem of community detection in directed networks. Most of the previous literature ignores edge directions and applies methods designed for community detection in undirected networks, which discards valuable information and often fails when different communities are defined on the basis of incoming and outgoing edges. We suggest extracting information about edge directions using a PageRank random walk and translating such information into edge weights. After extraction we obtain a new weighted directed network in which edge directions can then be safely ignored. We thus transform community detection in directed networks into community detection in reweighted undirected networks. Such an approach can benefit directly from the large volume of algorithms for the detection of communities in undirected networks already developed, since it is not obvious how to extend these algorithms to account for directed networks and the procedure is often difficult. Validations on synthetic and real-world networks demonstrate that the proposed framework can effectively detect communities in directed networks

  19. 78 FR 57639 - Request for Comments on Pediatric Planned Procedure Algorithm

    Science.gov (United States)

    2013-09-19

    ... Comments on Pediatric Planned Procedure Algorithm AGENCY: Agency for Healthcare Research and Quality (AHRQ), HHS. ACTION: Notice of request for comments on pediatric planned procedure algorithm from the members... Quality (AHRQ) is requesting comments from the public on an algorithm for identifying pediatric planned...

  20. An improved algorithm for connectivity analysis of distribution networks

    International Nuclear Information System (INIS)

    Kansal, M.L.; Devi, Sunita

    2007-01-01

    In the present paper, an efficient algorithm for connectivity analysis of moderately sized distribution networks has been suggested. Algorithm is based on generation of all possible minimal system cutsets. The algorithm is efficient as it identifies only the necessary and sufficient conditions of system failure conditions in n-out-of-n type of distribution networks. The proposed algorithm is demonstrated with the help of saturated and unsaturated distribution networks. The computational efficiency of the algorithm is justified by comparing the computational efforts with the previously suggested appended spanning tree (AST) algorithm. The proposed technique has the added advantage as it can be utilized for generation of system inequalities which is useful in reliability estimation of capacitated networks

  1. The Algorithm for Algorithms: An Evolutionary Algorithm Based on Automatic Designing of Genetic Operators

    Directory of Open Access Journals (Sweden)

    Dazhi Jiang

    2015-01-01

    Full Text Available At present there is a wide range of evolutionary algorithms available to researchers and practitioners. Despite the great diversity of these algorithms, virtually all of the algorithms share one feature: they have been manually designed. A fundamental question is “are there any algorithms that can design evolutionary algorithms automatically?” A more complete definition of the question is “can computer construct an algorithm which will generate algorithms according to the requirement of a problem?” In this paper, a novel evolutionary algorithm based on automatic designing of genetic operators is presented to address these questions. The resulting algorithm not only explores solutions in the problem space like most traditional evolutionary algorithms do, but also automatically generates genetic operators in the operator space. In order to verify the performance of the proposed algorithm, comprehensive experiments on 23 well-known benchmark optimization problems are conducted. The results show that the proposed algorithm can outperform standard differential evolution algorithm in terms of convergence speed and solution accuracy which shows that the algorithm designed automatically by computers can compete with the algorithms designed by human beings.

  2. A New Feedback-Analysis based Reputation Algorithm for E-Commerce Communities

    Directory of Open Access Journals (Sweden)

    Hasnae Rahimi

    2014-12-01

    Full Text Available Dealing with the ever-growing content generated by users in the e-commerce applications, Trust Reputation Systems (TRS are widely used online to provide the trust reputation of each product using the customers’ ratings. However, there is also a good number of online customer reviews and feedback that must be used by the TRS. As a result, we propose in this work a new architecture for TRS in e-commerce application which includes feedback’ mining in order to calculate reputation scores. This architecture is based on an intelligent layer that proposes to each user (i.e. “feedback provider” who has already given his recommendation, a collection of prefabricated feedback to like or dislike. Then the proposed reputation algorithm calculates the trust degree of the user, the feedback’s trustworthiness and generates the global reputation score of the product according to his ‘likes’ and ‘dislikes’. In this work, we present also a state of the art of text mining tools and algorithms that can be used to generate the prefabricated feedback and to classify them into different categories.

  3. Astronomy and big data a data clustering approach to identifying uncertain galaxy morphology

    CERN Document Server

    Edwards, Kieran Jay

    2014-01-01

    With the onset of massive cosmological data collection through media such as the Sloan Digital Sky Survey (SDSS), galaxy classification has been accomplished for the most part with the help of citizen science communities like Galaxy Zoo. Seeking the wisdom of the crowd for such Big Data processing has proved extremely beneficial. However, an analysis of one of the Galaxy Zoo morphological classification data sets has shown that a significant majority of all classified galaxies are labelled as “Uncertain”. This book reports on how to use data mining, more specifically clustering, to identify galaxies that the public has shown some degree of uncertainty for as to whether they belong to one morphology type or another. The book shows the importance of transitions between different data mining techniques in an insightful workflow. It demonstrates that Clustering enables to identify discriminating features in the analysed data sets, adopting a novel feature selection algorithms called Incremental Feature Select...

  4. Next Generation Suspension Dynamics Algorithms

    Energy Technology Data Exchange (ETDEWEB)

    Schunk, Peter Randall [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Higdon, Jonathon [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Chen, Steven [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2014-12-01

    This research project has the objective to extend the range of application, improve the efficiency and conduct simulations with the Fast Lubrication Dynamics (FLD) algorithm for concentrated particle suspensions in a Newtonian fluid solvent. The research involves a combination of mathematical development, new computational algorithms, and application to processing flows of relevance in materials processing. The mathematical developments clarify the underlying theory, facilitate verification against classic monographs in the field and provide the framework for a novel parallel implementation optimized for an OpenMP shared memory environment. The project considered application to consolidation flows of major interest in high throughput materials processing and identified hitherto unforeseen challenges in the use of FLD in these applications. Extensions to the algorithm have been developed to improve its accuracy in these applications.

  5. Evaluation of the performance of existing non-laboratory based cardiovascular risk assessment algorithms

    Science.gov (United States)

    2013-01-01

    Background The high burden and rising incidence of cardiovascular disease (CVD) in resource constrained countries necessitates implementation of robust and pragmatic primary and secondary prevention strategies. Many current CVD management guidelines recommend absolute cardiovascular (CV) risk assessment as a clinically sound guide to preventive and treatment strategies. Development of non-laboratory based cardiovascular risk assessment algorithms enable absolute risk assessment in resource constrained countries. The objective of this review is to evaluate the performance of existing non-laboratory based CV risk assessment algorithms using the benchmarks for clinically useful CV risk assessment algorithms outlined by Cooney and colleagues. Methods A literature search to identify non-laboratory based risk prediction algorithms was performed in MEDLINE, CINAHL, Ovid Premier Nursing Journals Plus, and PubMed databases. The identified algorithms were evaluated using the benchmarks for clinically useful cardiovascular risk assessment algorithms outlined by Cooney and colleagues. Results Five non-laboratory based CV risk assessment algorithms were identified. The Gaziano and Framingham algorithms met the criteria for appropriateness of statistical methods used to derive the algorithms and endpoints. The Swedish Consultation, Framingham and Gaziano algorithms demonstrated good discrimination in derivation datasets. Only the Gaziano algorithm was externally validated where it had optimal discrimination. The Gaziano and WHO algorithms had chart formats which made them simple and user friendly for clinical application. Conclusion Both the Gaziano and Framingham non-laboratory based algorithms met most of the criteria outlined by Cooney and colleagues. External validation of the algorithms in diverse samples is needed to ascertain their performance and applicability to different populations and to enhance clinicians’ confidence in them. PMID:24373202

  6. Community detection in complex networks using proximate support vector clustering

    Science.gov (United States)

    Wang, Feifan; Zhang, Baihai; Chai, Senchun; Xia, Yuanqing

    2018-03-01

    Community structure, one of the most attention attracting properties in complex networks, has been a cornerstone in advances of various scientific branches. A number of tools have been involved in recent studies concentrating on the community detection algorithms. In this paper, we propose a support vector clustering method based on a proximity graph, owing to which the introduced algorithm surpasses the traditional support vector approach both in accuracy and complexity. Results of extensive experiments undertaken on computer generated networks and real world data sets illustrate competent performances in comparison with the other counterparts.

  7. Identifying solutions to increase participation in physical activity interventions within a socio-economically disadvantaged community: a qualitative study.

    Science.gov (United States)

    Cleland, Claire L; Hunter, Ruth F; Tully, Mark A; Scott, David; Kee, Frank; Donnelly, Michael; Prior, Lindsay; Cupples, Margaret E

    2014-05-23

    There is an urgent need to increase population levels of physical activity, particularly amongst those who are socio-economically disadvantaged. Multiple factors influence physical activity behaviour but the generalisability of current evidence to such 'hard-to-reach' population subgroups is limited by difficulties in recruiting them into studies. Also, rigorous qualitative studies of lay perceptions and perceptions of community leaders about public health efforts to increase physical activity are sparse. We sought to explore, within a socio-economically disadvantaged community, residents' and community leaders' perceptions of physical activity (PA) interventions and issues regarding their implementation, in order to improve understanding of needs, expectations, and social/environmental factors relevant to future interventions. Within an ongoing regeneration project (Connswater Community Greenway), in a socio-economically disadvantaged community in Belfast, we collaborated with a Community Development Agency to purposively sample leaders from public- and voluntary-sector community groups and residents. Individual semi-structured interviews were conducted with 12 leaders. Residents (n = 113), of both genders and a range of ages (14 to 86 years) participated in focus groups (n = 14) in local facilities. Interviews and focus groups were recorded, transcribed verbatim and analysed using a thematic framework. Three main themes were identified: awareness of PA interventions; factors contributing to intervention effectiveness; and barriers to participation in PA interventions. Participants reported awareness only of interventions in which they were involved directly, highlighting a need for better communications, both inter- and intra-sectoral, and with residents. Meaningful engagement of residents in planning/organisation, tailoring to local context, supporting volunteers, providing relevant resources and an 'exit strategy' were perceived as important factors

  8. Generation of Referring Expressions: Assessing the Incremental Algorithm

    Science.gov (United States)

    van Deemter, Kees; Gatt, Albert; van der Sluis, Ielka; Power, Richard

    2012-01-01

    A substantial amount of recent work in natural language generation has focused on the generation of "one-shot" referring expressions whose only aim is to identify a target referent. Dale and Reiter's Incremental Algorithm (IA) is often thought to be the best algorithm for maximizing the similarity to referring expressions produced by people. We…

  9. Network-based recommendation algorithms: A review

    Science.gov (United States)

    Yu, Fei; Zeng, An; Gillard, Sébastien; Medo, Matúš

    2016-06-01

    Recommender systems are a vital tool that helps us to overcome the information overload problem. They are being used by most e-commerce web sites and attract the interest of a broad scientific community. A recommender system uses data on users' past preferences to choose new items that might be appreciated by a given individual user. While many approaches to recommendation exist, the approach based on a network representation of the input data has gained considerable attention in the past. We review here a broad range of network-based recommendation algorithms and for the first time compare their performance on three distinct real datasets. We present recommendation topics that go beyond the mere question of which algorithm to use-such as the possible influence of recommendation on the evolution of systems that use it-and finally discuss open research directions and challenges.

  10. Dementia Population Risk Tool (DemPoRT): study protocol for a predictive algorithm assessing dementia risk in the community.

    Science.gov (United States)

    Fisher, Stacey; Hsu, Amy; Mojaverian, Nassim; Taljaard, Monica; Huyer, Gregory; Manuel, Douglas G; Tanuseputro, Peter

    2017-10-24

    The burden of disease from dementia is a growing global concern as incidence increases dramatically with age, and average life expectancy has been increasing around the world. Planning for an ageing population requires reliable projections of dementia prevalence; however, existing population projections are simple and have poor predictive accuracy. The Dementia Population Risk Tool (DemPoRT) will predict incidence of dementia in the population setting using multivariable modelling techniques and will be used to project dementia prevalence. The derivation cohort will consist of elderly Ontario respondents of the Canadian Community Health Survey (CCHS) (2001, 2003, 2005 and 2007; 18 764 males and 25 288 females). Prespecified predictors include sociodemographic, general health, behavioural, functional and health condition variables. Incident dementia will be identified through individual linkage of survey respondents to population-level administrative healthcare databases (1797 and 3281 events, and 117 795 and 166 573 person-years of follow-up, for males and females, respectively, until 31 March 2014). Using time of first dementia capture as the primary outcome and death as a competing risk, sex-specific proportional hazards regression models will be estimated. The 2008/2009 CCHS survey will be used for validation (approximately 4600 males and 6300 females). Overall calibration and discrimination will be assessed as well as calibration within predefined subgroups of importance to clinicians and policy makers. Research ethics approval has been granted by the Ottawa Health Science Network Research Ethics Board. DemPoRT results will be submitted for publication in peer-review journals and presented at scientific meetings. The algorithm will be assessable online for both population and individual uses. ClinicalTrials.gov NCT03155815, pre-results. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No

  11. Assessment of a Novel Approach to Identify Trichiasis Cases Using Community Treatment Assistants in Tanzania.

    Directory of Open Access Journals (Sweden)

    Gregory S Greene

    2015-12-01

    Full Text Available Simple surgical intervention advocated by the World Health Organization can alleviate trachomatous trichiasis (TT and prevent subsequent blindness. A large backlog of TT cases remain unidentified and untreated. To increase identification and referral of TT cases, a novel approach using standard screening questions, a card, and simple training for Community Treatment Assistants (CTAs to use during Mass Drug Administration (MDA was developed and evaluated in Kongwa District, a trachoma-endemic area of central Tanzania.A community randomized trial was conducted in 36 communities during MDA. CTAs in intervention villages received an additional half-day of training and a TT screening card in addition to the training received by CTAs in villages assigned to usual care. All MDA participants 15 years and older were screened for TT, and senior TT graders confirmed case status by evaluating all screened-positive cases. A random sample of those screened negative for TT and those who did not present at MDA were also evaluated by the master graders. Intervention CTAs identified 5.6 times as many cases (n = 50 as those assigned to usual care (n = 9, p < 0.05. While specificity was above 90% for both groups, the sensitivity for the novel screening tool was 31.2% compared to 5.6% for the usual care group (p < 0.05.CTAs appear to be viable resources for the identification of TT cases. Additional training and use of a TT screening card significantly increased the ability of CTAs to recognize and refer TT cases during MDA; however, further efforts are needed to improve case detection and reduce the number of false positive cases.

  12. Can community members identify tropical tree species for REDD+ carbon and biodiversity measurements?

    DEFF Research Database (Denmark)

    Zhao, Mingxu; Brofeldt, Søren; Li, Qiaohong

    2016-01-01

    to take advantage of the same data for detecting changes in the tree diversity, using the richness and abundance of canopy trees as a proxy for biodiversity. If local community members are already assessing the above-ground biomass in a representative network of forest vegetation plots, it may require...... minimal further effort to collect data on the diversity of trees. We compare community members and trained scientists' data on tree diversity in permanent vegetation plots in montane forest in Yunnan, China. We show that local community members here can collect tree diversity data of comparable quality...... to trained botanists, at one third the cost. Without access to herbaria, identification guides or the Internet, community members could provide the ethno-taxonomical names for 95% of 1071 trees in 60 vegetation plots. Moreover, we show that the community-led survey spent 89% of the expenses at village level...

  13. Evolutionary algorithms for mobile ad hoc networks

    CERN Document Server

    Dorronsoro, Bernabé; Danoy, Grégoire; Pigné, Yoann; Bouvry, Pascal

    2014-01-01

    Describes how evolutionary algorithms (EAs) can be used to identify, model, and minimize day-to-day problems that arise for researchers in optimization and mobile networking. Mobile ad hoc networks (MANETs), vehicular networks (VANETs), sensor networks (SNs), and hybrid networks—each of these require a designer’s keen sense and knowledge of evolutionary algorithms in order to help with the common issues that plague professionals involved in optimization and mobile networking. This book introduces readers to both mobile ad hoc networks and evolutionary algorithms, presenting basic concepts as well as detailed descriptions of each. It demonstrates how metaheuristics and evolutionary algorithms (EAs) can be used to help provide low-cost operations in the optimization process—allowing designers to put some “intelligence” or sophistication into the design. It also offers efficient and accurate information on dissemination algorithms topology management, and mobility models to address challenges in the ...

  14. Data clustering algorithms and applications

    CERN Document Server

    Aggarwal, Charu C

    2013-01-01

    Research on the problem of clustering tends to be fragmented across the pattern recognition, database, data mining, and machine learning communities. Addressing this problem in a unified way, Data Clustering: Algorithms and Applications provides complete coverage of the entire area of clustering, from basic methods to more refined and complex data clustering approaches. It pays special attention to recent issues in graphs, social networks, and other domains.The book focuses on three primary aspects of data clustering: Methods, describing key techniques commonly used for clustering, such as fea

  15. Identify High-Quality Protein Structural Models by Enhanced K-Means.

    Science.gov (United States)

    Wu, Hongjie; Li, Haiou; Jiang, Min; Chen, Cheng; Lv, Qiang; Wu, Chuang

    2017-01-01

    Background. One critical issue in protein three-dimensional structure prediction using either ab initio or comparative modeling involves identification of high-quality protein structural models from generated decoys. Currently, clustering algorithms are widely used to identify near-native models; however, their performance is dependent upon different conformational decoys, and, for some algorithms, the accuracy declines when the decoy population increases. Results. Here, we proposed two enhanced K -means clustering algorithms capable of robustly identifying high-quality protein structural models. The first one employs the clustering algorithm SPICKER to determine the initial centroids for basic K -means clustering ( SK -means), whereas the other employs squared distance to optimize the initial centroids ( K -means++). Our results showed that SK -means and K -means++ were more robust as compared with SPICKER alone, detecting 33 (59%) and 42 (75%) of 56 targets, respectively, with template modeling scores better than or equal to those of SPICKER. Conclusions. We observed that the classic K -means algorithm showed a similar performance to that of SPICKER, which is a widely used algorithm for protein-structure identification. Both SK -means and K -means++ demonstrated substantial improvements relative to results from SPICKER and classical K -means.

  16. A database of phylogenetically atypical genes in archaeal and bacterial genomes, identified using the DarkHorse algorithm

    Directory of Open Access Journals (Sweden)

    Allen Eric E

    2008-10-01

    Full Text Available Abstract Background The process of horizontal gene transfer (HGT is believed to be widespread in Bacteria and Archaea, but little comparative data is available addressing its occurrence in complete microbial genomes. Collection of high-quality, automated HGT prediction data based on phylogenetic evidence has previously been impractical for large numbers of genomes at once, due to prohibitive computational demands. DarkHorse, a recently described statistical method for discovering phylogenetically atypical genes on a genome-wide basis, provides a means to solve this problem through lineage probability index (LPI ranking scores. LPI scores inversely reflect phylogenetic distance between a test amino acid sequence and its closest available database matches. Proteins with low LPI scores are good horizontal gene transfer candidates; those with high scores are not. Description The DarkHorse algorithm has been applied to 955 microbial genome sequences, and the results organized into a web-searchable relational database, called the DarkHorse HGT Candidate Resource http://darkhorse.ucsd.edu. Users can select individual genomes or groups of genomes to screen by LPI score, search for protein functions by descriptive annotation or amino acid sequence similarity, or select proteins with unusual G+C composition in their underlying coding sequences. The search engine reports LPI scores for match partners as well as query sequences, providing the opportunity to explore whether potential HGT donor sequences are phylogenetically typical or atypical within their own genomes. This information can be used to predict whether or not sufficient information is available to build a well-supported phylogenetic tree using the potential donor sequence. Conclusion The DarkHorse HGT Candidate database provides a powerful, flexible set of tools for identifying phylogenetically atypical proteins, allowing researchers to explore both individual HGT events in single genomes, and

  17. Can Community Members Identify Tropical Tree Species for REDD+ Carbon and Biodiversity Measurements?

    Science.gov (United States)

    Zhao, Mingxu; Brofeldt, Søren; Li, Qiaohong; Xu, Jianchu; Danielsen, Finn; Læssøe, Simon Bjarke Lægaard; Poulsen, Michael Køie; Gottlieb, Anna; Maxwell, James Franklin; Theilade, Ida

    2016-01-01

    Biodiversity conservation is a required co-benefit of REDD+. Biodiversity monitoring is therefore needed, yet in most areas it will be constrained by limitations in the available human professional and financial resources. REDD+ programs that use forest plots for biomass monitoring may be able to take advantage of the same data for detecting changes in the tree diversity, using the richness and abundance of canopy trees as a proxy for biodiversity. If local community members are already assessing the above-ground biomass in a representative network of forest vegetation plots, it may require minimal further effort to collect data on the diversity of trees. We compare community members and trained scientists' data on tree diversity in permanent vegetation plots in montane forest in Yunnan, China. We show that local community members here can collect tree diversity data of comparable quality to trained botanists, at one third the cost. Without access to herbaria, identification guides or the Internet, community members could provide the ethno-taxonomical names for 95% of 1071 trees in 60 vegetation plots. Moreover, we show that the community-led survey spent 89% of the expenses at village level as opposed to 23% of funds in the monitoring by botanists. In participatory REDD+ programs in areas where community members demonstrate great knowledge of forest trees, community-based collection of tree diversity data can be a cost-effective approach for obtaining tree diversity information.

  18. Can Community Members Identify Tropical Tree Species for REDD+ Carbon and Biodiversity Measurements?

    Directory of Open Access Journals (Sweden)

    Mingxu Zhao

    Full Text Available Biodiversity conservation is a required co-benefit of REDD+. Biodiversity monitoring is therefore needed, yet in most areas it will be constrained by limitations in the available human professional and financial resources. REDD+ programs that use forest plots for biomass monitoring may be able to take advantage of the same data for detecting changes in the tree diversity, using the richness and abundance of canopy trees as a proxy for biodiversity. If local community members are already assessing the above-ground biomass in a representative network of forest vegetation plots, it may require minimal further effort to collect data on the diversity of trees. We compare community members and trained scientists' data on tree diversity in permanent vegetation plots in montane forest in Yunnan, China. We show that local community members here can collect tree diversity data of comparable quality to trained botanists, at one third the cost. Without access to herbaria, identification guides or the Internet, community members could provide the ethno-taxonomical names for 95% of 1071 trees in 60 vegetation plots. Moreover, we show that the community-led survey spent 89% of the expenses at village level as opposed to 23% of funds in the monitoring by botanists. In participatory REDD+ programs in areas where community members demonstrate great knowledge of forest trees, community-based collection of tree diversity data can be a cost-effective approach for obtaining tree diversity information.

  19. Identifying paediatric nursing-sensitive outcomes in linked administrative health data

    Directory of Open Access Journals (Sweden)

    Wilson Sally

    2012-07-01

    Full Text Available Abstract Background There is increasing interest in the contribution of the quality of nursing care to patient outcomes. Due to different casemix and risk profiles, algorithms for administrative health data that identify nursing-sensitive outcomes in adult hospitalised patients may not be applicable to paediatric patients. The study purpose was to test adult algorithms in a paediatric hospital population and make amendments to increase the accuracy of identification of hospital acquired events. The study also aimed to determine whether the use of linked hospital records improved the likelihood of correctly identifying patient outcomes as nursing sensitive rather than being related to their pre-morbid conditions. Methods Using algorithms developed by Needleman et al. (2001, proportions and rates of records that identified nursing-sensitive outcomes for pressure ulcers, pneumonia and surgical wound infections were determined from administrative hospitalisation data for all paediatric patients discharged from a tertiary paediatric hospital in Western Australia between July 1999 and June 2009. The effects of changes to inclusion and exclusion criteria for each algorithm on the calculated proportion or rate in the paediatric population were explored. Linked records were used to identify comorbid conditions that increased nursing-sensitive outcome risk. Rates were calculated using algorithms revised for paediatric patients. Results Linked records of 129,719 hospital separations for 79,016 children were analysed. Identification of comorbid conditions was enhanced through access to prior and/or subsequent hospitalisation records (43% of children with pressure ulcers had a form of paralysis recorded only on a previous admission. Readmissions with a surgical wound infection were identified for 103 (4.8/1,000 surgical separations using linked data. After amendment of each algorithm for paediatric patients, rates of pressure ulcers and pneumonia reduced by

  20. Identifying paediatric nursing-sensitive outcomes in linked administrative health data.

    Science.gov (United States)

    Wilson, Sally; Bremner, Alexandra P; Hauck, Yvonne; Finn, Judith

    2012-07-20

    There is increasing interest in the contribution of the quality of nursing care to patient outcomes. Due to different casemix and risk profiles, algorithms for administrative health data that identify nursing-sensitive outcomes in adult hospitalised patients may not be applicable to paediatric patients. The study purpose was to test adult algorithms in a paediatric hospital population and make amendments to increase the accuracy of identification of hospital acquired events. The study also aimed to determine whether the use of linked hospital records improved the likelihood of correctly identifying patient outcomes as nursing sensitive rather than being related to their pre-morbid conditions. Using algorithms developed by Needleman et al. (2001), proportions and rates of records that identified nursing-sensitive outcomes for pressure ulcers, pneumonia and surgical wound infections were determined from administrative hospitalisation data for all paediatric patients discharged from a tertiary paediatric hospital in Western Australia between July 1999 and June 2009. The effects of changes to inclusion and exclusion criteria for each algorithm on the calculated proportion or rate in the paediatric population were explored. Linked records were used to identify comorbid conditions that increased nursing-sensitive outcome risk. Rates were calculated using algorithms revised for paediatric patients. Linked records of 129,719 hospital separations for 79,016 children were analysed. Identification of comorbid conditions was enhanced through access to prior and/or subsequent hospitalisation records (43% of children with pressure ulcers had a form of paralysis recorded only on a previous admission). Readmissions with a surgical wound infection were identified for 103 (4.8/1,000) surgical separations using linked data. After amendment of each algorithm for paediatric patients, rates of pressure ulcers and pneumonia reduced by 53% and 15% (from 1.3 to 0.6 and from 9.1 to 7.7 per

  1. The use of think-aloud protocols to identify a decision-making process of community pharmacists aimed at improving CMS Star Ratings scores.

    Science.gov (United States)

    George, David L; Smith, Michael J; Draugalis, JoLaine R; Tolma, Eleni L; Keast, Shellie L; Wilson, Justin B

    2018-03-01

    The Center for Medicare and Medicaid Services (CMS) created the Star Rating system based on multiple measures that indicate the overall quality of health plans. Community pharmacists can impact certain Star Ratings measure scores through medication adherence and patient safety interventions. To explore methods, needs, and workflow issues of community pharmacists to improve CMS Star Ratings measures. Think-aloud protocols (TAPs) were conducted with active community retail pharmacists in Oklahoma. Each TAP was audio recorded and transcribed to documents for analysis. Analysts agreed on common themes, illuminated differences in findings, and saturation of the data gathered. Methods, needs, and workflow themes of community pharmacists associated with improving Star Ratings measures were compiled and organized to exhibit a decision-making process. Five TAPs were performed among three independent pharmacy owners, one multi-store owner, and one chain-store administrator. A thematically common 4-step process to monitor and improve CMS Star Ratings scores among participants was identified. To improve Star Ratings measures, pharmacists: 1) used technology to access scores, 2) analyzed data to strategically set goals, 3) assessed individual patient information for comprehensive assessment, and 4) decided on interventions to best impact Star Ratings scores. Participants also shared common needs, workflow issues, and benefits associated with methods used in improving Star Ratings. TAPs were useful in exploring processes of pharmacists who improve CMS Star Ratings scores. Pharmacists demonstrated and verbalized their methods, workflow issues, needs, and benefits related to performing the task. The themes and decision-making process identified to improving CMS Star Ratings scores will assist in the development of training and education programs for pharmacists in the community setting. Published by Elsevier Inc.

  2. Embedded gamma spectrometry: new algorithms for spectral analysis

    International Nuclear Information System (INIS)

    Martin-Burtart, Nicolas

    2012-01-01

    Airborne gamma spectrometry was first used for mining prospecting. Three main families were looked for: K-40, U-238 and Th-232. The Chernobyl accident acted as a trigger and for the last fifteen years, a lot of new systems have been developed for intervention in case of nuclear accident or environmental purposes. Depending on their uses, new algorithms were developed, mainly for medium or high energy signal extraction. These spectral regions are characteristics of natural emissions (K-40, U-238 and Th-232 decay chains) and fissions products (mainly Cs-137 and Co-60). Below 400 keV, where special nuclear materials emit, these methods can still be used but are greatly imprecise. A new algorithm called 2-windows (extended to 3), was developed. It allows an accurate extraction, taking the flight altitude into account to minimize false detection. Watching radioactive materials traffic appeared with homeland security policy a few years ago. This particular use of dedicated sensors require a new type of algorithms. Before, one algorithm was very efficient for a particular nuclide or spectral region. Now, we need algorithm able to detect an anomaly wherever it is and whatever it is: industrial, medical or SNM. This work identified two families of methods working under these circumstances. Finally, anomalies have to be identified. IAEA recommend to watch around 30 radionuclides. A brand new identification algorithm was developed, using several rays per element and avoiding identifications conflicts. (author) [fr

  3. Image-reconstruction algorithms for positron-emission tomography systems

    International Nuclear Information System (INIS)

    Cheng, S.N.C.

    1982-01-01

    The positional uncertainty in the time-of-flight measurement of a positron-emission tomography system is modelled as a Gaussian distributed random variable and the image is assumed to be piecewise constant on a rectilinear lattice. A reconstruction algorithm using maximum-likelihood estimation is derived for the situation in which time-of-flight data are sorted as the most-likely-position array. The algorithm is formulated as a linear system described by a nonseparable, block-banded, Toeplitz matrix, and a sine-transform technique is used to implement this algorithm efficiently. The reconstruction algorithms for both the most-likely-position array and the confidence-weighted array are described by similar equations, hence similar linear systems can be used to described the reconstruction algorithm for a discrete, confidence-weighted array, when the matrix and the entries in the data array are properly identified. It is found that the mean square-error depends on the ratio of the full width at half the maximum of time-of-flight measurement over the size of a pixel. When other parameters are fixed, the larger the pixel size, the smaller is the mean square-error. In the study of resolution, parameters that affect the impulse response of time-of-flight reconstruction algorithms are identified. It is found that the larger the pixel size, the larger is the standard deviation of the impulse response. This shows that small mean square-error and fine resolution are two contradictory requirements

  4. Vector neural net identifying many strongly distorted and correlated patterns

    Science.gov (United States)

    Kryzhanovsky, Boris V.; Mikaelian, Andrei L.; Fonarev, Anatoly B.

    2005-01-01

    We suggest an effective and simple algorithm providing a polynomial storage capacity of a network of the form M ~ N2s+1, where N is the dimension of the stored binary patterns. In this problem the value of the free parameter s is restricted by the inequalities N >> slnN >= 1. The algorithm allows us to identify a large number of highly distorted similar patterns. The negative influence of correlations of the patterns is suppressed by choosing a sufficiently large value of the parameter s. We show the efficiency of the algorithm by the example of a perceptron identifier, but it also can be used to increase the storage capacity of full connected systems of associative memory.

  5. Community capacity building and sustainability: outcomes of community-based participatory research.

    Science.gov (United States)

    Hacker, Karen; Tendulkar, Shalini A; Rideout, Catlin; Bhuiya, Nazmim; Trinh-Shevrin, Chau; Savage, Clara P; Grullon, Milagro; Strelnick, Hal; Leung, Carolyn; DiGirolamo, Ann

    2012-01-01

    For communities, the value of community-based participatory research (CBPR) is often manifested in the outcomes of increased capacity and sustainable adoption of evidence-based practices for social change. Educational opportunities that promote discourse between community and academic partners can help to advance CBPR and better define these outcomes. This paper describes a community-academic conference to develop shared definitions of community capacity building and sustainability related to CBPR and to identify obstacles and facilitators to both. "Taking It to the Curbside: Engaging Communities to Create Sustainable Change for Health" was planned by five Clinical Translational Science Institutes and four community organizations. After a keynote presentation, breakout groups of community and academic members met to define community capacity building and sustainability, and to identify facilitators and barriers to achieving both. Groups were facilitated by researcher-community partner teams and conversations were recorded and transcribed. Qualitative analysis for thematic content was conducted by a subset of the planning committee. Important findings included learning that (1) the concepts of capacity and sustainability were considered interconnected; (2) partnership was perceived as both a facilitator and an outcome of CBPR; (3) sustainability was linked to "transfer of knowledge" from one generation to another within a community; and (4) capacity and sustainability were enhanced when goals were shared and health outcomes were achieved. Community capacity building and sustainability are key outcomes of CBPR for communities. Co-learning opportunities that engage and mutually educate both community members and academics can be useful strategies for identifying meaningful strategies to achieve these outcomes.

  6. A Simple Density with Distance Based Initial Seed Selection Technique for K Means Algorithm

    Directory of Open Access Journals (Sweden)

    Sajidha Syed Azimuddin

    2017-01-01

    Full Text Available Open issues with respect to K means algorithm are identifying the number of clusters, initial seed concept selection, clustering tendency, handling empty clusters, identifying outliers etc. In this paper we propose a novel and a simple technique considering both density and distance of the concepts in a dataset to identify initial seed concepts for clustering. Many authors have proposed different techniques to identify initial seed concepts; but our method ensures that the initial seed concepts are chosen from different clusters that are to be generated by the clustering solution. The hallmark of our algorithm is that it is a single pass algorithm that does not require any extra parameters to be estimated. Further, our seed concepts are one among the actual concepts and not the mean of representative concepts as is the case in many other algorithms. We have implemented our proposed algorithm and compared the results with the interval based technique of Fouad Khan. We see that our method outperforms the interval based method. We have also compared our method with the original random K means and K Means++ algorithms.

  7. Utility of an Algorithm to Increase the Accuracy of Medication History in an Obstetrical Setting.

    Science.gov (United States)

    Corbel, Aline; Baud, David; Chaouch, Aziz; Beney, Johnny; Csajka, Chantal; Panchaud, Alice

    2016-01-01

    In an obstetrical setting, inaccurate medication histories at hospital admission may result in failure to identify potentially harmful treatments for patients and/or their fetus(es). This prospective study was conducted to assess average concordance rates between (1) a medication list obtained with a one-page structured medication history algorithm developed for the obstetrical setting and (2) the medication list reported in medical records and obtained by open-ended questions based on standard procedures. Both lists were converted into concordance rate using a best possible medication history approach as the reference (information obtained by patients, prescribers and community pharmacists' interviews). The algorithm-based method obtained a higher average concordance rate than the standard method, with respectively 90.2% [CI95% 85.8-94.3] versus 24.6% [CI95%15.3-34.4] concordance rates (phistory in our obstetric population, without using substantial resources. Its implementation is an effective first step to the medication reconciliation process, which has been recognized as a very important component of patients' drug safety.

  8. Exploring anti-community structure in networks with application to incompatibility of traditional Chinese medicine

    Science.gov (United States)

    Zhu, Jiajing; Liu, Yongguo; Zhang, Yun; Liu, Xiaofeng; Xiao, Yonghua; Wang, Shidong; Wu, Xindong

    2017-11-01

    Community structure is one of the most important properties in networks, in which a node shares its most connections with the others in the same community. On the contrary, the anti-community structure means the nodes in the same group have few or no connections with each other. In Traditional Chinese Medicine (TCM), the incompatibility problem of herbs is a challenge to the clinical medication safety. In this paper, we propose a new anti-community detection algorithm, Random non-nEighboring nOde expansioN (REON), to find anti-communities in networks, in which a new evaluation criterion, anti-modularity, is designed to measure the quality of the obtained anti-community structure. In order to establish anti-communities in REON, we expand the node set by non-neighboring node expansion and regard the node set with the highest anti-modularity as an anti-community. Inspired by the phenomenon that the node with higher degree has greater contribution to the anti-modularity, an improved algorithm called REONI is developed by expanding node set by the non-neighboring node with the maximum degree, which greatly enhances the efficiency of REON. Experiments on synthetic and real-world networks demonstrate the superiority of the proposed algorithms over the existing methods. In addition, by applying REONI to the herb network, we find that it can discover incompatible herb combinations.

  9. Robustness of the ATLAS pixel clustering neural network algorithm

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00407780; The ATLAS collaboration

    2016-01-01

    Proton-proton collisions at the energy frontier puts strong constraints on track reconstruction algorithms. In the ATLAS track reconstruction algorithm, an artificial neural network is utilised to identify and split clusters of neighbouring read-out elements in the ATLAS pixel detector created by multiple charged particles. The robustness of the neural network algorithm is presented, probing its sensitivity to uncertainties in the detector conditions. The robustness is studied by evaluating the stability of the algorithm's performance under a range of variations in the inputs to the neural networks. Within reasonable variation magnitudes, the neural networks prove to be robust to most variation types.

  10. A risk profile for identifying community-dwelling elderly with a highrisk of recurrent falling: results of a 3-year prospective study

    NARCIS (Netherlands)

    Pluym, S.M.F.; Smit, J.H.; Tromp, A.M.; Stel, V.S.; Deeg, D.J.H.; Bouter, L.M.; Lips, P.T.A.M.

    2007-01-01

    Introduction: The aim of the prospective study reported here was to develop a risk profile that can be used to identify community-dwelling elderly at a high risk of recurrent falling. Materials and methods: The study was designed as a 3-year prospective cohort study. A total of 1365

  11. Tag SNP selection via a genetic algorithm.

    Science.gov (United States)

    Mahdevar, Ghasem; Zahiri, Javad; Sadeghi, Mehdi; Nowzari-Dalini, Abbas; Ahrabian, Hayedeh

    2010-10-01

    Single Nucleotide Polymorphisms (SNPs) provide valuable information on human evolutionary history and may lead us to identify genetic variants responsible for human complex diseases. Unfortunately, molecular haplotyping methods are costly, laborious, and time consuming; therefore, algorithms for constructing full haplotype patterns from small available data through computational methods, Tag SNP selection problem, are convenient and attractive. This problem is proved to be an NP-hard problem, so heuristic methods may be useful. In this paper we present a heuristic method based on genetic algorithm to find reasonable solution within acceptable time. The algorithm was tested on a variety of simulated and experimental data. In comparison with the exact algorithm, based on brute force approach, results show that our method can obtain optimal solutions in almost all cases and runs much faster than exact algorithm when the number of SNP sites is large. Our software is available upon request to the corresponding author.

  12. An Efficient, Noniterative Method of Identifying the Cost-Effectiveness Frontier.

    Science.gov (United States)

    Suen, Sze-chuan; Goldhaber-Fiebert, Jeremy D

    2016-01-01

    Cost-effectiveness analysis aims to identify treatments and policies that maximize benefits subject to resource constraints. However, the conventional process of identifying the efficient frontier (i.e., the set of potentially cost-effective options) can be algorithmically inefficient, especially when considering a policy problem with many alternative options or when performing an extensive suite of sensitivity analyses for which the efficient frontier must be found for each. Here, we describe an alternative one-pass algorithm that is conceptually simple, easier to implement, and potentially faster for situations that challenge the conventional approach. Our algorithm accomplishes this by exploiting the relationship between the net monetary benefit and the cost-effectiveness plane. To facilitate further evaluation and use of this approach, we also provide scripts in R and Matlab that implement our method and can be used to identify efficient frontiers for any decision problem. © The Author(s) 2015.

  13. Enhancements of LEACH Algorithm for Wireless Networks: A Review

    Directory of Open Access Journals (Sweden)

    M. Madheswaran

    2013-12-01

    Full Text Available Low Energy Adaptive Clustering Hierarchy (LEACH protocol is the first hierarchical cluster based routing protocol successfully used in the Wireless Sensor Networks (WSN. In this paper, various enhancements used in the original LEACH protocol are examined. The basic operations, advantages and limitations of the modified LEACH algorithms are compared to identify the research issues to be solved and to give the suggestions for the future proposed routing algorithms of wireless networks based on LEACH routing algorithm.

  14. A simple fall detection algorithm for Powered Two Wheelers

    OpenAIRE

    BOUBEZOUL, Abderrahmane; ESPIE, Stéphane; LARNAUDIE, Bruno; BOUAZIZ, Samir

    2013-01-01

    The aim of this study is to evaluate a low-complexity fall detection algorithm, that use both acceleration and angular velocity signals to trigger an alert-system or to inflate an airbag jacket. The proposed fall detection algorithm is a threshold-based algorithm, using data from 3-accelerometers and 3-gyroscopes sensors mounted on the motorcycle. During the first step, the commonly fall accident configurations were selected and analyzed in order to identify the main causation factors. On the...

  15. A space-efficient algorithm for local similarities.

    Science.gov (United States)

    Huang, X Q; Hardison, R C; Miller, W

    1990-10-01

    Existing dynamic-programming algorithms for identifying similar regions of two sequences require time and space proportional to the product of the sequence lengths. Often this space requirement is more limiting than the time requirement. We describe a dynamic-programming local-similarity algorithm that needs only space proportional to the sum of the sequence lengths. The method can also find repeats within a single long sequence. To illustrate the algorithm's potential, we discuss comparison of a 73,360 nucleotide sequence containing the human beta-like globin gene cluster and a corresponding 44,594 nucleotide sequence for rabbit, a problem well beyond the capabilities of other dynamic-programming software.

  16. Radioactivity nuclide identification based on BP and LM algorithm neural network

    International Nuclear Information System (INIS)

    Wang Jihong; Sun Jian; Wang Lianghou

    2012-01-01

    The paper provides the method which can identify radioactive nuclide based on the BP and LM algorithm neural network. Then, this paper compares the above-mentioned method with FR algorithm. Through the result of the Matlab simulation, the method of radioactivity nuclide identification based on the BP and LM algorithm neural network is superior to the FR algorithm. With the better effect and the higher accuracy, it will be the best choice. (authors)

  17. Using a community-driven approach to identify local forest and climate change priorities in Teslin, Yukon

    Directory of Open Access Journals (Sweden)

    Joleen Timko

    2015-12-01

    Full Text Available The likelihood of addressing the complex environmental, economic, and social/cultural issues associated with local climate change impacts is enhanced when collaborative partnerships with local people are established. Using a community-centered approach in the Teslin region of Canada’s Yukon Territory, we utilized our research skills to respond to local needs for information by facilitating both an internal community process to clarify traditional and local knowledge, values, and perceptions on locally identified priorities, while gathering external information to enable local people to make sound decisions. Specifically, we sought to clarify local perceptions surrounding climate change impacts on fire risk and wildlife habitat, and the potential adaptation strategies appropriate and feasible within the Teslin Tlingit Traditional Territory. This paper provides a characterization of the study region and our project team; provides background on the interview and data collection process; presents our key results; and discusses the importance of our findings and charts a way forward for our continued work with the people in the Teslin region. This approach presents an excellent opportunity to help people holistically connect a range of local values, including fire risk mitigation, habitat enhancement, economic development, and enhanced social health.

  18. Structure-Based Algorithms for Microvessel Classification

    KAUST Repository

    Smith, Amy F.; Secomb, Timothy W.; Pries, Axel R.; Smith, Nicolas P.; Shipley, Rebecca J.

    2015-01-01

    algorithm, developed for networks with one arteriolar and one venular tree, performs well in identifying arterioles and venules and is robust to parameter changes, but incorrectly labels a significant number of capillaries as arterioles or venules

  19. Parameter identification of PEMFC model based on hybrid adaptive differential evolution algorithm

    International Nuclear Information System (INIS)

    Sun, Zhe; Wang, Ning; Bi, Yunrui; Srinivasan, Dipti

    2015-01-01

    In this paper, a HADE (hybrid adaptive differential evolution) algorithm is proposed for the identification problem of PEMFC (proton exchange membrane fuel cell). Inspired by biological genetic strategy, a novel adaptive scaling factor and a dynamic crossover probability are presented to improve the adaptive and dynamic performance of differential evolution algorithm. Moreover, two kinds of neighborhood search operations based on the bee colony foraging mechanism are introduced for enhancing local search efficiency. Through testing the benchmark functions, the proposed algorithm exhibits better performance in convergent accuracy and speed. Finally, the HADE algorithm is applied to identify the nonlinear parameters of PEMFC stack model. Through experimental comparison with other identified methods, the PEMFC model based on the HADE algorithm shows better performance. - Highlights: • We propose a hybrid adaptive differential evolution algorithm (HADE). • The search efficiency is enhanced in low and high dimension search space. • The effectiveness is confirmed by testing benchmark functions. • The identification of the PEMFC model is conducted by adopting HADE.

  20. Community Health Workers in the United States: Challenges in Identifying, Surveying, and Supporting the Workforce.

    Science.gov (United States)

    Sabo, Samantha; Allen, Caitlin G; Sutkowi, Katherine; Wennerstrom, Ashley

    2017-12-01

    Community health workers (CHWs) are members of a growing profession in the United States. Studying this dynamic labor force is challenging, in part because its members have more than 100 different job titles. The demand for timely, accurate information about CHWs is increasing as the profession gains recognition for its ability to improve health outcomes and reduce costs. Although numerous surveys of CHWs have been conducted, the field lacks well-delineated methods for gaining access to this hard-to-identify workforce. We outline methods for surveying CHWs and promising approaches to engage the workforce and other stakeholders in conducting local, state, and national studies. We also highlight successful strategies to overcome challenges in CHW surveys and future directions for surveying the field.

  1. Involving the Community

    International Development Research Centre (IDRC) Digital Library (Canada)

    Step 3: Identifying the different community groups and other stakeholders concerned .... How can two-way communication enhance community participation in ...... for maintenance and the rights of specific community groups to drinkable water.

  2. Automatic identification of otological drilling faults: an intelligent recognition algorithm.

    Science.gov (United States)

    Cao, Tianyang; Li, Xisheng; Gao, Zhiqiang; Feng, Guodong; Shen, Peng

    2010-06-01

    This article presents an intelligent recognition algorithm that can recognize milling states of the otological drill by fusing multi-sensor information. An otological drill was modified by the addition of sensors. The algorithm was designed according to features of the milling process and is composed of a characteristic curve, an adaptive filter and a rule base. The characteristic curve can weaken the impact of the unstable normal milling process and reserve the features of drilling faults. The adaptive filter is capable of suppressing interference in the characteristic curve by fusing multi-sensor information. The rule base can identify drilling faults through the filtering result data. The experiments were repeated on fresh porcine scapulas, including normal milling and two drilling faults. The algorithm has high rates of identification. This study shows that the intelligent recognition algorithm can identify drilling faults under interference conditions. (c) 2010 John Wiley & Sons, Ltd.

  3. Verification test for on-line diagnosis algorithm based on noise analysis

    International Nuclear Information System (INIS)

    Tamaoki, T.; Naito, N.; Tsunoda, T.; Sato, M.; Kameda, A.

    1980-01-01

    An on-line diagnosis algorithm was developed and its verification test was performed using a minicomputer. This algorithm identifies the plant state by analyzing various system noise patterns, such as power spectral densities, coherence functions etc., in three procedure steps. Each obtained noise pattern is examined by using the distances from its reference patterns prepared for various plant states. Then, the plant state is identified by synthesizing each result with an evaluation weight. This weight is determined automatically from the reference noise patterns prior to on-line diagnosis. The test was performed with 50 MW (th) Steam Generator noise data recorded under various controller parameter values. The algorithm performance was evaluated based on a newly devised index. The results obtained with one kind of weight showed the algorithm efficiency under the proper selection of noise patterns. Results for another kind of weight showed the robustness of the algorithm to this selection. (orig.)

  4. PDES, Fips Standard Data Encryption Algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Nessett, D N [Lawrence Livermore National Laboratory (United States)

    1991-03-26

    Description of program or function: PDES performs the National Bureau of Standards FIPS Pub. 46 data encryption/decryption algorithm used for the cryptographic protection of computer data. The DES algorithm is designed to encipher and decipher blocks of data consisting of 64 bits under control of a 64-bit key. The key is generated in such a way that each of the 56 bits used directly by the algorithm are random and the remaining 8 error-detecting bits are set to make the parity of each 8-bit byte of the key odd, i. e. there is an odd number of '1' bits in each 8-bit byte. Each member of a group of authorized users of encrypted computer data must have the key that was used to encipher the data in order to use it. Data can be recovered from cipher only by using exactly the same key used to encipher it, but with the schedule of addressing the key bits altered so that the deciphering process is the reverse of the enciphering process. A block of data to be enciphered is subjected to an initial permutation, then to a complex key-dependent computation, and finally to a permutation which is the inverse of the initial permutation. Two PDES routines are included; both perform the same calculation. One, identified as FDES.MAR, is designed to achieve speed in execution, while the other identified as PDES.MAR, presents a clearer view of how the algorithm is executed

  5. PDES, Fips Standard Data Encryption Algorithm

    International Nuclear Information System (INIS)

    Nessett, D.N.

    1991-01-01

    Description of program or function: PDES performs the National Bureau of Standards FIPS Pub. 46 data encryption/decryption algorithm used for the cryptographic protection of computer data. The DES algorithm is designed to encipher and decipher blocks of data consisting of 64 bits under control of a 64-bit key. The key is generated in such a way that each of the 56 bits used directly by the algorithm are random and the remaining 8 error-detecting bits are set to make the parity of each 8-bit byte of the key odd, i. e. there is an odd number of '1' bits in each 8-bit byte. Each member of a group of authorized users of encrypted computer data must have the key that was used to encipher the data in order to use it. Data can be recovered from cipher only by using exactly the same key used to encipher it, but with the schedule of addressing the key bits altered so that the deciphering process is the reverse of the enciphering process. A block of data to be enciphered is subjected to an initial permutation, then to a complex key-dependent computation, and finally to a permutation which is the inverse of the initial permutation. Two PDES routines are included; both perform the same calculation. One, identified as FDES.MAR, is designed to achieve speed in execution, while the other identified as PDES.MAR, presents a clearer view of how the algorithm is executed

  6. Optimized hyperspectral band selection using hybrid genetic algorithm and gravitational search algorithm

    Science.gov (United States)

    Zhang, Aizhu; Sun, Genyun; Wang, Zhenjie

    2015-12-01

    The serious information redundancy in hyperspectral images (HIs) cannot contribute to the data analysis accuracy, instead it require expensive computational resources. Consequently, to identify the most useful and valuable information from the HIs, thereby improve the accuracy of data analysis, this paper proposed a novel hyperspectral band selection method using the hybrid genetic algorithm and gravitational search algorithm (GA-GSA). In the proposed method, the GA-GSA is mapped to the binary space at first. Then, the accuracy of the support vector machine (SVM) classifier and the number of selected spectral bands are utilized to measure the discriminative capability of the band subset. Finally, the band subset with the smallest number of spectral bands as well as covers the most useful and valuable information is obtained. To verify the effectiveness of the proposed method, studies conducted on an AVIRIS image against two recently proposed state-of-the-art GSA variants are presented. The experimental results revealed the superiority of the proposed method and indicated that the method can indeed considerably reduce data storage costs and efficiently identify the band subset with stable and high classification precision.

  7. A physarum-inspired prize-collecting steiner tree approach to identify subnetworks for drug repositioning.

    Science.gov (United States)

    Sun, Yahui; Hameed, Pathima Nusrath; Verspoor, Karin; Halgamuge, Saman

    2016-12-05

    Drug repositioning can reduce the time, costs and risks of drug development by identifying new therapeutic effects for known drugs. It is challenging to reposition drugs as pharmacological data is large and complex. Subnetwork identification has already been used to simplify the visualization and interpretation of biological data, but it has not been applied to drug repositioning so far. In this paper, we fill this gap by proposing a new Physarum-inspired Prize-Collecting Steiner Tree algorithm to identify subnetworks for drug repositioning. Drug Similarity Networks (DSN) are generated using the chemical, therapeutic, protein, and phenotype features of drugs. In DSNs, vertex prizes and edge costs represent the similarities and dissimilarities between drugs respectively, and terminals represent drugs in the cardiovascular class, as defined in the Anatomical Therapeutic Chemical classification system. A new Physarum-inspired Prize-Collecting Steiner Tree algorithm is proposed in this paper to identify subnetworks. We apply both the proposed algorithm and the widely-used GW algorithm to identify subnetworks in our 18 generated DSNs. In these DSNs, our proposed algorithm identifies subnetworks with an average Rand Index of 81.1%, while the GW algorithm can only identify subnetworks with an average Rand Index of 64.1%. We select 9 subnetworks with high Rand Index to find drug repositioning opportunities. 10 frequently occurring drugs in these subnetworks are identified as candidates to be repositioned for cardiovascular diseases. We find evidence to support previous discoveries that nitroglycerin, theophylline and acarbose may be able to be repositioned for cardiovascular diseases. Moreover, we identify seven previously unknown drug candidates that also may interact with the biological cardiovascular system. These discoveries show our proposed Prize-Collecting Steiner Tree approach as a promising strategy for drug repositioning.

  8. Evaluation of in silico algorithms for use with ACMG/AMP clinical variant interpretation guidelines.

    Science.gov (United States)

    Ghosh, Rajarshi; Oak, Ninad; Plon, Sharon E

    2017-11-28

    The American College of Medical Genetics and American College of Pathologists (ACMG/AMP) variant classification guidelines for clinical reporting are widely used in diagnostic laboratories for variant interpretation. The ACMG/AMP guidelines recommend complete concordance of predictions among all in silico algorithms used without specifying the number or types of algorithms. The subjective nature of this recommendation contributes to discordance of variant classification among clinical laboratories and prevents definitive classification of variants. Using 14,819 benign or pathogenic missense variants from the ClinVar database, we compared performance of 25 algorithms across datasets differing in distinct biological and technical variables. There was wide variability in concordance among different combinations of algorithms with particularly low concordance for benign variants. We also identify a previously unreported source of error in variant interpretation (false concordance) where concordant in silico predictions are opposite to the evidence provided by other sources. We identified recently developed algorithms with high predictive power and robust to variables such as disease mechanism, gene constraint, and mode of inheritance, although poorer performing algorithms are more frequently used based on review of the clinical genetics literature (2011-2017). Our analyses identify algorithms with high performance characteristics independent of underlying disease mechanisms. We describe combinations of algorithms with increased concordance that should improve in silico algorithm usage during assessment of clinically relevant variants using the ACMG/AMP guidelines.

  9. A robust human face detection algorithm

    Science.gov (United States)

    Raviteja, Thaluru; Karanam, Srikrishna; Yeduguru, Dinesh Reddy V.

    2012-01-01

    Human face detection plays a vital role in many applications like video surveillance, managing a face image database, human computer interface among others. This paper proposes a robust algorithm for face detection in still color images that works well even in a crowded environment. The algorithm uses conjunction of skin color histogram, morphological processing and geometrical analysis for detecting human faces. To reinforce the accuracy of face detection, we further identify mouth and eye regions to establish the presence/absence of face in a particular region of interest.

  10. [A new peak detection algorithm of Raman spectra].

    Science.gov (United States)

    Jiang, Cheng-Zhi; Sun, Qiang; Liu, Ying; Liang, Jing-Qiu; An, Yan; Liu, Bing

    2014-01-01

    The authors proposed a new Raman peak recognition method named bi-scale correlation algorithm. The algorithm uses the combination of the correlation coefficient and the local signal-to-noise ratio under two scales to achieve Raman peak identification. We compared the performance of the proposed algorithm with that of the traditional continuous wavelet transform method through MATLAB, and then tested the algorithm with real Raman spectra. The results show that the average time for identifying a Raman spectrum is 0.51 s with the algorithm, while it is 0.71 s with the continuous wavelet transform. When the signal-to-noise ratio of Raman peak is greater than or equal to 6 (modern Raman spectrometers feature an excellent signal-to-noise ratio), the recognition accuracy with the algorithm is higher than 99%, while it is less than 84% with the continuous wavelet transform method. The mean and the standard deviations of the peak position identification error of the algorithm are both less than that of the continuous wavelet transform method. Simulation analysis and experimental verification prove that the new algorithm possesses the following advantages: no needs of human intervention, no needs of de-noising and background removal operation, higher recognition speed and higher recognition accuracy. The proposed algorithm is operable in Raman peak identification.

  11. An objective procedure for evaluation of adaptive antifeedback algorithms in hearing aids.

    Science.gov (United States)

    Freed, Daniel J; Soli, Sigfrid D

    2006-08-01

    This study evaluated the performance of nine adaptive antifeedback algorithms. There were two goals: first, to identify objective procedures that are useful for evaluating these algorithms, and second, to identify strengths and weaknesses of existing algorithms. The algorithms were evaluated in behind-the-ear implementations on the Knowles Electronics Manikin for Acoustic Research (KEMAR). Different acoustic conditions were created by placing a telephone handset or a hat on KEMAR. Electroacoustic techniques were devised to measure the following performance aspects of each algorithm: (1) additional gain made available before oscillation, (2) gain lost in specific frequency regions, (3) reduction of suboscillatory peaks in the frequency response, (4) speed of adaptation to changing acoustic conditions, and (5) robustness in the presence of tonal input signals. For each measurement, performance varied widely across algorithms. No single algorithm was clearly superior or inferior to the others. Generally, the feedback cancellation algorithms were less likely to sacrifice gain in specific frequency regions and better at reducing suboscillatory peaks, whereas the algorithms that used noncancellation techniques were more tolerant of tonal input signals. For those algorithms equipped with special operational modes intended for music listening, the music mode improved the response to tonal inputs but sometimes sacrificed other performance aspects. Algorithms that required an acoustic measurement for initialization purposes tended to perform poorly in acoustic conditions dissimilar to the condition in which initialization was performed. The objective methods devised for this study appear useful for evaluating the performance of adaptive antifeedback algorithms. Currently available algorithms demonstrate a wide range of performance, and further research is required to develop new algorithms that combine the best features of existing algorithms.

  12. Search algorithms as a framework for the optimization of drug combinations.

    Directory of Open Access Journals (Sweden)

    Diego Calzolari

    2008-12-01

    Full Text Available Combination therapies are often needed for effective clinical outcomes in the management of complex diseases, but presently they are generally based on empirical clinical experience. Here we suggest a novel application of search algorithms -- originally developed for digital communication -- modified to optimize combinations of therapeutic interventions. In biological experiments measuring the restoration of the decline with age in heart function and exercise capacity in Drosophila melanogaster, we found that search algorithms correctly identified optimal combinations of four drugs using only one-third of the tests performed in a fully factorial search. In experiments identifying combinations of three doses of up to six drugs for selective killing of human cancer cells, search algorithms resulted in a highly significant enrichment of selective combinations compared with random searches. In simulations using a network model of cell death, we found that the search algorithms identified the optimal combinations of 6-9 interventions in 80-90% of tests, compared with 15-30% for an equivalent random search. These findings suggest that modified search algorithms from information theory have the potential to enhance the discovery of novel therapeutic drug combinations. This report also helps to frame a biomedical problem that will benefit from an interdisciplinary effort and suggests a general strategy for its solution.

  13. Algorithms

    Indian Academy of Sciences (India)

    polynomial) division have been found in Vedic Mathematics which are dated much before Euclid's algorithm. A programming language Is used to describe an algorithm for execution on a computer. An algorithm expressed using a programming.

  14. Identifying and Assessing Community-Based Social Behavior of Adolescents and Young Adults with EBD.

    Science.gov (United States)

    Bullis, Michael; And Others

    1994-01-01

    A battery of three measures for assessing the community-based social behavior of adolescents and young adults with emotional and behavioral disorders is described. The measures, in male and female forms, are "Test of Community-Based Social Skill Knowledge,""Scale of Community-Based Social Skill Performance," and "Behaviors That Are Undesirable for…

  15. An algorithm for modularization of MAPK and calcium signaling pathways: comparative analysis among different species.

    Science.gov (United States)

    Nayak, Losiana; De, Rajat K

    2007-12-01

    Signaling pathways are large complex biochemical networks. It is difficult to analyze the underlying mechanism of such networks as a whole. In the present article, we have proposed an algorithm for modularization of signal transduction pathways. Unlike studying a signaling pathway as a whole, this enables one to study the individual modules (less complex smaller units) easily and hence to study the entire pathway better. A comparative study of modules belonging to different species (for the same signaling pathway) has been made, which gives an overall idea about development of the signaling pathways over the taken set of species of calcium and MAPK signaling pathways. The superior performance, in terms of biological significance, of the proposed algorithm over an existing community finding algorithm of Newman [Newman MEJ. Modularity and community structure in networks. Proc Natl Acad Sci USA 2006;103(23):8577-82] has been demonstrated using the aforesaid pathways of H. sapiens.

  16. Data clustering theory, algorithms, and applications

    CERN Document Server

    Gan, Guojun; Wu, Jianhong

    2007-01-01

    Cluster analysis is an unsupervised process that divides a set of objects into homogeneous groups. This book starts with basic information on cluster analysis, including the classification of data and the corresponding similarity measures, followed by the presentation of over 50 clustering algorithms in groups according to some specific baseline methodologies such as hierarchical, center-based, and search-based methods. As a result, readers and users can easily identify an appropriate algorithm for their applications and compare novel ideas with existing results. The book also provides examples of clustering applications to illustrate the advantages and shortcomings of different clustering architectures and algorithms. Application areas include pattern recognition, artificial intelligence, information technology, image processing, biology, psychology, and marketing. Readers also learn how to perform cluster analysis with the C/C++ and MATLAB® programming languages.

  17. XTALOPT: An open-source evolutionary algorithm for crystal structure prediction

    Science.gov (United States)

    Lonie, David C.; Zurek, Eva

    2011-02-01

    The implementation and testing of XTALOPT, an evolutionary algorithm for crystal structure prediction, is outlined. We present our new periodic displacement (ripple) operator which is ideally suited to extended systems. It is demonstrated that hybrid operators, which combine two pure operators, reduce the number of duplicate structures in the search. This allows for better exploration of the potential energy surface of the system in question, while simultaneously zooming in on the most promising regions. A continuous workflow, which makes better use of computational resources as compared to traditional generation based algorithms, is employed. Various parameters in XTALOPT are optimized using a novel benchmarking scheme. XTALOPT is available under the GNU Public License, has been interfaced with various codes commonly used to study extended systems, and has an easy to use, intuitive graphical interface. Program summaryProgram title:XTALOPT Catalogue identifier: AEGX_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEGX_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GPL v2.1 or later [1] No. of lines in distributed program, including test data, etc.: 36 849 No. of bytes in distributed program, including test data, etc.: 1 149 399 Distribution format: tar.gz Programming language: C++ Computer: PCs, workstations, or clusters Operating system: Linux Classification: 7.7 External routines: QT [2], OpenBabel [3], AVOGADRO [4], SPGLIB [8] and one of: VASP [5], PWSCF [6], GULP [7]. Nature of problem: Predicting the crystal structure of a system from its stoichiometry alone remains a grand challenge in computational materials science, chemistry, and physics. Solution method: Evolutionary algorithms are stochastic search techniques which use concepts from biological evolution in order to locate the global minimum on their potential energy surface. Our evolutionary algorithm, XTALOPT, is freely

  18. In-depth analysis of protein inference algorithms using multiple search engines and well-defined metrics.

    Science.gov (United States)

    Audain, Enrique; Uszkoreit, Julian; Sachsenberg, Timo; Pfeuffer, Julianus; Liang, Xiao; Hermjakob, Henning; Sanchez, Aniel; Eisenacher, Martin; Reinert, Knut; Tabb, David L; Kohlbacher, Oliver; Perez-Riverol, Yasset

    2017-01-06

    In mass spectrometry-based shotgun proteomics, protein identifications are usually the desired result. However, most of the analytical methods are based on the identification of reliable peptides and not the direct identification of intact proteins. Thus, assembling peptides identified from tandem mass spectra into a list of proteins, referred to as protein inference, is a critical step in proteomics research. Currently, different protein inference algorithms and tools are available for the proteomics community. Here, we evaluated five software tools for protein inference (PIA, ProteinProphet, Fido, ProteinLP, MSBayesPro) using three popular database search engines: Mascot, X!Tandem, and MS-GF+. All the algorithms were evaluated using a highly customizable KNIME workflow using four different public datasets with varying complexities (different sample preparation, species and analytical instruments). We defined a set of quality control metrics to evaluate the performance of each combination of search engines, protein inference algorithm, and parameters on each dataset. We show that the results for complex samples vary not only regarding the actual numbers of reported protein groups but also concerning the actual composition of groups. Furthermore, the robustness of reported proteins when using databases of differing complexities is strongly dependant on the applied inference algorithm. Finally, merging the identifications of multiple search engines does not necessarily increase the number of reported proteins, but does increase the number of peptides per protein and thus can generally be recommended. Protein inference is one of the major challenges in MS-based proteomics nowadays. Currently, there are a vast number of protein inference algorithms and implementations available for the proteomics community. Protein assembly impacts in the final results of the research, the quantitation values and the final claims in the research manuscript. Even though protein

  19. Multi-scale analysis of the European airspace using network community detection.

    Directory of Open Access Journals (Sweden)

    Gérald Gurtner

    Full Text Available We show that the European airspace can be represented as a multi-scale traffic network whose nodes are airports, sectors, or navigation points and links are defined and weighted according to the traffic of flights between the nodes. By using a unique database of the air traffic in the European airspace, we investigate the architecture of these networks with a special emphasis on their community structure. We propose that unsupervised network community detection algorithms can be used to monitor the current use of the airspace and improve it by guiding the design of new ones. Specifically, we compare the performance of several community detection algorithms, both with fixed and variable resolution, and also by using a null model which takes into account the spatial distance between nodes, and we discuss their ability to find communities that could be used to define new control units of the airspace.

  20. Traveling salesman problems with PageRank Distance on complex networks reveal community structure

    Science.gov (United States)

    Jiang, Zhongzhou; Liu, Jing; Wang, Shuai

    2016-12-01

    In this paper, we propose a new algorithm for community detection problems (CDPs) based on traveling salesman problems (TSPs), labeled as TSP-CDA. Since TSPs need to find a tour with minimum cost, cities close to each other are usually clustered in the tour. This inspired us to model CDPs as TSPs by taking each vertex as a city. Then, in the final tour, the vertices in the same community tend to cluster together, and the community structure can be obtained by cutting the tour into a couple of paths. There are two challenges. The first is to define a suitable distance between each pair of vertices which can reflect the probability that they belong to the same community. The second is to design a suitable strategy to cut the final tour into paths which can form communities. In TSP-CDA, we deal with these two challenges by defining a PageRank Distance and an automatic threshold-based cutting strategy. The PageRank Distance is designed with the intrinsic properties of CDPs in mind, and can be calculated efficiently. In the experiments, benchmark networks with 1000-10,000 nodes and varying structures are used to test the performance of TSP-CDA. A comparison is also made between TSP-CDA and two well-established community detection algorithms. The results show that TSP-CDA can find accurate community structure efficiently and outperforms the two existing algorithms.

  1. Improved Ant Colony Clustering Algorithm and Its Performance Study

    Science.gov (United States)

    Gao, Wei

    2016-01-01

    Clustering analysis is used in many disciplines and applications; it is an important tool that descriptively identifies homogeneous groups of objects based on attribute values. The ant colony clustering algorithm is a swarm-intelligent method used for clustering problems that is inspired by the behavior of ant colonies that cluster their corpses and sort their larvae. A new abstraction ant colony clustering algorithm using a data combination mechanism is proposed to improve the computational efficiency and accuracy of the ant colony clustering algorithm. The abstraction ant colony clustering algorithm is used to cluster benchmark problems, and its performance is compared with the ant colony clustering algorithm and other methods used in existing literature. Based on similar computational difficulties and complexities, the results show that the abstraction ant colony clustering algorithm produces results that are not only more accurate but also more efficiently determined than the ant colony clustering algorithm and the other methods. Thus, the abstraction ant colony clustering algorithm can be used for efficient multivariate data clustering. PMID:26839533

  2. Ethics in Community-University-Artist Partnered Research: Tensions, Contradictions and Gaps Identified in an 'Arts for Social Change' Project.

    Science.gov (United States)

    Yassi, Annalee; Spiegel, Jennifer Beth; Lockhart, Karen; Fels, Lynn; Boydell, Katherine; Marcuse, Judith

    Academics from diverse disciplines are recognizing not only the procedural ethical issues involved in research, but also the complexity of everyday "micro" ethical issues that arise. While ethical guidelines are being developed for research in aboriginal populations and low-and-middle-income countries, multi-partnered research initiatives examining arts-based interventions to promote social change pose a unique set of ethical dilemmas not yet fully explored. Our research team, comprising health, education, and social scientists, critical theorists, artists and community-activists launched a five-year research partnership on arts-for-social change. Funded by the Social Science and Humanities Research Council in Canada and based in six universities, including over 40 community-based collaborators, and informed by five main field projects (circus with street youth, theatre by people with disabilities, dance for people with Parkinson's disease, participatory theatre with refugees and artsinfused dialogue), we set out to synthesize existing knowledge and lessons we learned. We summarized these learnings into 12 key points for reflection, grouped into three categories: community-university partnership concerns ( n  = 3), dilemmas related to the arts ( n  = 5), and team issues ( n  = 4). In addition to addressing previous concerns outlined in the literature (e.g., related to consent, anonymity, dangerous emotional terrain, etc.), we identified power dynamics (visible and hidden) hindering meaningful participation of community partners and university-based teams that need to be addressed within a reflective critical framework of ethical practice. We present how our team has been addressing these issues, as examples of how such concerns could be approached in community-university partnerships in arts for social change.

  3. Social Graph Community Differentiated by Node Features with Partly Missing Information

    Directory of Open Access Journals (Sweden)

    V. O. Chesnokov

    2015-01-01

    Full Text Available This paper proposes a new algorithm for community differentiation in social graphs, which uses information both on the graph structure and on the vertices. We consider user's ego-network i.e. his friends, with no himself, where each vertex has a set of features such as details on a workplace, institution, etc. The task is to determine missing or unspecified features of the vertices, based on their neighbors' features, and use these features to differentiate the communities in the social graph. Two vertices are believed to belong to the same community if they have a common feature. A hypothesis has been put forward that if most neighbors of a vertex have a common feature, there is a good probability that the vertex has this feature as well. The proposed algorithm is iterative and updates features of vertices, based on its neighbors, according to the hypothesis. Share of neighbors that form a majority is specified by the algorithm parameter. Complexity of single iteration depends linearly on the number of edges in the graph.To assess the quality of clustering three normalized metrics were used, namely: expected density, silhouette index, and Hubert's Gamma Statistic. The paper describes a method for test sampling of 2.000 graphs of the user's social network \\VKontakte". The API requests addressed \\VKontakte" and parsing HTML-pages of user's profiles and search results provided crawling. Information on user's group membership, secondary and higher education, and workplace was used as features. To store data the PostgreSQL DBMS was used, and the gexf format was used for data processing. For the test sample, metrics for several values of algorithm parameter were estimated: the value of index silhouettes was low (0.14-0.20, but within the normal range; the value of expected density was high, i.e. 1.17-1.52; the value of Hubert's gamma statistic was 0.94-0.95 that is close to the maximum. The number of vertices with no features was calculated before

  4. STAR Algorithm Integration Team - Facilitating operational algorithm development

    Science.gov (United States)

    Mikles, V. J.

    2015-12-01

    The NOAA/NESDIS Center for Satellite Research and Applications (STAR) provides technical support of the Joint Polar Satellite System (JPSS) algorithm development and integration tasks. Utilizing data from the S-NPP satellite, JPSS generates over thirty Environmental Data Records (EDRs) and Intermediate Products (IPs) spanning atmospheric, ocean, cryosphere, and land weather disciplines. The Algorithm Integration Team (AIT) brings technical expertise and support to product algorithms, specifically in testing and validating science algorithms in a pre-operational environment. The AIT verifies that new and updated algorithms function in the development environment, enforces established software development standards, and ensures that delivered packages are functional and complete. AIT facilitates the development of new JPSS-1 algorithms by implementing a review approach based on the Enterprise Product Lifecycle (EPL) process. Building on relationships established during the S-NPP algorithm development process and coordinating directly with science algorithm developers, the AIT has implemented structured reviews with self-contained document suites. The process has supported algorithm improvements for products such as ozone, active fire, vegetation index, and temperature and moisture profiles.

  5. Environmental drivers of viral community composition in Antarctic soils identified by viromics.

    Science.gov (United States)

    Adriaenssens, Evelien M; Kramer, Rolf; Van Goethem, Marc W; Makhalanyane, Thulani P; Hogg, Ian; Cowan, Don A

    2017-07-19

    The Antarctic continent is considered the coldest and driest place on earth with simple ecosystems, devoid of higher plants. Soils in the ice-free regions of Antarctica are known to harbor a wide range of microorganisms from primary producers to grazers, yet their ecology and particularly the role of viruses is poorly understood. In this study, we examined the virus community structures of 14 soil samples from the Mackay Glacier region. Viral communities were extracted from soil and the dsDNA was extracted, amplified using single-primer amplification, and sequenced using the Ion Torrent Proton platform. Metadata on soil physico-chemistry was collected from all sites. Both read and contig datasets were analyzed with reference-independent and reference-dependent methods to assess viral community structures and the influence of environmental parameters on their distribution. We observed a high heterogeneity in virus signatures, independent of geographical proximity. Tailed bacteriophages were dominant in all samples, but the incidences of the affiliated families Siphoviridae and Myoviridae were inversely correlated, suggesting direct competition for hosts. Viruses of the families Phycodnaviridae and Mimiviridae were present at significant levels in high-diversity soil samples and were found to co-occur, implying little competition between them. Combinations of soil factors, including pH, calcium content, and site altitude, were found to be the main drivers of viral community structure. The pattern of viral community structure with higher levels of diversity at lower altitude and pH, and co-occurring viral families, suggests that these cold desert soil viruses interact with each other, the host, and the environment in an intricate manner, playing a potentially crucial role in maintaining host diversity and functioning of the microbial ecosystem in the extreme environments of Antarctic soil.

  6. Identifying frauds and anomalies in Medicare-B dataset.

    Science.gov (United States)

    Jiwon Seo; Mendelevitch, Ofer

    2017-07-01

    Healthcare industry is growing at a rapid rate to reach a market value of $7 trillion dollars world wide. At the same time, fraud in healthcare is becoming a serious problem, amounting to 5% of the total healthcare spending, or $100 billion dollars each year in US. Manually detecting healthcare fraud requires much effort. Recently, machine learning and data mining techniques are applied to automatically detect healthcare frauds. This paper proposes a novel PageRank-based algorithm to detect healthcare frauds and anomalies. We apply the algorithm to Medicare-B dataset, a real-life data with 10 million healthcare insurance claims. The algorithm successfully identifies tens of previously unreported anomalies.

  7. A spectral algorithm for the seriation problem

    Energy Technology Data Exchange (ETDEWEB)

    Atkins, J.E. [Michigan Univ., Ann Arbor, MI (United States). Dept. of Mathematics; Boman, E.G. [Stanford Univ., CA (United States). Dept. of Computer Science; Hendrickson, B. [Sandia National Labs., Albuquerque, NM (United States)

    1994-11-01

    Given a set of objects and a correlation function f reflecting the desire for two items to be near each other, find all sequences {pi} of the items so that correlation preferences are preserved; that is if {pi}(i) < {pi}(j) < {pi}(k) then f(i,j) {ge} f(i,k) and f(j,k) {ge} f(i,k). This seriation problem has numerous applications, for instance, solving it yields a solution to the consecutive ones problem. We present a spectral algorithm for this problem that has a number of interesting features. Whereas most previous applications of spectral techniques provided bounds or heuristics, our result is an algorithm for a nontrivial combinatorial problem. Our analysis introduces powerful tools from matrix theory to the theoretical computer science community. Also, spectral methods are being applied as heuristics for a variety of sequencing problems and our result helps explain and justify these applications. Although the worst case running time for our approach is not competitive with that of existing methods for well posed problem instances, unlike combinatorial approaches our algorithm remains a credible heuristic for the important cases where there are errors in the data.

  8. An Efficient, Non-iterative Method of Identifying the Cost-Effectiveness Frontier

    Science.gov (United States)

    Suen, Sze-chuan; Goldhaber-Fiebert, Jeremy D.

    2015-01-01

    Cost-effectiveness analysis aims to identify treatments and policies that maximize benefits subject to resource constraints. However, the conventional process of identifying the efficient frontier (i.e., the set of potentially cost-effective options) can be algorithmically inefficient, especially when considering a policy problem with many alternative options or when performing an extensive suite of sensitivity analyses for which the efficient frontier must be found for each. Here, we describe an alternative one-pass algorithm that is conceptually simple, easier to implement, and potentially faster for situations that challenge the conventional approach. Our algorithm accomplishes this by exploiting the relationship between the net monetary benefit and the cost-effectiveness plane. To facilitate further evaluation and use of this approach, we additionally provide scripts in R and Matlab that implement our method and can be used to identify efficient frontiers for any decision problem. PMID:25926282

  9. An algorithm for determination of peak regions and baseline elimination in spectroscopic data

    International Nuclear Information System (INIS)

    Morhac, Miroslav

    2009-01-01

    In the paper we propose a new algorithm for the determination of peaks containing regions and their separation from peak-free regions. Further based on this algorithm we propose a new background elimination algorithm which allows more accurate estimate of the background beneath the peaks than the algorithms known so far. The algorithm is based on a clipping operation with the window adjustable automatically to the widths of identified peak regions. The illustrative examples presented in the paper prove in favor of the proposed algorithms.

  10. Searching for the majority: algorithms of voluntary control.

    Directory of Open Access Journals (Sweden)

    Jin Fan

    Full Text Available Voluntary control of information processing is crucial to allocate resources and prioritize the processes that are most important under a given situation; the algorithms underlying such control, however, are often not clear. We investigated possible algorithms of control for the performance of the majority function, in which participants searched for and identified one of two alternative categories (left or right pointing arrows as composing the majority in each stimulus set. We manipulated the amount (set size of 1, 3, and 5 and content (ratio of left and right pointing arrows within a set of the inputs to test competing hypotheses regarding mental operations for information processing. Using a novel measure based on computational load, we found that reaction time was best predicted by a grouping search algorithm as compared to alternative algorithms (i.e., exhaustive or self-terminating search. The grouping search algorithm involves sampling and resampling of the inputs before a decision is reached. These findings highlight the importance of investigating the implications of voluntary control via algorithms of mental operations.

  11. Robust Selection Algorithm (RSA) for Multi-Omic Biomarker Discovery; Integration with Functional Network Analysis to Identify miRNA Regulated Pathways in Multiple Cancers.

    Science.gov (United States)

    Sehgal, Vasudha; Seviour, Elena G; Moss, Tyler J; Mills, Gordon B; Azencott, Robert; Ram, Prahlad T

    2015-01-01

    MicroRNAs (miRNAs) play a crucial role in the maintenance of cellular homeostasis by regulating the expression of their target genes. As such, the dysregulation of miRNA expression has been frequently linked to cancer. With rapidly accumulating molecular data linked to patient outcome, the need for identification of robust multi-omic molecular markers is critical in order to provide clinical impact. While previous bioinformatic tools have been developed to identify potential biomarkers in cancer, these methods do not allow for rapid classification of oncogenes versus tumor suppressors taking into account robust differential expression, cutoffs, p-values and non-normality of the data. Here, we propose a methodology, Robust Selection Algorithm (RSA) that addresses these important problems in big data omics analysis. The robustness of the survival analysis is ensured by identification of optimal cutoff values of omics expression, strengthened by p-value computed through intensive random resampling taking into account any non-normality in the data and integration into multi-omic functional networks. Here we have analyzed pan-cancer miRNA patient data to identify functional pathways involved in cancer progression that are associated with selected miRNA identified by RSA. Our approach demonstrates the way in which existing survival analysis techniques can be integrated with a functional network analysis framework to efficiently identify promising biomarkers and novel therapeutic candidates across diseases.

  12. Selfish Gene Algorithm Vs Genetic Algorithm: A Review

    Science.gov (United States)

    Ariff, Norharyati Md; Khalid, Noor Elaiza Abdul; Hashim, Rathiah; Noor, Noorhayati Mohamed

    2016-11-01

    Evolutionary algorithm is one of the algorithms inspired by the nature. Within little more than a decade hundreds of papers have reported successful applications of EAs. In this paper, the Selfish Gene Algorithms (SFGA), as one of the latest evolutionary algorithms (EAs) inspired from the Selfish Gene Theory which is an interpretation of Darwinian Theory ideas from the biologist Richards Dawkins on 1989. In this paper, following a brief introduction to the Selfish Gene Algorithm (SFGA), the chronology of its evolution is presented. It is the purpose of this paper is to present an overview of the concepts of Selfish Gene Algorithm (SFGA) as well as its opportunities and challenges. Accordingly, the history, step involves in the algorithm are discussed and its different applications together with an analysis of these applications are evaluated.

  13. MIIB: A Metric to Identify Top Influential Bloggers in a Community.

    Science.gov (United States)

    Khan, Hikmat Ullah; Daud, Ali; Malik, Tahir Afzal

    2015-01-01

    Social networking has revolutionized the use of conventional web and has converted World Wide Web into the social web as users can generate their own content. This change has been possible due to social web platforms like forums, wikis, and blogs. Blogs are more commonly being used as a form of virtual communication to express an opinion about an event, product or experience and can reach a large audience. Users can influence others to buy a product, have certain political or social views, etc. Therefore, identifying the most influential bloggers has become very significant as this can help us in the fields of commerce, advertisement and product knowledge searching. Existing approaches consider some basic features, but lack to consider some other features like the importance of the blog on which the post has been created. This paper presents a new metric, MIIB (Metric for Identification of Influential Bloggers), based on various features of bloggers' productivity and popularity. Productivity refers to bloggers' blogging activity and popularity measures bloggers' influence in the blogging community. The novel module of BlogRank depicts the importance of blog sites where bloggers create their posts. The MIIB has been evaluated against the standard model and existing metrics for finding the influential bloggers using dataset from the real-world blogosphere. The obtained results confirm that the MIIB is able to find the most influential bloggers in a more effective manner.

  14. MIIB: A Metric to Identify Top Influential Bloggers in a Community.

    Directory of Open Access Journals (Sweden)

    Hikmat Ullah Khan

    Full Text Available Social networking has revolutionized the use of conventional web and has converted World Wide Web into the social web as users can generate their own content. This change has been possible due to social web platforms like forums, wikis, and blogs. Blogs are more commonly being used as a form of virtual communication to express an opinion about an event, product or experience and can reach a large audience. Users can influence others to buy a product, have certain political or social views, etc. Therefore, identifying the most influential bloggers has become very significant as this can help us in the fields of commerce, advertisement and product knowledge searching. Existing approaches consider some basic features, but lack to consider some other features like the importance of the blog on which the post has been created. This paper presents a new metric, MIIB (Metric for Identification of Influential Bloggers, based on various features of bloggers' productivity and popularity. Productivity refers to bloggers' blogging activity and popularity measures bloggers' influence in the blogging community. The novel module of BlogRank depicts the importance of blog sites where bloggers create their posts. The MIIB has been evaluated against the standard model and existing metrics for finding the influential bloggers using dataset from the real-world blogosphere. The obtained results confirm that the MIIB is able to find the most influential bloggers in a more effective manner.

  15. A Systematic Review of Validated Methods for Identifying Cerebrovascular Accident or Transient Ischemic Attack Using Administrative Data

    Science.gov (United States)

    Andrade, Susan E.; Harrold, Leslie R.; Tjia, Jennifer; Cutrona, Sarah L.; Saczynski, Jane S.; Dodd, Katherine S.; Goldberg, Robert J.; Gurwitz, Jerry H.

    2012-01-01

    Purpose To perform a systematic review of the validity of algorithms for identifying cerebrovascular accidents (CVAs) or transient ischemic attacks (TIAs) using administrative and claims data. Methods PubMed and Iowa Drug Information Service (IDIS) searches of the English language literature were performed to identify studies published between 1990 and 2010 that evaluated the validity of algorithms for identifying CVAs (ischemic and hemorrhagic strokes, intracranial hemorrhage and subarachnoid hemorrhage) and/or TIAs in administrative data. Two study investigators independently reviewed the abstracts and articles to determine relevant studies according to pre-specified criteria. Results A total of 35 articles met the criteria for evaluation. Of these, 26 articles provided data to evaluate the validity of stroke, 7 reported the validity of TIA, 5 reported the validity of intracranial bleeds (intracerebral hemorrhage and subarachnoid hemorrhage), and 10 studies reported the validity of algorithms to identify the composite endpoints of stroke/TIA or cerebrovascular disease. Positive predictive values (PPVs) varied depending on the specific outcomes and algorithms evaluated. Specific algorithms to evaluate the presence of stroke and intracranial bleeds were found to have high PPVs (80% or greater). Algorithms to evaluate TIAs in adult populations were generally found to have PPVs of 70% or greater. Conclusions The algorithms and definitions to identify CVAs and TIAs using administrative and claims data differ greatly in the published literature. The choice of the algorithm employed should be determined by the stroke subtype of interest. PMID:22262598

  16. Algorithms

    Indian Academy of Sciences (India)

    to as 'divide-and-conquer'. Although there has been a large effort in realizing efficient algorithms, there are not many universally accepted algorithm design paradigms. In this article, we illustrate algorithm design techniques such as balancing, greedy strategy, dynamic programming strategy, and backtracking or traversal of ...

  17. Weighted community detection and data clustering using message passing

    Science.gov (United States)

    Shi, Cheng; Liu, Yanchen; Zhang, Pan

    2018-03-01

    Grouping objects into clusters based on the similarities or weights between them is one of the most important problems in science and engineering. In this work, by extending message-passing algorithms and spectral algorithms proposed for an unweighted community detection problem, we develop a non-parametric method based on statistical physics, by mapping the problem to the Potts model at the critical temperature of spin-glass transition and applying belief propagation to solve the marginals corresponding to the Boltzmann distribution. Our algorithm is robust to over-fitting and gives a principled way to determine whether there are significant clusters in the data and how many clusters there are. We apply our method to different clustering tasks. In the community detection problem in weighted and directed networks, we show that our algorithm significantly outperforms existing algorithms. In the clustering problem, where the data were generated by mixture models in the sparse regime, we show that our method works all the way down to the theoretical limit of detectability and gives accuracy very close to that of the optimal Bayesian inference. In the semi-supervised clustering problem, our method only needs several labels to work perfectly in classic datasets. Finally, we further develop Thouless-Anderson-Palmer equations which heavily reduce the computation complexity in dense networks but give almost the same performance as belief propagation.

  18. The secondary vertex finding algorithm with the ATLAS detector

    CERN Document Server

    Heer, Sebastian; The ATLAS collaboration

    2017-01-01

    A high performance identification of jets, produced via fragmentation of bottom quarks, is crucial for the ATLAS physics program. These jets can be identified by exploiting the presence of cascade decay vertices from bottom hadrons. A general vertex-finding algorithm is introduced and its ap- plication to the search for secondary vertices inside jets is described. Kinematic properties of the reconstructed vertices are used to construct several b-jet identification algorithms. The features and performance of the secondary vertex finding algorithm in a jet, as well as the performance of the jet tagging algorithms, are studied using simulated $pp$ -> $t\\bar{t}$ events at a centre-of-mass energy of 13 TeV.

  19. A New Algorithm for Identifying Cis-Regulatory Modules Based on Hidden Markov Model

    Directory of Open Access Journals (Sweden)

    Haitao Guo

    2017-01-01

    Full Text Available The discovery of cis-regulatory modules (CRMs is the key to understanding mechanisms of transcription regulation. Since CRMs have specific regulatory structures that are the basis for the regulation of gene expression, how to model the regulatory structure of CRMs has a considerable impact on the performance of CRM identification. The paper proposes a CRM discovery algorithm called ComSPS. ComSPS builds a regulatory structure model of CRMs based on HMM by exploring the rules of CRM transcriptional grammar that governs the internal motif site arrangement of CRMs. We test ComSPS on three benchmark datasets and compare it with five existing methods. Experimental results show that ComSPS performs better than them.

  20. A New Algorithm for Identifying Cis-Regulatory Modules Based on Hidden Markov Model

    Science.gov (United States)

    2017-01-01

    The discovery of cis-regulatory modules (CRMs) is the key to understanding mechanisms of transcription regulation. Since CRMs have specific regulatory structures that are the basis for the regulation of gene expression, how to model the regulatory structure of CRMs has a considerable impact on the performance of CRM identification. The paper proposes a CRM discovery algorithm called ComSPS. ComSPS builds a regulatory structure model of CRMs based on HMM by exploring the rules of CRM transcriptional grammar that governs the internal motif site arrangement of CRMs. We test ComSPS on three benchmark datasets and compare it with five existing methods. Experimental results show that ComSPS performs better than them. PMID:28497059

  1. Enhanced backpropagation training algorithm for transient event identification

    International Nuclear Information System (INIS)

    Vitela, J.; Reifman, J.

    1993-01-01

    We present an enhanced backpropagation (BP) algorithm for training feedforward neural networks that avoids the undesirable premature saturation of the network output nodes and accelerates the training process even in cases where premature saturation is not present. When the standard BP algorithm is applied to train patterns of nuclear power plant (NPP) transients, the network output nodes often become prematurely saturated causing the already slow rate of convergence of the algorithm to become even slower. When premature saturation occurs, the gradient of the prediction error becomes very small, although the prediction error itself is still large, yielding negligible weight updates and hence no significant decrease in the prediction error until the eventual recovery of the output nodes from saturation. By defining the onset of premature saturation and systematically modifying the gradient of the prediction error at saturation, we developed an enhanced BP algorithm that is compared with the standard BP algorithm in training a network to identify NPP transients

  2. Chiari malformation Type I surgery in pediatric patients. Part 1: validation of an ICD-9-CM code search algorithm.

    Science.gov (United States)

    Ladner, Travis R; Greenberg, Jacob K; Guerrero, Nicole; Olsen, Margaret A; Shannon, Chevis N; Yarbrough, Chester K; Piccirillo, Jay F; Anderson, Richard C E; Feldstein, Neil A; Wellons, John C; Smyth, Matthew D; Park, Tae Sung; Limbrick, David D

    2016-05-01

    OBJECTIVE Administrative billing data may facilitate large-scale assessments of treatment outcomes for pediatric Chiari malformation Type I (CM-I). Validated International Classification of Diseases, Ninth Revision, Clinical Modification (ICD-9-CM) code algorithms for identifying CM-I surgery are critical prerequisites for such studies but are currently only available for adults. The objective of this study was to validate two ICD-9-CM code algorithms using hospital billing data to identify pediatric patients undergoing CM-I decompression surgery. METHODS The authors retrospectively analyzed the validity of two ICD-9-CM code algorithms for identifying pediatric CM-I decompression surgery performed at 3 academic medical centers between 2001 and 2013. Algorithm 1 included any discharge diagnosis code of 348.4 (CM-I), as well as a procedure code of 01.24 (cranial decompression) or 03.09 (spinal decompression or laminectomy). Algorithm 2 restricted this group to the subset of patients with a primary discharge diagnosis of 348.4. The positive predictive value (PPV) and sensitivity of each algorithm were calculated. RESULTS Among 625 first-time admissions identified by Algorithm 1, the overall PPV for CM-I decompression was 92%. Among the 581 admissions identified by Algorithm 2, the PPV was 97%. The PPV for Algorithm 1 was lower in one center (84%) compared with the other centers (93%-94%), whereas the PPV of Algorithm 2 remained high (96%-98%) across all subgroups. The sensitivity of Algorithms 1 (91%) and 2 (89%) was very good and remained so across subgroups (82%-97%). CONCLUSIONS An ICD-9-CM algorithm requiring a primary diagnosis of CM-I has excellent PPV and very good sensitivity for identifying CM-I decompression surgery in pediatric patients. These results establish a basis for utilizing administrative billing data to assess pediatric CM-I treatment outcomes.

  3. Algorithms for optimizing drug therapy

    Directory of Open Access Journals (Sweden)

    Martin Lene

    2004-07-01

    Full Text Available Abstract Background Drug therapy has become increasingly efficient, with more drugs available for treatment of an ever-growing number of conditions. Yet, drug use is reported to be sub optimal in several aspects, such as dosage, patient's adherence and outcome of therapy. The aim of the current study was to investigate the possibility to optimize drug therapy using computer programs, available on the Internet. Methods One hundred and ten officially endorsed text documents, published between 1996 and 2004, containing guidelines for drug therapy in 246 disorders, were analyzed with regard to information about patient-, disease- and drug-related factors and relationships between these factors. This information was used to construct algorithms for identifying optimum treatment in each of the studied disorders. These algorithms were categorized in order to define as few models as possible that still could accommodate the identified factors and the relationships between them. The resulting program prototypes were implemented in HTML (user interface and JavaScript (program logic. Results Three types of algorithms were sufficient for the intended purpose. The simplest type is a list of factors, each of which implies that the particular patient should or should not receive treatment. This is adequate in situations where only one treatment exists. The second type, a more elaborate model, is required when treatment can by provided using drugs from different pharmacological classes and the selection of drug class is dependent on patient characteristics. An easily implemented set of if-then statements was able to manage the identified information in such instances. The third type was needed in the few situations where the selection and dosage of drugs were depending on the degree to which one or more patient-specific factors were present. In these cases the implementation of an established decision model based on fuzzy sets was required. Computer programs

  4. Algorithmic mathematics

    CERN Document Server

    Hougardy, Stefan

    2016-01-01

    Algorithms play an increasingly important role in nearly all fields of mathematics. This book allows readers to develop basic mathematical abilities, in particular those concerning the design and analysis of algorithms as well as their implementation. It presents not only fundamental algorithms like the sieve of Eratosthenes, the Euclidean algorithm, sorting algorithms, algorithms on graphs, and Gaussian elimination, but also discusses elementary data structures, basic graph theory, and numerical questions. In addition, it provides an introduction to programming and demonstrates in detail how to implement algorithms in C++. This textbook is suitable for students who are new to the subject and covers a basic mathematical lecture course, complementing traditional courses on analysis and linear algebra. Both authors have given this "Algorithmic Mathematics" course at the University of Bonn several times in recent years.

  5. Multi-scale graph-cut algorithm for efficient water-fat separation.

    Science.gov (United States)

    Berglund, Johan; Skorpil, Mikael

    2017-09-01

    To improve the accuracy and robustness to noise in water-fat separation by unifying the multiscale and graph cut based approaches to B 0 -correction. A previously proposed water-fat separation algorithm that corrects for B 0 field inhomogeneity in 3D by a single quadratic pseudo-Boolean optimization (QPBO) graph cut was incorporated into a multi-scale framework, where field map solutions are propagated from coarse to fine scales for voxels that are not resolved by the graph cut. The accuracy of the single-scale and multi-scale QPBO algorithms was evaluated against benchmark reference datasets. The robustness to noise was evaluated by adding noise to the input data prior to water-fat separation. Both algorithms achieved the highest accuracy when compared with seven previously published methods, while computation times were acceptable for implementation in clinical routine. The multi-scale algorithm was more robust to noise than the single-scale algorithm, while causing only a small increase (+10%) of the reconstruction time. The proposed 3D multi-scale QPBO algorithm offers accurate water-fat separation, robustness to noise, and fast reconstruction. The software implementation is freely available to the research community. Magn Reson Med 78:941-949, 2017. © 2016 International Society for Magnetic Resonance in Medicine. © 2016 International Society for Magnetic Resonance in Medicine.

  6. Application-Specific Graph Sampling for Frequent Subgraph Mining and Community Detection

    Energy Technology Data Exchange (ETDEWEB)

    Purohit, Sumit; Choudhury, Sutanay; Holder, Lawrence B.

    2017-12-11

    Graph mining is an important data analysis methodology, but struggles as the input graph size increases. The scalability and usability challenges posed by such large graphs make it imperative to sample the input graph and reduce its size. The critical challenge in sampling is to identify the appropriate algorithm to insure the resulting analysis does not suffer heavily from the data reduction. Predicting the expected performance degradation for a given graph and sampling algorithm is also useful. In this paper, we present different sampling approaches for graph mining applications such as Frequent Subgrpah Mining (FSM), and Community Detection (CD). We explore graph metrics such as PageRank, Triangles, and Diversity to sample a graph and conclude that for heterogeneous graphs Triangles and Diversity perform better than degree based metrics. We also present two new sampling variations for targeted graph mining applications. We present empirical results to show that knowledge of the target application, along with input graph properties can be used to select the best sampling algorithm. We also conclude that performance degradation is an abrupt, rather than gradual phenomena, as the sample size decreases. We present the empirical results to show that the performance degradation follows a logistic function.

  7. Genetic algorithms with memory- and elitism-based immigrants in dynamic environments.

    Science.gov (United States)

    Yang, Shengxiang

    2008-01-01

    In recent years the genetic algorithm community has shown a growing interest in studying dynamic optimization problems. Several approaches have been devised. The random immigrants and memory schemes are two major ones. The random immigrants scheme addresses dynamic environments by maintaining the population diversity while the memory scheme aims to adapt genetic algorithms quickly to new environments by reusing historical information. This paper investigates a hybrid memory and random immigrants scheme, called memory-based immigrants, and a hybrid elitism and random immigrants scheme, called elitism-based immigrants, for genetic algorithms in dynamic environments. In these schemes, the best individual from memory or the elite from the previous generation is retrieved as the base to create immigrants into the population by mutation. This way, not only can diversity be maintained but it is done more efficiently to adapt genetic algorithms to the current environment. Based on a series of systematically constructed dynamic problems, experiments are carried out to compare genetic algorithms with the memory-based and elitism-based immigrants schemes against genetic algorithms with traditional memory and random immigrants schemes and a hybrid memory and multi-population scheme. The sensitivity analysis regarding some key parameters is also carried out. Experimental results show that the memory-based and elitism-based immigrants schemes efficiently improve the performance of genetic algorithms in dynamic environments.

  8. Active Noise Control Using Modified FsLMS and Hybrid PSOFF Algorithm

    Directory of Open Access Journals (Sweden)

    Ranjan Walia

    2018-04-01

    Full Text Available Active noise control is an efficient technique for noise cancellation of the system, which has been defined in this paper with the aid of Modified Filtered-s Least Mean Square (MFsLMS algorithm. The Hybrid Particle Swarm Optimization and Firefly (HPSOFF algorithm are used to identify the stability factor of the MFsLMS algorithm. The computational difficulty of the modified algorithm is reduced when compared with the original Filtered-s Least Mean Square (FsLMS algorithm. The noise sources are removed from the signal and it is compared with the existing FsLMS algorithm. The performance of the system is established with the normalized mean square error for two different types of noises. The proposed method has also been compared with the existing algorithms for the same purposes.

  9. Multilabel user classification using the community structure of online networks.

    Science.gov (United States)

    Rizos, Georgios; Papadopoulos, Symeon; Kompatsiaris, Yiannis

    2017-01-01

    We study the problem of semi-supervised, multi-label user classification of networked data in the online social platform setting. We propose a framework that combines unsupervised community extraction and supervised, community-based feature weighting before training a classifier. We introduce Approximate Regularized Commute-Time Embedding (ARCTE), an algorithm that projects the users of a social graph onto a latent space, but instead of packing the global structure into a matrix of predefined rank, as many spectral and neural representation learning methods do, it extracts local communities for all users in the graph in order to learn a sparse embedding. To this end, we employ an improvement of personalized PageRank algorithms for searching locally in each user's graph structure. Then, we perform supervised community feature weighting in order to boost the importance of highly predictive communities. We assess our method performance on the problem of user classification by performing an extensive comparative study among various recent methods based on graph embeddings. The comparison shows that ARCTE significantly outperforms the competition in almost all cases, achieving up to 35% relative improvement compared to the second best competing method in terms of F1-score.

  10. Multilabel user classification using the community structure of online networks.

    Directory of Open Access Journals (Sweden)

    Georgios Rizos

    Full Text Available We study the problem of semi-supervised, multi-label user classification of networked data in the online social platform setting. We propose a framework that combines unsupervised community extraction and supervised, community-based feature weighting before training a classifier. We introduce Approximate Regularized Commute-Time Embedding (ARCTE, an algorithm that projects the users of a social graph onto a latent space, but instead of packing the global structure into a matrix of predefined rank, as many spectral and neural representation learning methods do, it extracts local communities for all users in the graph in order to learn a sparse embedding. To this end, we employ an improvement of personalized PageRank algorithms for searching locally in each user's graph structure. Then, we perform supervised community feature weighting in order to boost the importance of highly predictive communities. We assess our method performance on the problem of user classification by performing an extensive comparative study among various recent methods based on graph embeddings. The comparison shows that ARCTE significantly outperforms the competition in almost all cases, achieving up to 35% relative improvement compared to the second best competing method in terms of F1-score.

  11. COMMUNITY BASED HOME ENERGY MANAGEMENT SYSTEM

    Directory of Open Access Journals (Sweden)

    Muhammad Adnan Aziz

    2017-05-01

    Full Text Available In a Smart Grid (SG scenario, domestic consumers can gain cost reduction benefit by scheduling their Appliance Activation Time (AAT towards the slots of low charge. Minimization in cost is essential in Home Energy Management Systems (HEMS to induce consumers acceptance for power scheduling to accommodate for a Demand Response (DR at peak hours. Despite the fact that many algorithms address the power scheduling for HEMS, community based optimization has not been the focus. This paper presents an algorithm that targets the minimization of energy costs of whole community while keeping a low Peak to Average Ratio (PAR and smooth Power Usage Pattern (PUP. Objective of cost reduction is accomplished by finding most favorable AAT by Particle Swarm Optimization (PSO in conjunction with Inclined Block Rate (IBR approach and Circular Price Shift (CPS. Simulated numerical results demonstrate the effectiveness of CPS to assist the merger of PSO & IBR to enhance the reduction/stability of PAR and cost reduction.

  12. DNA Microarray Data Analysis: A Novel Biclustering Algorithm Approach

    Directory of Open Access Journals (Sweden)

    Tewfik Ahmed H

    2006-01-01

    Full Text Available Biclustering algorithms refer to a distinct class of clustering algorithms that perform simultaneous row-column clustering. Biclustering problems arise in DNA microarray data analysis, collaborative filtering, market research, information retrieval, text mining, electoral trends, exchange analysis, and so forth. When dealing with DNA microarray experimental data for example, the goal of biclustering algorithms is to find submatrices, that is, subgroups of genes and subgroups of conditions, where the genes exhibit highly correlated activities for every condition. In this study, we develop novel biclustering algorithms using basic linear algebra and arithmetic tools. The proposed biclustering algorithms can be used to search for all biclusters with constant values, biclusters with constant values on rows, biclusters with constant values on columns, and biclusters with coherent values from a set of data in a timely manner and without solving any optimization problem. We also show how one of the proposed biclustering algorithms can be adapted to identify biclusters with coherent evolution. The algorithms developed in this study discover all valid biclusters of each type, while almost all previous biclustering approaches will miss some.

  13. Algorithms for Cytoplasm Segmentation of Fluorescence Labelled Cells

    OpenAIRE

    Carolina Wählby; Joakim Lindblad; Mikael Vondrus; Ewert Bengtsson; Lennart Björkesten

    2002-01-01

    Automatic cell segmentation has various applications in cytometry, and while the nucleus is often very distinct and easy to identify, the cytoplasm provides a lot more challenge. A new combination of image analysis algorithms for segmentation of cells imaged by fluorescence microscopy is presented. The algorithm consists of an image pre?processing step, a general segmentation and merging step followed by a segmentation quality measurement. The quality measurement consists of a statistical ana...

  14. Use of a quality improvement tool, the prioritization matrix, to identify and prioritize triage software algorithm enhancement.

    Science.gov (United States)

    North, Frederick; Varkey, Prathiba; Caraballo, Pedro; Vsetecka, Darlene; Bartel, Greg

    2007-10-11

    Complex decision support software can require significant effort in maintenance and enhancement. A quality improvement tool, the prioritization matrix, was successfully used to guide software enhancement of algorithms in a symptom assessment call center.

  15. Essential algorithms a practical approach to computer algorithms

    CERN Document Server

    Stephens, Rod

    2013-01-01

    A friendly and accessible introduction to the most useful algorithms Computer algorithms are the basic recipes for programming. Professional programmers need to know how to use algorithms to solve difficult programming problems. Written in simple, intuitive English, this book describes how and when to use the most practical classic algorithms, and even how to create new algorithms to meet future needs. The book also includes a collection of questions that can help readers prepare for a programming job interview. Reveals methods for manipulating common data structures s

  16. Algorithms for spectral calibration of energy-resolving small-pixel detectors

    International Nuclear Information System (INIS)

    Scuffham, J; Veale, M C; Wilson, M D; Seller, P

    2013-01-01

    Small pixel Cd(Zn)Te detectors often suffer from inter-pixel variations in gain, resulting in shifts in the individual energy spectra. These gain variations are mainly caused by inclusions and defects within the crystal structure, which affect the charge transport within the material causing a decrease in the signal pulse height. In imaging applications, spectra are commonly integrated over a particular peak of interest. This means that the individual pixels must be accurately calibrated to ensure that the same portion of the spectrum is integrated in every pixel. The development of large-area detectors with fine pixel pitch necessitates automated algorithms for this spectral calibration, due to the very large number of pixels. Algorithms for automatic spectral calibration require accurate determination of characteristic x-ray or photopeak positions on a pixelwise basis. In this study, we compare two peak searching spectral calibration algorithms for a small-pixel CdTe detector in gamma spectroscopic imaging. The first algorithm uses rigid search ranges to identify peaks in each pixel spectrum, based on the average peak positions across all pixels. The second algorithm scales the search ranges on the basis of the position of the highest-energy peak relative to the average across all pixels. In test spectra acquired with Tc-99m, we found that the rigid search algorithm failed to correctly identify the target calibraton peaks in up to 4% of pixels. In contrast, the scaled search algorithm failed in only 0.16% of pixels. Failures in the scaled search algorithm were attributed to the presence of noise events above the main photopeak, and possible non-linearities in the spectral response in a small number of pixels. We conclude that a peak searching algorithm based on scaling known peak spacings is simple to implement and performs well for the spectral calibration of pixellated radiation detectors

  17. Intersectionality in the Lives of LGBTQ Youth: Identifying as LGBTQ and Finding Community in Small Cities and Rural Towns.

    Science.gov (United States)

    Hulko, Wendy; Hovanes, Jessica

    2018-01-01

    This article presents an analysis of the views of younger bisexual and lesbian women and transgender youth living in a western Canadian small city on their sexual and gender identities. Data were collected through focus groups and interviews and analyzed thematically through an intersectional lens. The purposive sample was composed of 13 youth who identified as lesbian, gay, bisexual, transgender, or queer (LGBTQ) and whose average age was 19.8 years. The analytical themes of (1) living in a small town, (2) identifying and being identified, and (3) talking intersectionality indicate that the sexual identities and gender identities and expressions of LGBTQ youth change across time and context and are impacted by often overlooked factors including faith, Indigenous ancestry, disability, and class. Further, the size and character of the community significantly impacts LGBTQ youth identity development and expression. This research demonstrates the uniqueness of individual youth's experiences-opposing notions of milestone events as singularly important in queer youth identity development.

  18. Algorithmic cryptanalysis

    CERN Document Server

    Joux, Antoine

    2009-01-01

    Illustrating the power of algorithms, Algorithmic Cryptanalysis describes algorithmic methods with cryptographically relevant examples. Focusing on both private- and public-key cryptographic algorithms, it presents each algorithm either as a textual description, in pseudo-code, or in a C code program.Divided into three parts, the book begins with a short introduction to cryptography and a background chapter on elementary number theory and algebra. It then moves on to algorithms, with each chapter in this section dedicated to a single topic and often illustrated with simple cryptographic applic

  19. Active Learning with Rationales for Identifying Operationally Significant Anomalies in Aviation

    Science.gov (United States)

    Sharma, Manali; Das, Kamalika; Bilgic, Mustafa; Matthews, Bryan; Nielsen, David Lynn; Oza, Nikunj C.

    2016-01-01

    A major focus of the commercial aviation community is discovery of unknown safety events in flight operations data. Data-driven unsupervised anomaly detection methods are better at capturing unknown safety events compared to rule-based methods which only look for known violations. However, not all statistical anomalies that are discovered by these unsupervised anomaly detection methods are operationally significant (e.g., represent a safety concern). Subject Matter Experts (SMEs) have to spend significant time reviewing these statistical anomalies individually to identify a few operationally significant ones. In this paper we propose an active learning algorithm that incorporates SME feedback in the form of rationales to build a classifier that can distinguish between uninteresting and operationally significant anomalies. Experimental evaluation on real aviation data shows that our approach improves detection of operationally significant events by as much as 75% compared to the state-of-the-art. The learnt classifier also generalizes well to additional validation data sets.

  20. An Ontology for Identifying Cyber Intrusion Induced Faults in Process Control Systems

    Science.gov (United States)

    Hieb, Jeffrey; Graham, James; Guan, Jian

    This paper presents an ontological framework that permits formal representations of process control systems, including elements of the process being controlled and the control system itself. A fault diagnosis algorithm based on the ontological model is also presented. The algorithm can identify traditional process elements as well as control system elements (e.g., IP network and SCADA protocol) as fault sources. When these elements are identified as a likely fault source, the possibility exists that the process fault is induced by a cyber intrusion. A laboratory-scale distillation column is used to illustrate the model and the algorithm. Coupled with a well-defined statistical process model, this fault diagnosis approach provides cyber security enhanced fault diagnosis information to plant operators and can help identify that a cyber attack is underway before a major process failure is experienced.

  1. Comprehensive eye evaluation algorithm

    Science.gov (United States)

    Agurto, C.; Nemeth, S.; Zamora, G.; Vahtel, M.; Soliz, P.; Barriga, S.

    2016-03-01

    In recent years, several research groups have developed automatic algorithms to detect diabetic retinopathy (DR) in individuals with diabetes (DM), using digital retinal images. Studies have indicated that diabetics have 1.5 times the annual risk of developing primary open angle glaucoma (POAG) as do people without DM. Moreover, DM patients have 1.8 times the risk for age-related macular degeneration (AMD). Although numerous investigators are developing automatic DR detection algorithms, there have been few successful efforts to create an automatic algorithm that can detect other ocular diseases, such as POAG and AMD. Consequently, our aim in the current study was to develop a comprehensive eye evaluation algorithm that not only detects DR in retinal images, but also automatically identifies glaucoma suspects and AMD by integrating other personal medical information with the retinal features. The proposed system is fully automatic and provides the likelihood of each of the three eye disease. The system was evaluated in two datasets of 104 and 88 diabetic cases. For each eye, we used two non-mydriatic digital color fundus photographs (macula and optic disc centered) and, when available, information about age, duration of diabetes, cataracts, hypertension, gender, and laboratory data. Our results show that the combination of multimodal features can increase the AUC by up to 5%, 7%, and 8% in the detection of AMD, DR, and glaucoma respectively. Marked improvement was achieved when laboratory results were combined with retinal image features.

  2. Denni Algorithm An Enhanced Of SMS (Scan, Move and Sort) Algorithm

    Science.gov (United States)

    Aprilsyah Lubis, Denni; Salim Sitompul, Opim; Marwan; Tulus; Andri Budiman, M.

    2017-12-01

    Sorting has been a profound area for the algorithmic researchers, and many resources are invested to suggest a more working sorting algorithm. For this purpose many existing sorting algorithms were observed in terms of the efficiency of the algorithmic complexity. Efficient sorting is important to optimize the use of other algorithms that require sorted lists to work correctly. Sorting has been considered as a fundamental problem in the study of algorithms that due to many reasons namely, the necessary to sort information is inherent in many applications, algorithms often use sorting as a key subroutine, in algorithm design there are many essential techniques represented in the body of sorting algorithms, and many engineering issues come to the fore when implementing sorting algorithms., Many algorithms are very well known for sorting the unordered lists, and one of the well-known algorithms that make the process of sorting to be more economical and efficient is SMS (Scan, Move and Sort) algorithm, an enhancement of Quicksort invented Rami Mansi in 2010. This paper presents a new sorting algorithm called Denni-algorithm. The Denni algorithm is considered as an enhancement on the SMS algorithm in average, and worst cases. The Denni algorithm is compared with the SMS algorithm and the results were promising.

  3. CAMPAIGN: an open-source library of GPU-accelerated data clustering algorithms.

    Science.gov (United States)

    Kohlhoff, Kai J; Sosnick, Marc H; Hsu, William T; Pande, Vijay S; Altman, Russ B

    2011-08-15

    Data clustering techniques are an essential component of a good data analysis toolbox. Many current bioinformatics applications are inherently compute-intense and work with very large datasets. Sequential algorithms are inadequate for providing the necessary performance. For this reason, we have created Clustering Algorithms for Massively Parallel Architectures, Including GPU Nodes (CAMPAIGN), a central resource for data clustering algorithms and tools that are implemented specifically for execution on massively parallel processing architectures. CAMPAIGN is a library of data clustering algorithms and tools, written in 'C for CUDA' for Nvidia GPUs. The library provides up to two orders of magnitude speed-up over respective CPU-based clustering algorithms and is intended as an open-source resource. New modules from the community will be accepted into the library and the layout of it is such that it can easily be extended to promising future platforms such as OpenCL. Releases of the CAMPAIGN library are freely available for download under the LGPL from https://simtk.org/home/campaign. Source code can also be obtained through anonymous subversion access as described on https://simtk.org/scm/?group_id=453. kjk33@cantab.net.

  4. (Automated) software modularization using community detection

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Manikas, Konstantinos

    2015-01-01

    The modularity of a software system is known to have an effect on, among other, development effort, change impact, and technical debt. Modularizing a specific system and evaluating this modularization is, however, challenging. In this paper, we apply community detection methods to the graph...... of class dependencies in software systems to find optimal modularizations through communities. We evaluate this approach through a study of 111 Java systems contained in the Qualitas Corpus. We found that using the modularity function of Newman with an Erdős-Rényi null-model and using the community...... detection algorithm of Reichardt and Bornholdt improved community quality for all systems, that coupling decreased for 99 of the systems, and that coherence increased for 102 of the systems. Furthermore, the modularity function correlates with existing metrics for coupling and coherence....

  5. Comparison of Two Phenotypic Algorithms To Detect Carbapenemase-Producing Enterobacteriaceae

    Science.gov (United States)

    Dortet, Laurent; Bernabeu, Sandrine; Gonzalez, Camille

    2017-01-01

    ABSTRACT A novel algorithm designed for the screening of carbapenemase-producing Enterobacteriaceae (CPE), based on faropenem and temocillin disks, was compared to that of the Committee of the Antibiogram of the French Society of Microbiology (CA-SFM), which is based on ticarcillin-clavulanate, imipenem, and temocillin disks. The two algorithms presented comparable negative predictive values (98.6% versus 97.5%) for CPE screening among carbapenem-nonsusceptible Enterobacteriaceae. However, since 46.2% (n = 49) of the CPE were correctly identified as OXA-48-like producers by the faropenem/temocillin-based algorithm, it significantly decreased the number of complementary tests needed (42.2% versus 62.6% with the CA-SFM algorithm). PMID:28607010

  6. Benchmarking motion planning algorithms for bin-picking applications

    DEFF Research Database (Denmark)

    Iversen, Thomas Fridolin; Ellekilde, Lars-Peter

    2017-01-01

    Purpose For robot motion planning there exists a large number of different algorithms, each appropriate for a certain domain, and the right choice of planner depends on the specific use case. The purpose of this paper is to consider the application of bin picking and benchmark a set of motion...... planning algorithms to identify which are most suited in the given context. Design/methodology/approach The paper presents a selection of motion planning algorithms and defines benchmarks based on three different bin-picking scenarios. The evaluation is done based on a fixed set of tasks, which are planned...... and executed on a real and a simulated robot. Findings The benchmarking shows a clear difference between the planners and generally indicates that algorithms integrating optimization, despite longer planning time, perform better due to a faster execution. Originality/value The originality of this work lies...

  7. Python algorithms mastering basic algorithms in the Python language

    CERN Document Server

    Hetland, Magnus Lie

    2014-01-01

    Python Algorithms, Second Edition explains the Python approach to algorithm analysis and design. Written by Magnus Lie Hetland, author of Beginning Python, this book is sharply focused on classical algorithms, but it also gives a solid understanding of fundamental algorithmic problem-solving techniques. The book deals with some of the most important and challenging areas of programming and computer science in a highly readable manner. It covers both algorithmic theory and programming practice, demonstrating how theory is reflected in real Python programs. Well-known algorithms and data struc

  8. The Soil Moisture Active Passive Mission (SMAP) Science Data Products: Results of Testing with Field Experiment and Algorithm Testbed Simulation Environment Data

    Science.gov (United States)

    Entekhabi, Dara; Njoku, Eni E.; O'Neill, Peggy E.; Kellogg, Kent H.; Entin, Jared K.

    2010-01-01

    Talk outline 1. Derivation of SMAP basic and applied science requirements from the NRC Earth Science Decadal Survey applications 2. Data products and latencies 3. Algorithm highlights 4. SMAP Algorithm Testbed 5. SMAP Working Groups and community engagement

  9. MXLKID: a maximum likelihood parameter identifier

    International Nuclear Information System (INIS)

    Gavel, D.T.

    1980-07-01

    MXLKID (MaXimum LiKelihood IDentifier) is a computer program designed to identify unknown parameters in a nonlinear dynamic system. Using noisy measurement data from the system, the maximum likelihood identifier computes a likelihood function (LF). Identification of system parameters is accomplished by maximizing the LF with respect to the parameters. The main body of this report briefly summarizes the maximum likelihood technique and gives instructions and examples for running the MXLKID program. MXLKID is implemented LRLTRAN on the CDC7600 computer at LLNL. A detailed mathematical description of the algorithm is given in the appendices. 24 figures, 6 tables

  10. Using a service sector segmented approach to identify community stakeholders who can improve access to suicide prevention services for veterans.

    Science.gov (United States)

    Matthieu, Monica M; Gardiner, Giovanina; Ziegemeier, Ellen; Buxton, Miranda

    2014-04-01

    Veterans in need of social services may access many different community agencies within the public and private sectors. Each of these settings has the potential to be a pipeline for attaining needed health, mental health, and benefits services; however, many service providers lack information on how to conceptualize where Veterans go for services within their local community. This article describes a conceptual framework for outreach that uses a service sector segmented approach. This framework was developed to aid recruitment of a provider-based sample of stakeholders (N = 70) for a study on improving access to the Department of Veterans Affairs and community-based suicide prevention services. Results indicate that although there are statistically significant differences in the percent of Veterans served by the different service sectors (F(9, 55) = 2.71, p = 0.04), exposure to suicidal Veterans and providers' referral behavior is consistent across the sectors. Challenges to using this framework include isolating the appropriate sectors for targeted outreach efforts. The service sector segmented approach holds promise for identifying and referring at-risk Veterans in need of services. Reprint & Copyright © 2014 Association of Military Surgeons of the U.S.

  11. Quantum algorithms for topological and geometric analysis of data

    Science.gov (United States)

    Lloyd, Seth; Garnerone, Silvano; Zanardi, Paolo

    2016-01-01

    Extracting useful information from large data sets can be a daunting task. Topological methods for analysing data sets provide a powerful technique for extracting such information. Persistent homology is a sophisticated tool for identifying topological features and for determining how such features persist as the data is viewed at different scales. Here we present quantum machine learning algorithms for calculating Betti numbers—the numbers of connected components, holes and voids—in persistent homology, and for finding eigenvectors and eigenvalues of the combinatorial Laplacian. The algorithms provide an exponential speed-up over the best currently known classical algorithms for topological data analysis. PMID:26806491

  12. Seismotectonic models and CN algorithm: The case of Italy

    International Nuclear Information System (INIS)

    Costa, G.; Orozova Stanishkova, I.; Panza, G.F.; Rotwain, I.M.

    1995-07-01

    The CN algorithm is here utilized both for the intermediate term earthquake prediction and to validate the seismotectonic model of the Italian territory. Using the results of the analysis, made through the CN algorithm and taking into account the seismotectonic model, three areas, one for Northern Italy, one for Central Italy and one for Southern Italy, are defined. Two transition areas, between the three main areas are delineated. The earthquakes which occurred in these two areas contribute to the precursor phenomena identified by the CN algorithm in each main area. (author). 26 refs, 6 figs, 2 tabs

  13. A semi-supervised classification algorithm using the TAD-derived background as training data

    Science.gov (United States)

    Fan, Lei; Ambeau, Brittany; Messinger, David W.

    2013-05-01

    In general, spectral image classification algorithms fall into one of two categories: supervised and unsupervised. In unsupervised approaches, the algorithm automatically identifies clusters in the data without a priori information about those clusters (except perhaps the expected number of them). Supervised approaches require an analyst to identify training data to learn the characteristics of the clusters such that they can then classify all other pixels into one of the pre-defined groups. The classification algorithm presented here is a semi-supervised approach based on the Topological Anomaly Detection (TAD) algorithm. The TAD algorithm defines background components based on a mutual k-Nearest Neighbor graph model of the data, along with a spectral connected components analysis. Here, the largest components produced by TAD are used as regions of interest (ROI's),or training data for a supervised classification scheme. By combining those ROI's with a Gaussian Maximum Likelihood (GML) or a Minimum Distance to the Mean (MDM) algorithm, we are able to achieve a semi supervised classification method. We test this classification algorithm against data collected by the HyMAP sensor over the Cooke City, MT area and University of Pavia scene.

  14. A Constrained Genetic Algorithm with Adaptively Defined Fitness Function in MRS Quantification

    Science.gov (United States)

    Papakostas, G. A.; Karras, D. A.; Mertzios, B. G.; Graveron-Demilly, D.; van Ormondt, D.

    MRS Signal quantification is a rather involved procedure and has attracted the interest of the medical engineering community, regarding the development of computationally efficient methodologies. Significant contributions based on Computational Intelligence tools, such as Neural Networks (NNs), demonstrated a good performance but not without drawbacks already discussed by the authors. On the other hand preliminary application of Genetic Algorithms (GA) has already been reported in the literature by the authors regarding the peak detection problem encountered in MRS quantification using the Voigt line shape model. This paper investigates a novel constrained genetic algorithm involving a generic and adaptively defined fitness function which extends the simple genetic algorithm methodology in case of noisy signals. The applicability of this new algorithm is scrutinized through experimentation in artificial MRS signals interleaved with noise, regarding its signal fitting capabilities. Although extensive experiments with real world MRS signals are necessary, the herein shown performance illustrates the method's potential to be established as a generic MRS metabolites quantification procedure.

  15. Hybrid Cryptosystem Using Tiny Encryption Algorithm and LUC Algorithm

    Science.gov (United States)

    Rachmawati, Dian; Sharif, Amer; Jaysilen; Andri Budiman, Mohammad

    2018-01-01

    Security becomes a very important issue in data transmission and there are so many methods to make files more secure. One of that method is cryptography. Cryptography is a method to secure file by writing the hidden code to cover the original file. Therefore, if the people do not involve in cryptography, they cannot decrypt the hidden code to read the original file. There are many methods are used in cryptography, one of that method is hybrid cryptosystem. A hybrid cryptosystem is a method that uses a symmetric algorithm to secure the file and use an asymmetric algorithm to secure the symmetric algorithm key. In this research, TEA algorithm is used as symmetric algorithm and LUC algorithm is used as an asymmetric algorithm. The system is tested by encrypting and decrypting the file by using TEA algorithm and using LUC algorithm to encrypt and decrypt the TEA key. The result of this research is by using TEA Algorithm to encrypt the file, the cipher text form is the character from ASCII (American Standard for Information Interchange) table in the form of hexadecimal numbers and the cipher text size increase by sixteen bytes as the plaintext length is increased by eight characters.

  16. TURING MACHINE AS UNIVERSAL ALGORITHM EXECUTOR AND ITS APPLICATION IN THE PROCESS OF HIGH-SCHOOL STUDENTS` ADVANCED STUDY OF ALGORITHMIZATION AND PROGRAMMING FUNDAMENTALS

    Directory of Open Access Journals (Sweden)

    Oleksandr B. Yashchyk

    2016-05-01

    Full Text Available The article discusses the importance of studying the notion of algorithm and its formal specification using Turing machines. In the article it was identified the basic hypothesis of the theory of algorithms for Turing as well as reviewed scientific research of modern scientists devoted to this issue and found the main principles of the Turing machine as an abstract mathematical model. The process of forming information competencies components, information culture and students` logical thinking development with the inclusion of the topic “Study and Application of Turing machine as Universal Algorithm Executor” in the course of Informatics was analyzed.

  17. Investigation of energy windowing algorithms for effective cargo screening with radiation portal monitors

    International Nuclear Information System (INIS)

    Hevener, Ryne; Yim, Man-Sung; Baird, Ken

    2013-01-01

    Radiation portal monitors (RPMs) are distributed across the globe in an effort to decrease the illicit trafficking of nuclear materials. Many current generation RPMs utilizes large polyvinyltoluene (PVT) plastic scintillators. These detectors are low cost and reliable but have very poor energy resolution. The lack of spectroscopic detail available from PVT spectra has restricted these systems primarily to performing simple gross counting measurements in the past. A common approach to extend the capability of PVT detectors beyond simple “gross-gamma” use is to apply a technique known as energy windowing (EW) to perform rough nuclide identification with limited spectral information. An approach to creating EW algorithms was developed in this work utilizing a specific set of calibration sources and modified EW equations; this algorithm provided a degree of increased identification capability. A simulated real-time emulation of the algorithm utilizing actual port-of-entry RPM data supplied by ORNL provided an extensive proving ground for the algorithm. This algorithm is able to identify four potential threat nuclides and the major NORM source with a high degree of accuracy. High-energy masking, a major detriment of EW algorithms, is reduced by the algorithm's design. - Highlights: • Gross counting algorithms do not produce detailed screenings. • Energy windowing algorithms enhance nuclide identification capability. • Proper use of EW algorithm can identify multiple threat nuclides. • Utilizing specific set of calibration sources is important for nuclide identification

  18. An Enhanced K-Means Algorithm for Water Quality Analysis of The Haihe River in China.

    Science.gov (United States)

    Zou, Hui; Zou, Zhihong; Wang, Xiaojing

    2015-11-12

    The increase and the complexity of data caused by the uncertain environment is today's reality. In order to identify water quality effectively and reliably, this paper presents a modified fast clustering algorithm for water quality analysis. The algorithm has adopted a varying weights K-means cluster algorithm to analyze water monitoring data. The varying weights scheme was the best weighting indicator selected by a modified indicator weight self-adjustment algorithm based on K-means, which is named MIWAS-K-means. The new clustering algorithm avoids the margin of the iteration not being calculated in some cases. With the fast clustering analysis, we can identify the quality of water samples. The algorithm is applied in water quality analysis of the Haihe River (China) data obtained by the monitoring network over a period of eight years (2006-2013) with four indicators at seven different sites (2078 samples). Both the theoretical and simulated results demonstrate that the algorithm is efficient and reliable for water quality analysis of the Haihe River. In addition, the algorithm can be applied to more complex data matrices with high dimensionality.

  19. A flexible fuzzy regression algorithm for forecasting oil consumption estimation

    International Nuclear Information System (INIS)

    Azadeh, A.; Khakestani, M.; Saberi, M.

    2009-01-01

    Oil consumption plays a vital role in socio-economic development of most countries. This study presents a flexible fuzzy regression algorithm for forecasting oil consumption based on standard economic indicators. The standard indicators are annual population, cost of crude oil import, gross domestic production (GDP) and annual oil production in the last period. The proposed algorithm uses analysis of variance (ANOVA) to select either fuzzy regression or conventional regression for future demand estimation. The significance of the proposed algorithm is three fold. First, it is flexible and identifies the best model based on the results of ANOVA and minimum absolute percentage error (MAPE), whereas previous studies consider the best fitted fuzzy regression model based on MAPE or other relative error results. Second, the proposed model may identify conventional regression as the best model for future oil consumption forecasting because of its dynamic structure, whereas previous studies assume that fuzzy regression always provide the best solutions and estimation. Third, it utilizes the most standard independent variables for the regression models. To show the applicability and superiority of the proposed flexible fuzzy regression algorithm the data for oil consumption in Canada, United States, Japan and Australia from 1990 to 2005 are used. The results show that the flexible algorithm provides accurate solution for oil consumption estimation problem. The algorithm may be used by policy makers to accurately foresee the behavior of oil consumption in various regions.

  20. Common and Critical Components Among Community Health Assessment and Community Health Improvement Planning Models.

    Science.gov (United States)

    Pennel, Cara L; Burdine, James N; Prochaska, John D; McLeroy, Kenneth R

    Community health assessment and community health improvement planning are continuous, systematic processes for assessing and addressing health needs in a community. Since there are different models to guide assessment and planning, as well as a variety of organizations and agencies that carry out these activities, there may be confusion in choosing among approaches. By examining the various components of the different assessment and planning models, we are able to identify areas for coordination, ways to maximize collaboration, and strategies to further improve community health. We identified 11 common assessment and planning components across 18 models and requirements, with a particular focus on health department, health system, and hospital models and requirements. These common components included preplanning; developing partnerships; developing vision and scope; collecting, analyzing, and interpreting data; identifying community assets; identifying priorities; developing and implementing an intervention plan; developing and implementing an evaluation plan; communicating and receiving feedback on the assessment findings and/or the plan; planning for sustainability; and celebrating success. Within several of these components, we discuss characteristics that are critical to improving community health. Practice implications include better understanding of different models and requirements by health departments, hospitals, and others involved in assessment and planning to improve cross-sector collaboration, collective impact, and community health. In addition, federal and state policy and accreditation requirements may be revised or implemented to better facilitate assessment and planning collaboration between health departments, hospitals, and others for the purpose of improving community health.

  1. Sound algorithms

    OpenAIRE

    De Götzen , Amalia; Mion , Luca; Tache , Olivier

    2007-01-01

    International audience; We call sound algorithms the categories of algorithms that deal with digital sound signal. Sound algorithms appeared in the very infancy of computer. Sound algorithms present strong specificities that are the consequence of two dual considerations: the properties of the digital sound signal itself and its uses, and the properties of auditory perception.

  2. How well do discharge diagnoses identify hospitalised patients with community-acquired infections? - a validation study

    DEFF Research Database (Denmark)

    Henriksen, Daniel Pilsgaard; Nielsen, Stig Lønberg; Laursen, Christian Borbjerg

    2014-01-01

    -10 diagnoses was 79.9% (95%CI: 78.1-81.3%), specificity 83.9% (95%CI: 82.6-85.1%), positive likelihood ratio 4.95 (95%CI: 4.58-5.36) and negative likelihood ratio 0.24 (95%CI: 0.22-0.26). The two most common sites of infection, the lower respiratory tract and urinary tract, had positive likelihood......BACKGROUND: Credible measures of disease incidence, trends and mortality can be obtained through surveillance using manual chart review, but this is both time-consuming and expensive. ICD-10 discharge diagnoses are used as surrogate markers of infection, but knowledge on the validity of infections...... in general is sparse. The aim of the study was to determine how well ICD-10 discharge diagnoses identify patients with community-acquired infections in a medical emergency department (ED), overall and related to sites of infection and patient characteristics. METHODS: We manually reviewed 5977 patients...

  3. CSA: An efficient algorithm to improve circular DNA multiple alignment

    Directory of Open Access Journals (Sweden)

    Pereira Luísa

    2009-07-01

    Full Text Available Abstract Background The comparison of homologous sequences from different species is an essential approach to reconstruct the evolutionary history of species and of the genes they harbour in their genomes. Several complete mitochondrial and nuclear genomes are now available, increasing the importance of using multiple sequence alignment algorithms in comparative genomics. MtDNA has long been used in phylogenetic analysis and errors in the alignments can lead to errors in the interpretation of evolutionary information. Although a large number of multiple sequence alignment algorithms have been proposed to date, they all deal with linear DNA and cannot handle directly circular DNA. Researchers interested in aligning circular DNA sequences must first rotate them to the "right" place using an essentially manual process, before they can use multiple sequence alignment tools. Results In this paper we propose an efficient algorithm that identifies the most interesting region to cut circular genomes in order to improve phylogenetic analysis when using standard multiple sequence alignment algorithms. This algorithm identifies the largest chain of non-repeated longest subsequences common to a set of circular mitochondrial DNA sequences. All the sequences are then rotated and made linear for multiple alignment purposes. To evaluate the effectiveness of this new tool, three different sets of mitochondrial DNA sequences were considered. Other tests considering randomly rotated sequences were also performed. The software package Arlequin was used to evaluate the standard genetic measures of the alignments obtained with and without the use of the CSA algorithm with two well known multiple alignment algorithms, the CLUSTALW and the MAVID tools, and also the visualization tool SinicView. Conclusion The results show that a circularization and rotation pre-processing step significantly improves the efficiency of public available multiple sequence alignment

  4. Application of particle swarm optimization to identify gamma spectrum with neural network

    International Nuclear Information System (INIS)

    Shi Dongsheng; Di Yuming; Zhou Chunlin

    2007-01-01

    In applying neural network to identification of gamma spectra back propagation (BP) algorithm is usually trapped to a local optimum and has a low speed of convergence, whereas particle swarm optimization (PSO) is advantageous in terms of globe optimal searching. In this paper, we propose a new algorithm for neural network training, i.e. combined BP and PSO optimization, or PSO-BP algorithm. Practical example shows that the new algorithm can overcome shortcomings of BP algorithm and the neural network trained by it has a high ability of generalization with identification result of 100% correctness. It can be used effectively and reliably to identify gamma spectra. (authors)

  5. Multiobjective optimization of classifiers by means of 3D convex-hull-based evolutionary algorithms

    NARCIS (Netherlands)

    Zhao, J.; Basto, Fernandes V.; Jiao, L.; Yevseyeva, I.; Asep, Maulana A.; Li, R.; Bäck, T.H.W.; Tang, T.; Michael, Emmerich T. M.

    2016-01-01

    The receiver operating characteristic (ROC) and detection error tradeoff(DET) curves are frequently used in the machine learning community to analyze the performance of binary classifiers. Recently, the convex-hull-based multiobjective genetic programming algorithm was proposed and successfully

  6. A novel self-organizing E-Learner community model with award and exchange mechanisms.

    Science.gov (United States)

    Yang, Fan; Shen, Rui-min; Han, Peng

    2004-11-01

    How to share experience and resources among learners is becoming one of the hottest topics in the field of E-Learning collaborative techniques. An intuitive way to achieve this objective is to group learners which can help each other into the same community and help them learn collaboratively. In this paper, we proposed a novel community self-organization model based on multi-agent mechanism, which can automatically group learners with similar preferences and capabilities. In particular, we proposed award and exchange schemas with evaluation and preference track records to raise the performance of this algorithm. The description of learner capability, the matchmaking process, the definition of evaluation and preference track records, the rules of award and exchange schemas and the self-organization algorithm are all discussed in this paper. Meanwhile, a prototype has been built to verify the validity and efficiency of the algorithm. Experiments based on real learner data showed that this mechanism can organize learner communities properly and efficiently; and that it has sustainable improved efficiency and scalability.

  7. A speedup technique for (l, d-motif finding algorithms

    Directory of Open Access Journals (Sweden)

    Dinh Hieu

    2011-03-01

    Full Text Available Abstract Background The discovery of patterns in DNA, RNA, and protein sequences has led to the solution of many vital biological problems. For instance, the identification of patterns in nucleic acid sequences has resulted in the determination of open reading frames, identification of promoter elements of genes, identification of intron/exon splicing sites, identification of SH RNAs, location of RNA degradation signals, identification of alternative splicing sites, etc. In protein sequences, patterns have proven to be extremely helpful in domain identification, location of protease cleavage sites, identification of signal peptides, protein interactions, determination of protein degradation elements, identification of protein trafficking elements, etc. Motifs are important patterns that are helpful in finding transcriptional regulatory elements, transcription factor binding sites, functional genomics, drug design, etc. As a result, numerous papers have been written to solve the motif search problem. Results Three versions of the motif search problem have been proposed in the literature: Simple Motif Search (SMS, (l, d-motif search (or Planted Motif Search (PMS, and Edit-distance-based Motif Search (EMS. In this paper we focus on PMS. Two kinds of algorithms can be found in the literature for solving the PMS problem: exact and approximate. An exact algorithm identifies the motifs always and an approximate algorithm may fail to identify some or all of the motifs. The exact version of PMS problem has been shown to be NP-hard. Exact algorithms proposed in the literature for PMS take time that is exponential in some of the underlying parameters. In this paper we propose a generic technique that can be used to speedup PMS algorithms. Conclusions We present a speedup technique that can be used on any PMS algorithm. We have tested our speedup technique on a number of algorithms. These experimental results show that our speedup technique is indeed very

  8. Community detection, link prediction, and layer interdependence in multilayer networks

    Science.gov (United States)

    De Bacco, Caterina; Power, Eleanor A.; Larremore, Daniel B.; Moore, Cristopher

    2017-04-01

    Complex systems are often characterized by distinct types of interactions between the same entities. These can be described as a multilayer network where each layer represents one type of interaction. These layers may be interdependent in complicated ways, revealing different kinds of structure in the network. In this work we present a generative model, and an efficient expectation-maximization algorithm, which allows us to perform inference tasks such as community detection and link prediction in this setting. Our model assumes overlapping communities that are common between the layers, while allowing these communities to affect each layer in a different way, including arbitrary mixtures of assortative, disassortative, or directed structure. It also gives us a mathematically principled way to define the interdependence between layers, by measuring how much information about one layer helps us predict links in another layer. In particular, this allows us to bundle layers together to compress redundant information and identify small groups of layers which suffice to predict the remaining layers accurately. We illustrate these findings by analyzing synthetic data and two real multilayer networks, one representing social support relationships among villagers in South India and the other representing shared genetic substring material between genes of the malaria parasite.

  9. General simulation algorithm for autocorrelated binary processes.

    Science.gov (United States)

    Serinaldi, Francesco; Lombardo, Federico

    2017-02-01

    The apparent ubiquity of binary random processes in physics and many other fields has attracted considerable attention from the modeling community. However, generation of binary sequences with prescribed autocorrelation is a challenging task owing to the discrete nature of the marginal distributions, which makes the application of classical spectral techniques problematic. We show that such methods can effectively be used if we focus on the parent continuous process of beta distributed transition probabilities rather than on the target binary process. This change of paradigm results in a simulation procedure effectively embedding a spectrum-based iterative amplitude-adjusted Fourier transform method devised for continuous processes. The proposed algorithm is fully general, requires minimal assumptions, and can easily simulate binary signals with power-law and exponentially decaying autocorrelation functions corresponding, for instance, to Hurst-Kolmogorov and Markov processes. An application to rainfall intermittency shows that the proposed algorithm can also simulate surrogate data preserving the empirical autocorrelation.

  10. General simulation algorithm for autocorrelated binary processes

    Science.gov (United States)

    Serinaldi, Francesco; Lombardo, Federico

    2017-02-01

    The apparent ubiquity of binary random processes in physics and many other fields has attracted considerable attention from the modeling community. However, generation of binary sequences with prescribed autocorrelation is a challenging task owing to the discrete nature of the marginal distributions, which makes the application of classical spectral techniques problematic. We show that such methods can effectively be used if we focus on the parent continuous process of beta distributed transition probabilities rather than on the target binary process. This change of paradigm results in a simulation procedure effectively embedding a spectrum-based iterative amplitude-adjusted Fourier transform method devised for continuous processes. The proposed algorithm is fully general, requires minimal assumptions, and can easily simulate binary signals with power-law and exponentially decaying autocorrelation functions corresponding, for instance, to Hurst-Kolmogorov and Markov processes. An application to rainfall intermittency shows that the proposed algorithm can also simulate surrogate data preserving the empirical autocorrelation.

  11. Store-Carry and Forward-Type M2M Communication Protocol Enabling Guide Robots to Work together and the Method of Identifying Malfunctioning Robots Using the Byzantine Algorithm

    Directory of Open Access Journals (Sweden)

    Yoshio Suga

    2016-11-01

    Full Text Available This paper concerns a service in which multiple guide robots in an area display arrows to guide individual users to their destinations. It proposes a method of identifying malfunctioning robots and robots that give wrong directions to users. In this method, users’ mobile terminals and robots form a store-carry and forward-type M2M communication network, and a distributed cooperative protocol is used to enable robots to share information and identify malfunctioning robots using the Byzantine algorithm. The robots do not directly communicate with each other, but through users’ mobile terminals. We have introduced the concept of the quasi-synchronous number, so whether a certain robot is malfunctioning can be determined even when items of information held by all of the robots are not synchronized. Using simulation, we have evaluated the proposed method in terms of the rate of identifying malfunctioning robots, the rate of reaching the destination and the average length of time to reach the destination.

  12. Genetic algorithms

    Science.gov (United States)

    Wang, Lui; Bayer, Steven E.

    1991-01-01

    Genetic algorithms are mathematical, highly parallel, adaptive search procedures (i.e., problem solving methods) based loosely on the processes of natural genetics and Darwinian survival of the fittest. Basic genetic algorithms concepts are introduced, genetic algorithm applications are introduced, and results are presented from a project to develop a software tool that will enable the widespread use of genetic algorithm technology.

  13. Impact of Reconstruction Algorithms on CT Radiomic Features of Pulmonary Tumors: Analysis of Intra- and Inter-Reader Variability and Inter-Reconstruction Algorithm Variability.

    Science.gov (United States)

    Kim, Hyungjin; Park, Chang Min; Lee, Myunghee; Park, Sang Joon; Song, Yong Sub; Lee, Jong Hyuk; Hwang, Eui Jin; Goo, Jin Mo

    2016-01-01

    To identify the impact of reconstruction algorithms on CT radiomic features of pulmonary tumors and to reveal and compare the intra- and inter-reader and inter-reconstruction algorithm variability of each feature. Forty-two patients (M:F = 19:23; mean age, 60.43±10.56 years) with 42 pulmonary tumors (22.56±8.51mm) underwent contrast-enhanced CT scans, which were reconstructed with filtered back projection and commercial iterative reconstruction algorithm (level 3 and 5). Two readers independently segmented the whole tumor volume. Fifteen radiomic features were extracted and compared among reconstruction algorithms. Intra- and inter-reader variability and inter-reconstruction algorithm variability were calculated using coefficients of variation (CVs) and then compared. Among the 15 features, 5 first-order tumor intensity features and 4 gray level co-occurrence matrix (GLCM)-based features showed significant differences (palgorithms. As for the variability, effective diameter, sphericity, entropy, and GLCM entropy were the most robust features (CV≤5%). Inter-reader variability was larger than intra-reader or inter-reconstruction algorithm variability in 9 features. However, for entropy, homogeneity, and 4 GLCM-based features, inter-reconstruction algorithm variability was significantly greater than inter-reader variability (palgorithms. Inter-reconstruction algorithm variability was greater than inter-reader variability for entropy, homogeneity, and GLCM-based features.

  14. Could a brief assessment of negative emotions and self-esteem identify adolescents at current and future risk of self-harm in the community? A prospective cohort analysis.

    Science.gov (United States)

    Phillips, Rhiannon; Spears, Melissa R; Montgomery, Alan A; Millings, Abigail; Sayal, Kapil; Stallard, Paul

    2013-06-22

    Self-harm is common in adolescents, but it is often unreported and undetected. Available screening tools typically ask directly about self-harm and suicidal ideation. Although in an ideal world, direct enquiry and open discussion around self-harm would be advocated, non-psychiatric professionals in community settings are often reluctant to ask about this directly and disclosure can be met with feeling of intense anxiety. Training non-specialist staff to directly ask about self-harm has limited effects suggesting that alternative approaches are required. This study investigated whether a targeted analysis of negative emotions and self-esteem could identify young adolescents at risk of self-harm in community settings. Data were collected as part of a clinical trial from young people in school years 8-11 (aged 12-16) at eight UK secondary schools (N = 4503 at baseline, N = 3263 in prospective analysis). The Short Mood and Feelings Questionnaire, Revised Child Anxiety and Depression Scale, Rosenberg Self-Esteem Scale, personal failure (Children's Automatic Thoughts Scale), and two items on self-harm were completed at baseline, 6 and 12 months. Following a process of Principal Components Analysis, item reduction, and logistic regression analysis, three internally reliable factors were identified from the original measures that were independently associated with current and future self-harm; personal failure (3 items), physical symptoms of depression/anxiety (6 items), positive self-esteem (5 items). The summed score of these 14 items had good accuracy in identifying current self-harm (AUC 0.87 girls, 0.81 boys) and at six months for girls (0.81), and fair accuracy at six months for boys (AUC 0.74) and 12 months for girls (AUC 0.77). A brief and targeted assessment of negative emotions and self-esteem, focusing on factors that are strongly associated with current and future self-harm, could potentially be used to help identify adolescents who are at risk in

  15. Enhancing Breast Cancer Recurrence Algorithms Through Selective Use of Medical Record Data.

    Science.gov (United States)

    Kroenke, Candyce H; Chubak, Jessica; Johnson, Lisa; Castillo, Adrienne; Weltzien, Erin; Caan, Bette J

    2016-03-01

    The utility of data-based algorithms in research has been questioned because of errors in identification of cancer recurrences. We adapted previously published breast cancer recurrence algorithms, selectively using medical record (MR) data to improve classification. We evaluated second breast cancer event (SBCE) and recurrence-specific algorithms previously published by Chubak and colleagues in 1535 women from the Life After Cancer Epidemiology (LACE) and 225 women from the Women's Health Initiative cohorts and compared classification statistics to published values. We also sought to improve classification with minimal MR examination. We selected pairs of algorithms-one with high sensitivity/high positive predictive value (PPV) and another with high specificity/high PPV-using MR information to resolve discrepancies between algorithms, properly classifying events based on review; we called this "triangulation." Finally, in LACE, we compared associations between breast cancer survival risk factors and recurrence using MR data, single Chubak algorithms, and triangulation. The SBCE algorithms performed well in identifying SBCE and recurrences. Recurrence-specific algorithms performed more poorly than published except for the high-specificity/high-PPV algorithm, which performed well. The triangulation method (sensitivity = 81.3%, specificity = 99.7%, PPV = 98.1%, NPV = 96.5%) improved recurrence classification over two single algorithms (sensitivity = 57.1%, specificity = 95.5%, PPV = 71.3%, NPV = 91.9%; and sensitivity = 74.6%, specificity = 97.3%, PPV = 84.7%, NPV = 95.1%), with 10.6% MR review. Triangulation performed well in survival risk factor analyses vs analyses using MR-identified recurrences. Use of multiple recurrence algorithms in administrative data, in combination with selective examination of MR data, may improve recurrence data quality and reduce research costs. © The Author 2015. Published by Oxford University Press. All rights reserved. For

  16. An experiment framework to identify community functional components driving ecosystem processes and services delivery.

    NARCIS (Netherlands)

    Dias, A.; Berg, M.P.; de Bello, F.; van Oosten, A.R.; Bila, K.; Moretti, M.

    2013-01-01

    There is a growing consensus that the distribution of species trait values in a community can greatly determine ecosystem processes and services delivery. Two distinct components of community trait composition are hypothesized to chiefly affect ecosystem processes: (i) the average trait value of the

  17. Sampling from complex networks with high community structures.

    Science.gov (United States)

    Salehi, Mostafa; Rabiee, Hamid R; Rajabi, Arezo

    2012-06-01

    In this paper, we propose a novel link-tracing sampling algorithm, based on the concepts from PageRank vectors, to sample from networks with high community structures. Our method has two phases; (1) Sampling the closest nodes to the initial nodes by approximating personalized PageRank vectors and (2) Jumping to a new community by using PageRank vectors and unknown neighbors. Empirical studies on several synthetic and real-world networks show that the proposed method improves the performance of network sampling compared to the popular link-based sampling methods in terms of accuracy and visited communities.

  18. Quantum Google algorithm. Construction and application to complex networks

    Science.gov (United States)

    Paparo, G. D.; Müller, M.; Comellas, F.; Martin-Delgado, M. A.

    2014-07-01

    We review the main findings on the ranking capabilities of the recently proposed Quantum PageRank algorithm (G.D. Paparo et al., Sci. Rep. 2, 444 (2012) and G.D. Paparo et al., Sci. Rep. 3, 2773 (2013)) applied to large complex networks. The algorithm has been shown to identify unambiguously the underlying topology of the network and to be capable of clearly highlighting the structure of secondary hubs of networks. Furthermore, it can resolve the degeneracy in importance of the low-lying part of the list of rankings. Examples of applications include real-world instances from the WWW, which typically display a scale-free network structure and models of hierarchical networks. The quantum algorithm has been shown to display an increased stability with respect to a variation of the damping parameter, present in the Google algorithm, and a more clearly pronounced power-law behaviour in the distribution of importance among the nodes, as compared to the classical algorithm.

  19. Kriging-based algorithm for nuclear reactor neutronic design optimization

    International Nuclear Information System (INIS)

    Kempf, Stephanie; Forget, Benoit; Hu, Lin-Wen

    2012-01-01

    Highlights: ► A Kriging-based algorithm was selected to guide research reactor optimization. ► We examined impacts of parameter values upon the algorithm. ► The best parameter values were incorporated into a set of best practices. ► Algorithm with best practices used to optimize thermal flux of concept. ► Final design produces thermal flux 30% higher than other 5 MW reactors. - Abstract: Kriging, a geospatial interpolation technique, has been used in the present work to drive a search-and-optimization algorithm which produces the optimum geometric parameters for a 5 MW research reactor design. The technique has been demonstrated to produce an optimal neutronic solution after a relatively small number of core calculations. It has additionally been successful in producing a design which significantly improves thermal neutron fluxes by 30% over existing reactors of the same power rating. Best practices for use of this algorithm in reactor design were identified and indicated the importance of selecting proper correlation functions.

  20. ANOMALY DETECTION IN NETWORKING USING HYBRID ARTIFICIAL IMMUNE ALGORITHM

    Directory of Open Access Journals (Sweden)

    D. Amutha Guka

    2012-01-01

    Full Text Available Especially in today’s network scenario, when computers are interconnected through internet, security of an information system is very important issue. Because no system can be absolutely secure, the timely and accurate detection of anomalies is necessary. The main aim of this research paper is to improve the anomaly detection by using Hybrid Artificial Immune Algorithm (HAIA which is based on Artificial Immune Systems (AIS and Genetic Algorithm (GA. In this research work, HAIA approach is used to develop Network Anomaly Detection System (NADS. The detector set is generated by using GA and the anomalies are identified using Negative Selection Algorithm (NSA which is based on AIS. The HAIA algorithm is tested with KDD Cup 99 benchmark dataset. The detection rate is used to measure the effectiveness of the NADS. The results and consistency of the HAIA are compared with earlier approaches and the results are presented. The proposed algorithm gives best results when compared to the earlier approaches.

  1. Modeling the heterogeneous traffic correlations in urban road systems using traffic-enhanced community detection approach

    Science.gov (United States)

    Lu, Feng; Liu, Kang; Duan, Yingying; Cheng, Shifen; Du, Fei

    2018-07-01

    A better characterization of the traffic influence among urban roads is crucial for traffic control and traffic forecasting. The existence of spatial heterogeneity imposes great influence on modeling the extent and degree of road traffic correlation, which is usually neglected by the traditional distance based method. In this paper, we propose a traffic-enhanced community detection approach to spatially reveal the traffic correlation in city road networks. First, the road network is modeled as a traffic-enhanced dual graph with the closeness between two road segments determined not only by their topological connection, but also by the traffic correlation between them. Then a flow-based community detection algorithm called Infomap is utilized to identify the road segment clusters. Evaluated by Moran's I, Calinski-Harabaz Index and the traffic interpolation application, we find that compared to the distance based method and the community based method, our proposed traffic-enhanced community based method behaves better in capturing the extent of traffic relevance as both the topological structure of the road network and the traffic correlations among urban roads are considered. It can be used in more traffic-related applications, such as traffic forecasting, traffic control and guidance.

  2. Distributed Coordination of Electric Vehicle Charging in a Community Microgrid Considering Real-Time Price

    DEFF Research Database (Denmark)

    Li, Chendan; Schaltz, Erik; Quintero, Juan Carlos Vasquez

    2016-01-01

    The predictable increasing adoption of EV by residential users imposes the necessity of Electric Vehicle charging coordination, in order to charge effectively while minimizing the impact on the grid. In this paper, a two-stage distributed coordination algorithm for electric vehicle charging...... management in a community microgrid is proposed. Each local EV charging controller is taken as an agent, which can manage the charging to achieve the optimization of the whole community by communicating in a sparse network. The proposed algorithm aims at optimizing real-time, which manages the charging...

  3. Algorithm aversion: people erroneously avoid algorithms after seeing them err.

    Science.gov (United States)

    Dietvorst, Berkeley J; Simmons, Joseph P; Massey, Cade

    2015-02-01

    Research shows that evidence-based algorithms more accurately predict the future than do human forecasters. Yet when forecasters are deciding whether to use a human forecaster or a statistical algorithm, they often choose the human forecaster. This phenomenon, which we call algorithm aversion, is costly, and it is important to understand its causes. We show that people are especially averse to algorithmic forecasters after seeing them perform, even when they see them outperform a human forecaster. This is because people more quickly lose confidence in algorithmic than human forecasters after seeing them make the same mistake. In 5 studies, participants either saw an algorithm make forecasts, a human make forecasts, both, or neither. They then decided whether to tie their incentives to the future predictions of the algorithm or the human. Participants who saw the algorithm perform were less confident in it, and less likely to choose it over an inferior human forecaster. This was true even among those who saw the algorithm outperform the human.

  4. Optimization of Multiple Traveling Salesman Problem Based on Simulated Annealing Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Xu Mingji

    2017-01-01

    Full Text Available It is very effective to solve the multi variable optimization problem by using hierarchical genetic algorithm. This thesis analyzes both advantages and disadvantages of hierarchical genetic algorithm and puts forward an improved simulated annealing genetic algorithm. The new algorithm is applied to solve the multiple traveling salesman problem, which can improve the performance of the solution. First, it improves the design of chromosomes hierarchical structure in terms of redundant hierarchical algorithm, and it suggests a suffix design of chromosomes; Second, concerning to some premature problems of genetic algorithm, it proposes a self-identify crossover operator and mutation; Third, when it comes to the problem of weak ability of local search of genetic algorithm, it stretches the fitness by mixing genetic algorithm with simulated annealing algorithm. Forth, it emulates the problems of N traveling salesmen and M cities so as to verify its feasibility. The simulation and calculation shows that this improved algorithm can be quickly converged to a best global solution, which means the algorithm is encouraging in practical uses.

  5. Kmer-SSR: a fast and exhaustive SSR search algorithm.

    Science.gov (United States)

    Pickett, Brandon D; Miller, Justin B; Ridge, Perry G

    2017-12-15

    One of the main challenges with bioinformatics software is that the size and complexity of datasets necessitate trading speed for accuracy, or completeness. To combat this problem of computational complexity, a plethora of heuristic algorithms have arisen that report a 'good enough' solution to biological questions. However, in instances such as Simple Sequence Repeats (SSRs), a 'good enough' solution may not accurately portray results in population genetics, phylogenetics and forensics, which require accurate SSRs to calculate intra- and inter-species interactions. We present Kmer-SSR, which finds all SSRs faster than most heuristic SSR identification algorithms in a parallelized, easy-to-use manner. The exhaustive Kmer-SSR option has 100% precision and 100% recall and accurately identifies every SSR of any specified length. To identify more biologically pertinent SSRs, we also developed several filters that allow users to easily view a subset of SSRs based on user input. Kmer-SSR, coupled with the filter options, accurately and intuitively identifies SSRs quickly and in a more user-friendly manner than any other SSR identification algorithm. The source code is freely available on GitHub at https://github.com/ridgelab/Kmer-SSR. perry.ridge@byu.edu. © The Author(s) 2017. Published by Oxford University Press.

  6. Utility of Neck Circumference for Identifying Metabolic Syndrome by Different Definitions in Chinese Subjects over 50 Years Old: A Community-Based Study

    Directory of Open Access Journals (Sweden)

    Shuo Lin

    2018-01-01

    Full Text Available Aims. Whether neck circumference (NC could be used as a valuable tool for identifying metabolic syndrome (MS by different criteria in Chinese is still unclear. Methods. We conducted a cross-sectional survey from October 2010 to January 2011 in Shipai community, Guangzhou, Guangdong Province, China. A total of 1473 subjects aged over 50 years were investigated. We measured height, weight, NC, waist circumference, blood pressure, blood glucose, and lipids in all subjects. MS was identified by criteria of the National Cholesterol Education Program-Adult Treatment Panel III (NCEP-ATP III, Chinese Diabetes Society (CDS, and International Diabetes Federation (IDF. Results. Mean NC was 38.0 ± 2.7 cm in men and 34.2 ± 2.5 cm in women. By using receiver operating characteristic curves, the area under the curve (AUC of NC for identifying MS (IDF was 0.823 in men and 0.777 in women, while for identifying MS (CDS, it was 0.788 in men and 0.762 in women. The AUC of NC for diagnosing MS (ATP III was 0.776 in men and 0.752 in women. The optimal cut points of NC for MS were 38.5 cm by three definitions in men, while those were 34.2 cm, 33.4 cm, and 34.0 cm in women by IDF, ATP III, and CDS definitions, respectively. No significant difference was observed between the AUC of NC and BMI for diagnosing MS by using different criteria (all p>0.05. Conclusions. NC is associated with MS by different definitions in Chinese subjects over 50 years old. It may be a useful tool to identify MS in a community population.

  7. Identifying community healthcare supports for the elderly and the factors affecting their aging care model preference: evidence from three districts of Beijing

    Directory of Open Access Journals (Sweden)

    Tianyang Liu

    2016-11-01

    Full Text Available Abstract Background The Chinese tradition of filial piety, which prioritized family-based care for the elderly, is transitioning and elders can no longer necessarily rely on their children. The purpose of this study was to identify community support for the elderly, and analyze the factors that affect which model of old-age care elderly people dwelling in communities prefer. Methods We used the database “Health and Social Support of Elderly Population in Community”. Questionnaires were issued in 2013, covering 3 districts in Beijing. A group of 1036 people over 60 years in age were included in the study. The respondents’ profile variables were organized in Andersen’s Model and community healthcare resource factors were added. A multinomial logistic model was applied to analyze the factors associated with the desired aging care models. Results Cohabiting with children and relying on care from family was still the primary desired aging care model for seniors (78 %, followed by living in institutions (14.8 % and living at home independently while relying on community resources (7.2 %. The regression result indicated that predisposing, enabling and community factors were significantly associated with the aging care model preference. Specifically, compared with those who preferred to cohabit with children, those having higher education, fewer available family and friend helpers, and shorter distance to healthcare center were more likely to prefer to live independently and rely on community support. And compared with choosing to live in institutions, those having fewer available family and friend helpers and those living alone were more likely to prefer to live independently and rely on community. Need factors (health and disability condition were not significantly associated with desired aging care models, indicating that desired aging care models were passive choices resulted from the balancing of family and social caring resources

  8. ERGC: an efficient referential genome compression algorithm.

    Science.gov (United States)

    Saha, Subrata; Rajasekaran, Sanguthevar

    2015-11-01

    Genome sequencing has become faster and more affordable. Consequently, the number of available complete genomic sequences is increasing rapidly. As a result, the cost to store, process, analyze and transmit the data is becoming a bottleneck for research and future medical applications. So, the need for devising efficient data compression and data reduction techniques for biological sequencing data is growing by the day. Although there exists a number of standard data compression algorithms, they are not efficient in compressing biological data. These generic algorithms do not exploit some inherent properties of the sequencing data while compressing. To exploit statistical and information-theoretic properties of genomic sequences, we need specialized compression algorithms. Five different next-generation sequencing data compression problems have been identified and studied in the literature. We propose a novel algorithm for one of these problems known as reference-based genome compression. We have done extensive experiments using five real sequencing datasets. The results on real genomes show that our proposed algorithm is indeed competitive and performs better than the best known algorithms for this problem. It achieves compression ratios that are better than those of the currently best performing algorithms. The time to compress and decompress the whole genome is also very promising. The implementations are freely available for non-commercial purposes. They can be downloaded from http://engr.uconn.edu/∼rajasek/ERGC.zip. rajasek@engr.uconn.edu. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  9. The Texas Medication Algorithm Project (TMAP) schizophrenia algorithms.

    Science.gov (United States)

    Miller, A L; Chiles, J A; Chiles, J K; Crismon, M L; Rush, A J; Shon, S P

    1999-10-01

    In the Texas Medication Algorithm Project (TMAP), detailed guidelines for medication management of schizophrenia and related disorders, bipolar disorders, and major depressive disorders have been developed and implemented. This article describes the algorithms developed for medication treatment of schizophrenia and related disorders. The guidelines recommend a sequence of medications and discuss dosing, duration, and switch-over tactics. They also specify response criteria at each stage of the algorithm for both positive and negative symptoms. The rationale and evidence for each aspect of the algorithms are presented.

  10. Community detection for networks with unipartite and bipartite structure

    Science.gov (United States)

    Chang, Chang; Tang, Chao

    2014-09-01

    Finding community structures in networks is important in network science, technology, and applications. To date, most algorithms that aim to find community structures only focus either on unipartite or bipartite networks. A unipartite network consists of one set of nodes and a bipartite network consists of two nonoverlapping sets of nodes with only links joining the nodes in different sets. However, a third type of network exists, defined here as the mixture network. Just like a bipartite network, a mixture network also consists of two sets of nodes, but some nodes may simultaneously belong to two sets, which breaks the nonoverlapping restriction of a bipartite network. The mixture network can be considered as a general case, with unipartite and bipartite networks viewed as its limiting cases. A mixture network can represent not only all the unipartite and bipartite networks, but also a wide range of real-world networks that cannot be properly represented as either unipartite or bipartite networks in fields such as biology and social science. Based on this observation, we first propose a probabilistic model that can find modules in unipartite, bipartite, and mixture networks in a unified framework based on the link community model for a unipartite undirected network [B Ball et al (2011 Phys. Rev. E 84 036103)]. We test our algorithm on synthetic networks (both overlapping and nonoverlapping communities) and apply it to two real-world networks: a southern women bipartite network and a human transcriptional regulatory mixture network. The results suggest that our model performs well for all three types of networks, is competitive with other algorithms for unipartite or bipartite networks, and is applicable to real-world networks.

  11. Applications of expectation maximization algorithm for coherent optical communication

    DEFF Research Database (Denmark)

    Carvalho, L.; Oliveira, J.; Zibar, Darko

    2014-01-01

    In this invited paper, we present powerful statistical signal processing methods, used by machine learning community, and link them to current problems in optical communication. In particular, we will look into iterative maximum likelihood parameter estimation based on expectation maximization...... algorithm and its application in coherent optical communication systems for linear and nonlinear impairment mitigation. Furthermore, the estimated parameters are used to build the probabilistic model of the system for the synthetic impairment generation....

  12. Innovative Software Algorithms and Tools parallel sessions summary

    International Nuclear Information System (INIS)

    Gaines, Irwin

    2001-01-01

    A variety of results were presented in the poster and 5 parallel sessions of the Innovative Software, Algorithms and Tools (ISAT) sessions. I will briefly summarize these presentations and attempt to identify some unifying trends

  13. Loss-minimal Algorithmic Trading Based on Levy Processes

    Directory of Open Access Journals (Sweden)

    Farhad Kia

    2014-08-01

    Full Text Available In this paper we optimize portfolios assuming that the value of the portfolio follows a Lévy process. First we identify the parameters of the underlying Lévy process and then portfolio optimization is performed by maximizing the probability of positive return. The method has been tested by extensive performance analysis on Forex and SP 500 historical time series. The proposed trading algorithm has achieved 4.9\\% percent yearly return on average without leverage which proves its applicability to algorithmic trading.

  14. Improvement of the SEP protocol based on community structure of node degree

    Science.gov (United States)

    Li, Donglin; Wei, Suyuan

    2017-05-01

    Analyzing the Stable election protocol (SEP) in wireless sensor networks and aiming at the problem of inhomogeneous cluster-heads distribution and unreasonable cluster-heads selectivity and single hop transmission in the SEP, a SEP Protocol based on community structure of node degree (SEP-CSND) is proposed. In this algorithm, network node deployed by using grid deployment model, and the connection between nodes established by setting up the communication threshold. The community structure constructed by node degree, then cluster head is elected in the community structure. On the basis of SEP, the node's residual energy and node degree is added in cluster-heads election. The information is transmitted with mode of multiple hops between network nodes. The simulation experiments showed that compared to the classical LEACH and SEP, this algorithm balances the energy consumption of the entire network and significantly prolongs network lifetime.

  15. Performance Analysis of Binary Search Algorithm in RFID

    Directory of Open Access Journals (Sweden)

    Xiangmei SONG

    2014-12-01

    Full Text Available Binary search algorithm (BS is a kind of important anti-collision algorithm in the Radio Frequency Identification (RFID, is also one of the key technologies which determine whether the information in the tag is identified by the reader-writer fast and reliably. The performance of BS directly affects the quality of service in Internet of Things. This paper adopts an automated formal technology: probabilistic model checking to analyze the performance of BS algorithm formally. Firstly, according to the working principle of BS algorithm, its dynamic behavior is abstracted into a Discrete Time Markov Chains which can describe deterministic, discrete time and the probability selection. And then on the model we calculate the probability of the data sent successfully and the expected time of tags completing the data transmission. Compared to the another typical anti-collision protocol S-ALOHA in RFID, experimental results show that with an increase in the number of tags the BS algorithm has a less space and time consumption, the average number of conflicts increases slower than the S-ALOHA protocol standard, BS algorithm needs fewer expected time to complete the data transmission, and the average speed of the data transmission in BS is as 1.6 times as the S-ALOHA protocol.

  16. Semantic integration to identify overlapping functional modules in protein interaction networks

    Directory of Open Access Journals (Sweden)

    Ramanathan Murali

    2007-07-01

    Full Text Available Abstract Background The systematic analysis of protein-protein interactions can enable a better understanding of cellular organization, processes and functions. Functional modules can be identified from the protein interaction networks derived from experimental data sets. However, these analyses are challenging because of the presence of unreliable interactions and the complex connectivity of the network. The integration of protein-protein interactions with the data from other sources can be leveraged for improving the effectiveness of functional module detection algorithms. Results We have developed novel metrics, called semantic similarity and semantic interactivity, which use Gene Ontology (GO annotations to measure the reliability of protein-protein interactions. The protein interaction networks can be converted into a weighted graph representation by assigning the reliability values to each interaction as a weight. We presented a flow-based modularization algorithm to efficiently identify overlapping modules in the weighted interaction networks. The experimental results show that the semantic similarity and semantic interactivity of interacting pairs were positively correlated with functional co-occurrence. The effectiveness of the algorithm for identifying modules was evaluated using functional categories from the MIPS database. We demonstrated that our algorithm had higher accuracy compared to other competing approaches. Conclusion The integration of protein interaction networks with GO annotation data and the capability of detecting overlapping modules substantially improve the accuracy of module identification.

  17. Pheochromocytoma in Denmark 1977-2016: Identifying a national cohort using patterns of health registrations

    DEFF Research Database (Denmark)

    Ebbehoj, A; Jacobsen, Sarah Forslund; Trolle, Christian

    for all persons living in the North and Central Regions of Denmark to validate the diagnosis of PPGL. We tested a number of algorithms for accurately identifying true cases of PPGL to maximize positive predictive values (PPV) and completeness. The best algorithm was validated in an external sample of 110...... algorithm-positive patients. Results: We identified 2626 persons with a PPGL diagnosis code in Denmark, including 787 (30.0%) in North and Central Denmark. We retrieved the health records of 771/787 (98.0%) persons and confirmed 198 incident PPGL patients (25.3%). By combining patterns of registrations, we...... created an algorithm with a PPV for PPGL of 93.1% (CI95%: 88.5-96.3) and completeness of 88.9% (CI95%: 83.7-92.9), thus creating a national PPGL cohort of 587 patients. PPV for the optimal algorithm was 94.3% (CI95%: 87.1-98.1) in the external validation sample. Conclusion: Diagnosis codes...

  18. Algorithming the Algorithm

    DEFF Research Database (Denmark)

    Mahnke, Martina; Uprichard, Emma

    2014-01-01

    Imagine sailing across the ocean. The sun is shining, vastness all around you. And suddenly [BOOM] you’ve hit an invisible wall. Welcome to the Truman Show! Ever since Eli Pariser published his thoughts on a potential filter bubble, this movie scenario seems to have become reality, just with slight...... changes: it’s not the ocean, it’s the internet we’re talking about, and it’s not a TV show producer, but algorithms that constitute a sort of invisible wall. Building on this assumption, most research is trying to ‘tame the algorithmic tiger’. While this is a valuable and often inspiring approach, we...

  19. Validation of a clinical practice-based algorithm for the diagnosis of autosomal recessive cerebellar ataxias based on NGS identified cases.

    Science.gov (United States)

    Mallaret, Martial; Renaud, Mathilde; Redin, Claire; Drouot, Nathalie; Muller, Jean; Severac, Francois; Mandel, Jean Louis; Hamza, Wahiba; Benhassine, Traki; Ali-Pacha, Lamia; Tazir, Meriem; Durr, Alexandra; Monin, Marie-Lorraine; Mignot, Cyril; Charles, Perrine; Van Maldergem, Lionel; Chamard, Ludivine; Thauvin-Robinet, Christel; Laugel, Vincent; Burglen, Lydie; Calvas, Patrick; Fleury, Marie-Céline; Tranchant, Christine; Anheim, Mathieu; Koenig, Michel

    2016-07-01

    Establishing a molecular diagnosis of autosomal recessive cerebellar ataxias (ARCA) is challenging due to phenotype and genotype heterogeneity. We report the validation of a previously published clinical practice-based algorithm to diagnose ARCA. Two assessors performed a blind analysis to determine the most probable mutated gene based on comprehensive clinical and paraclinical data, without knowing the molecular diagnosis of 23 patients diagnosed by targeted capture of 57 ataxia genes and high-throughput sequencing coming from a 145 patients series. The correct gene was predicted in 61 and 78 % of the cases by the two assessors, respectively. There was a high inter-rater agreement [K = 0.85 (0.55-0.98) p < 0.001] confirming the algorithm's reproducibility. Phenotyping patients with proper clinical examination, imaging, biochemical investigations and nerve conduction studies remain crucial for the guidance of molecular analysis and to interpret next generation sequencing results. The proposed algorithm should be helpful for diagnosing ARCA in clinical practice.

  20. Pseudo-deterministic Algorithms

    OpenAIRE

    Goldwasser , Shafi

    2012-01-01

    International audience; In this talk we describe a new type of probabilistic algorithm which we call Bellagio Algorithms: a randomized algorithm which is guaranteed to run in expected polynomial time, and to produce a correct and unique solution with high probability. These algorithms are pseudo-deterministic: they can not be distinguished from deterministic algorithms in polynomial time by a probabilistic polynomial time observer with black box access to the algorithm. We show a necessary an...

  1. Motion Cueing Algorithm Development: Human-Centered Linear and Nonlinear Approaches

    Science.gov (United States)

    Houck, Jacob A. (Technical Monitor); Telban, Robert J.; Cardullo, Frank M.

    2005-01-01

    While the performance of flight simulator motion system hardware has advanced substantially, the development of the motion cueing algorithm, the software that transforms simulated aircraft dynamics into realizable motion commands, has not kept pace. Prior research identified viable features from two algorithms: the nonlinear "adaptive algorithm", and the "optimal algorithm" that incorporates human vestibular models. A novel approach to motion cueing, the "nonlinear algorithm" is introduced that combines features from both approaches. This algorithm is formulated by optimal control, and incorporates a new integrated perception model that includes both visual and vestibular sensation and the interaction between the stimuli. Using a time-varying control law, the matrix Riccati equation is updated in real time by a neurocomputing approach. Preliminary pilot testing resulted in the optimal algorithm incorporating a new otolith model, producing improved motion cues. The nonlinear algorithm vertical mode produced a motion cue with a time-varying washout, sustaining small cues for longer durations and washing out large cues more quickly compared to the optimal algorithm. The inclusion of the integrated perception model improved the responses to longitudinal and lateral cues. False cues observed with the NASA adaptive algorithm were absent. The neurocomputing approach was crucial in that the number of presentations of an input vector could be reduced to meet the real time requirement without degrading the quality of the motion cues.

  2. Effects of multiple spreaders in community networks

    Science.gov (United States)

    Hu, Zhao-Long; Ren, Zhuo-Ming; Yang, Guang-Yong; Liu, Jian-Guo

    2014-12-01

    Human contact networks exhibit the community structure. Understanding how such community structure affects the epidemic spreading could provide insights for preventing the spreading of epidemics between communities. In this paper, we explore the spreading of multiple spreaders in community networks. A network based on the clustering preferential mechanism is evolved, whose communities are detected by the Girvan-Newman (GN) algorithm. We investigate the spreading effectiveness by selecting the nodes as spreaders in the following ways: nodes with the largest degree in each community (community hubs), the same number of nodes with the largest degree from the global network (global large-degree) and randomly selected one node within each community (community random). The experimental results on the SIR model show that the spreading effectiveness based on the global large-degree and community hubs methods is the same in the early stage of the infection and the method of community random is the worst. However, when the infection rate exceeds the critical value, the global large-degree method embodies the worst spreading effectiveness. Furthermore, the discrepancy of effectiveness for the three methods will decrease as the infection rate increases. Therefore, we should immunize the hubs in each community rather than those hubs in the global network to prevent the outbreak of epidemics.

  3. The Algorithmic Imaginary

    DEFF Research Database (Denmark)

    Bucher, Taina

    2017-01-01

    the notion of the algorithmic imaginary. It is argued that the algorithmic imaginary – ways of thinking about what algorithms are, what they should be and how they function – is not just productive of different moods and sensations but plays a generative role in moulding the Facebook algorithm itself...... of algorithms affect people's use of these platforms, if at all? To help answer these questions, this article examines people's personal stories about the Facebook algorithm through tweets and interviews with 25 ordinary users. To understand the spaces where people and algorithms meet, this article develops...

  4. Algorithms in practice: Comparing web journalism and criminal justice

    Directory of Open Access Journals (Sweden)

    Angèle Christin

    2017-07-01

    Full Text Available Big Data evangelists often argue that algorithms make decision-making more informed and objective—a promise hotly contested by critics of these technologies. Yet, to date, most of the debate has focused on the instruments themselves, rather than on how they are used. This article addresses this lack by examining the actual practices surrounding algorithmic technologies. Specifically, drawing on multi-sited ethnographic data, I compare how algorithms are used and interpreted in two institutional contexts with markedly different characteristics: web journalism and criminal justice. I find that there are surprising similarities in how web journalists and legal professionals use algorithms in their work. In both cases, I document a gap between the intended and actual effects of algorithms—a process I analyze as “decoupling.” Second, I identify a gamut of buffering strategies used by both web journalists and legal professionals to minimize the impact of algorithms in their daily work. Those include foot-dragging, gaming, and open critique. Of course, these similarities do not exhaust the differences between the two cases, which are explored in the discussion section. I conclude with a call for further ethnographic work on algorithms in practice as an important empirical check against the dominant rhetoric of algorithmic power.

  5. Rural Embedded Assistants for Community Health (REACH) network: first-person accounts in a community-university partnership.

    Science.gov (United States)

    Brown, Louis D; Alter, Theodore R; Brown, Leigh Gordon; Corbin, Marilyn A; Flaherty-Craig, Claire; McPhail, Lindsay G; Nevel, Pauline; Shoop, Kimbra; Sterner, Glenn; Terndrup, Thomas E; Weaver, M Ellen

    2013-03-01

    Community research and action projects undertaken by community-university partnerships can lead to contextually appropriate and sustainable community improvements in rural and urban localities. However, effective implementation is challenging and prone to failure when poorly executed. The current paper seeks to inform rural community-university partnership practice through consideration of first-person accounts from five stakeholders in the Rural Embedded Assistants for Community Health (REACH) Network. The REACH Network is a unique community-university partnership aimed at improving rural health services by identifying, implementing, and evaluating innovative health interventions delivered by local caregivers. The first-person accounts provide an insider's perspective on the nature of collaboration. The unique perspectives identify three critical challenges facing the REACH Network: trust, coordination, and sustainability. Through consideration of the challenges, we identified several strategies for success. We hope readers can learn their own lessons when considering the details of our partnership's efforts to improve the delivery infrastructure for rural healthcare.

  6. Time-Delay System Identification Using Genetic Algorithm

    DEFF Research Database (Denmark)

    Yang, Zhenyu; Seested, Glen Thane

    2013-01-01

    problem through an identification approach using the real coded Genetic Algorithm (GA). The desired FOPDT/SOPDT model is directly identified based on the measured system's input and output data. In order to evaluate the quality and performance of this GA-based approach, the proposed method is compared...

  7. Applied economic model development algorithm for electronics company

    Directory of Open Access Journals (Sweden)

    Mikhailov I.

    2017-01-01

    Full Text Available The purpose of this paper is to report about received experience in the field of creating the actual methods and algorithms that help to simplify development of applied decision support systems. It reports about an algorithm, which is a result of two years research and have more than one-year practical verification. In a case of testing electronic components, the time of the contract conclusion is crucial point to make the greatest managerial mistake. At this stage, it is difficult to achieve a realistic assessment of time-limit and of wage-fund for future work. The creation of estimating model is possible way to solve this problem. In the article is represented an algorithm for creation of those models. The algorithm is based on example of the analytical model development that serves for amount of work estimation. The paper lists the algorithm’s stages and explains their meanings with participants’ goals. The implementation of the algorithm have made possible twofold acceleration of these models development and fulfilment of management’s requirements. The resulting models have made a significant economic effect. A new set of tasks was identified to be further theoretical study.

  8. Detecting the overlapping and hierarchical community structure in complex networks

    International Nuclear Information System (INIS)

    Lancichinetti, Andrea; Fortunato, Santo; Kertesz, Janos

    2009-01-01

    Many networks in nature, society and technology are characterized by a mesoscopic level of organization, with groups of nodes forming tightly connected units, called communities or modules, that are only weakly linked to each other. Uncovering this community structure is one of the most important problems in the field of complex networks. Networks often show a hierarchical organization, with communities embedded within other communities; moreover, nodes can be shared between different communities. Here, we present the first algorithm that finds both overlapping communities and the hierarchical structure. The method is based on the local optimization of a fitness function. Community structure is revealed by peaks in the fitness histogram. The resolution can be tuned by a parameter enabling different hierarchical levels of organization to be investigated. Tests on real and artificial networks give excellent results.

  9. Identifying online user reputation of user-object bipartite networks

    Science.gov (United States)

    Liu, Xiao-Lu; Liu, Jian-Guo; Yang, Kai; Guo, Qiang; Han, Jing-Ti

    2017-02-01

    Identifying online user reputation based on the rating information of the user-object bipartite networks is important for understanding online user collective behaviors. Based on the Bayesian analysis, we present a parameter-free algorithm for ranking online user reputation, where the user reputation is calculated based on the probability that their ratings are consistent with the main part of all user opinions. The experimental results show that the AUC values of the presented algorithm could reach 0.8929 and 0.8483 for the MovieLens and Netflix data sets, respectively, which is better than the results generated by the CR and IARR methods. Furthermore, the experimental results for different user groups indicate that the presented algorithm outperforms the iterative ranking methods in both ranking accuracy and computation complexity. Moreover, the results for the synthetic networks show that the computation complexity of the presented algorithm is a linear function of the network size, which suggests that the presented algorithm is very effective and efficient for the large scale dynamic online systems.

  10. An Adaptive Filtering Algorithm Based on Genetic Algorithm-Backpropagation Network

    Directory of Open Access Journals (Sweden)

    Kai Hu

    2013-01-01

    Full Text Available A new image filtering algorithm is proposed. GA-BPN algorithm uses genetic algorithm (GA to decide weights in a back propagation neural network (BPN. It has better global optimal characteristics than traditional optimal algorithm. In this paper, we used GA-BPN to do image noise filter researching work. Firstly, this paper uses training samples to train GA-BPN as the noise detector. Then, we utilize the well-trained GA-BPN to recognize noise pixels in target image. And at last, an adaptive weighted average algorithm is used to recover noise pixels recognized by GA-BPN. Experiment data shows that this algorithm has better performance than other filters.

  11. Nature-inspired optimization algorithms

    CERN Document Server

    Yang, Xin-She

    2014-01-01

    Nature-Inspired Optimization Algorithms provides a systematic introduction to all major nature-inspired algorithms for optimization. The book's unified approach, balancing algorithm introduction, theoretical background and practical implementation, complements extensive literature with well-chosen case studies to illustrate how these algorithms work. Topics include particle swarm optimization, ant and bee algorithms, simulated annealing, cuckoo search, firefly algorithm, bat algorithm, flower algorithm, harmony search, algorithm analysis, constraint handling, hybrid methods, parameter tuning

  12. Application of Genetic Algorithms in Seismic Tomography

    Science.gov (United States)

    Soupios, Pantelis; Akca, Irfan; Mpogiatzis, Petros; Basokur, Ahmet; Papazachos, Constantinos

    2010-05-01

    In the earth sciences several inverse problems that require data fitting and parameter estimation are nonlinear and can involve a large number of unknown parameters. Consequently, the application of analytical inversion or optimization techniques may be quite restrictive. In practice, most analytical methods are local in nature and rely on a linearized form of the problem in question, adopting an iterative procedure using partial derivatives to improve an initial model. This approach can lead to a dependence of the final model solution on the starting model and is prone to entrapment in local misfit minima. Moreover, the calculation of derivatives can be computationally inefficient and create instabilities when numerical approximations are used. In contrast to these local minimization methods, global techniques that do not rely on partial derivatives, are independent of the form of the data misfit criterion, and are computationally robust. Such methods often use random processes to sample a selected wider span of the model space. In this situation, randomly generated models are assessed in terms of their data-fitting quality and the process may be stopped after a certain number of acceptable models is identified or continued until a satisfactory data fit is achieved. A new class of methods known as genetic algorithms achieves the aforementioned approximation through novel model representation and manipulations. Genetic algorithms (GAs) were originally developed in the field of artificial intelligence by John Holland more than 20 years ago, but even in this field it is less than a decade that the methodology has been more generally applied and only recently did the methodology attract the attention of the earth sciences community. Applications have been generally concentrated in geophysics and in particular seismology. As awareness of genetic algorithms grows there surely will be many more and varied applications to earth science problems. In the present work, the

  13. Novel image reconstruction algorithm for multi-phase flow tomography system using γ ray method

    International Nuclear Information System (INIS)

    Hao Kuihong; Wang Huaxiang; Gao Mei

    2007-01-01

    After analyzing the reason of image reconstructed algorithm by using the conventional back projection (IBP) is prone to produce spurious line, and considering the characteristic of multi-phase flow tomography, a novel image reconstruction algorithm is proposed, which carries out the intersection calculation using back projection data. This algorithm can obtain a perfect system point spread function, and can eliminate spurious line better. Simulating results show that the algorithm is effective for identifying multi-phase flow pattern. (authors)

  14. Assistance algorithm of nursing for amiodarone intravenous infusion

    Directory of Open Access Journals (Sweden)

    Francimar Tinoco de Oliveira

    2014-12-01

    Full Text Available This study aimed at identifying scientific publication on phlebitis caused by amiodarone and proposes a nursing care algorithm for interventions in intravenous amiodarone administration grounded in the Infusion Nursing Society and the Center for Disease Control and Prevention. It is a descriptive study mediated by integrative review in MedLine, LILACS, IBECS, BDENF, Cochrane Library and Scielo bases, published from 2006 to 2013. The sample consisted of nine articles. The evidence pointed the incidence of phlebitis due to the infusion of amiodarone and the need to control this event. The algorithm proposed shows the materials to be used and the procedure of drug administration in order to minimize injury. Besides subsidizing the development of future studies, this algorithm also promotes the incorporation of the best recommendation for the interventionist clinical practice.

  15. Convex hull ranking algorithm for multi-objective evolutionary algorithms

    NARCIS (Netherlands)

    Davoodi Monfrared, M.; Mohades, A.; Rezaei, J.

    2012-01-01

    Due to many applications of multi-objective evolutionary algorithms in real world optimization problems, several studies have been done to improve these algorithms in recent years. Since most multi-objective evolutionary algorithms are based on the non-dominated principle, and their complexity

  16. Identifying Primary Spontaneous Pneumothorax from Administrative Databases: A Validation Study

    Directory of Open Access Journals (Sweden)

    Eric Frechette

    2016-01-01

    Full Text Available Introduction. Primary spontaneous pneumothorax (PSP is a disorder commonly encountered in healthy young individuals. There is no differentiation between PSP and secondary pneumothorax (SP in the current version of the International Classification of Diseases (ICD-10. This complicates the conduct of epidemiological studies on the subject. Objective. To validate the accuracy of an algorithm that identifies cases of PSP from administrative databases. Methods. The charts of 150 patients who consulted the emergency room (ER with a recorded main diagnosis of pneumothorax were reviewed to define the type of pneumothorax that occurred. The corresponding hospital administrative data collected during previous hospitalizations and ER visits were processed through the proposed algorithm. The results were compared over two different age groups. Results. There were 144 cases of pneumothorax correctly coded (96%. The results obtained from the PSP algorithm demonstrated a significantly higher sensitivity (97% versus 81%, p=0.038 and positive predictive value (87% versus 46%, p<0.001 in patients under 40 years of age than in older patients. Conclusions. The proposed algorithm is adequate to identify cases of PSP from administrative databases in the age group classically associated with the disease. This makes possible its utilization in large population-based studies.

  17. Analysis of the Command and Control Segment (CCS) attitude estimation algorithm

    Science.gov (United States)

    Stockwell, Catherine

    1993-01-01

    This paper categorizes the qualitative behavior of the Command and Control Segment (CCS) differential correction algorithm as applied to attitude estimation using simultaneous spin axis sun angle and Earth cord length measurements. The categories of interest are the domains of convergence, divergence, and their boundaries. Three series of plots are discussed that show the dependence of the estimation algorithm on the vehicle radius, the sun/Earth angle, and the spacecraft attitude. Common qualitative dynamics to all three series are tabulated and discussed. Out-of-limits conditions for the estimation algorithm are identified and discussed.

  18. An empirical approach to selecting community-based alcohol interventions: combining research evidence, rural community views and professional opinion

    Directory of Open Access Journals (Sweden)

    Shakeshaft Anthony

    2012-01-01

    Full Text Available Abstract Background Given limited research evidence for community-based alcohol interventions, this study examines the intervention preferences of rural communities and alcohol professionals, and factors that influence their choices. Method Community preferences were identified by a survey of randomly selected individuals across 20 regional Australian communities. The preferences of alcohol professionals were identified by a survey of randomly selected members of the Australasian Professional Society on Alcohol and Other Drugs. To identify preferred interventions and the extent of support for them, a budget allocation exercise was embedded in both surveys, asking respondents to allocate a given budget to different interventions. Tobit regression models were estimated to identify the characteristics that explain differences in intervention preferences. Results Community respondents selected school programs most often (88.0% and allocated it the largest proportion of funds, followed by promotion of safer drinking (71.3%, community programs (61.4% and police enforcement of alcohol laws (60.4%. Professionals selected GP training most often (61.0% and allocated it the largest proportion of funds, followed by school programs (36.6%, community programs (33.8% and promotion of safer drinking (31.7%. Community views were susceptible to response bias. There were no significant predictors of professionals' preferences. Conclusions In the absence of sufficient research evidence for effective community-based alcohol interventions, rural communities and professionals both strongly support school programs, promotion of safer drinking and community programs. Rural communities also supported police enforcement of alcohol laws and professionals supported GP training. The impact of a combination of these strategies needs to be rigorously evaluated.

  19. Total algorithms

    NARCIS (Netherlands)

    Tel, G.

    We define the notion of total algorithms for networks of processes. A total algorithm enforces that a "decision" is taken by a subset of the processes, and that participation of all processes is required to reach this decision. Total algorithms are an important building block in the design of

  20. Assessment of available integration algorithms for initial value ordinary differential equations

    International Nuclear Information System (INIS)

    Carver, M.B.; Stewart, D.G.

    1979-11-01

    There exists an extremely large number of algorithms designed for the ordinary differential equation initial value problem. The integration is normally done by a finite sum at time intervals which are chosen dynamically to satisfy an imposed error tolerance. This report describes the basic logistics of the integration process, identifies common areas of difficulty, and establishes a comprehensive test profile for integration algorithms. A number of algorithms are described, and selected published subroutines are evaluated using the test profile. It concludes that an effective library for general use need have only two such routines. The two selected are versions of the well-known Gear and Runge-Kutta-Fehlberg algorithms. Full documentation and listings are included. (auth)