Waagepetersen, Rasmus; Ibánēz-Escriche, Noelia; Sorensen, Daniel
another extension of the linear mixed model introducing genetic random effects influencing the log residual variances of the observations thereby producing a genetically structured variance heterogeneity. Considerable computational problems arise when abandoning the standard linear mixed model. Maximum...... the various algorithms in the context of the heterogeneous variance model. Apart from being a model of great interest in its own right, this model has proven to be a hard test for MCMC methods. We compare the performances of the different algorithms when applied to three real datasets which differ markedly...... results of applying two MCMC schemes to data sets with pig litter sizes, rabbit litter sizes, and snail weights. Some concluding remarks are given in Section 5....
Mitra, Sushmita; Das, Ranajit; Hayashi, Yoichi
The analysis of gene regulatory networks provides enormous information on various fundamental cellular processes involving growth, development, hormone secretion, and cellular communication. Their extraction from available gene expression profiles is a challenging problem. Such reverse engineering of genetic networks offers insight into cellular activity toward prediction of adverse effects of new drugs or possible identification of new drug targets. Tasks such as classification, clustering, and feature selection enable efficient mining of knowledge about gene interactions in the form of networks. It is known that biological data is prone to different kinds of noise and ambiguity. Soft computing tools, such as fuzzy sets, evolutionary strategies, and neurocomputing, have been found to be helpful in providing low-cost, acceptable solutions in the presence of various types of uncertainties. In this paper, we survey the role of these soft methodologies and their hybridizations, for the purpose of generating genetic networks.
Kozlov, Nicolay N.; Kozlova, Olga N.
One of the problems in the development of mathematical theory of the genetic code (summary is presented in , the detailed -to ) is the problem of the calculation of the genetic code. Similar problems in the world is unknown and could be delivered only in the 21st century. One approach to solving this problem is devoted to this work. For the first time provides a detailed description of the method of calculation of the genetic code, the idea of which was first published earlier ), and the choice of one of the most important sets for the calculation was based on an article . Such a set of amino acid corresponds to a complete set of representations of the plurality of overlapping triple gene belonging to the same DNA strand. A separate issue was the initial point, triggering an iterative search process all codes submitted by the initial data. Mathematical analysis has shown that the said set contains some ambiguities, which have been founded because of our proposed compressed representation of the set. As a result, the developed method of calculation was limited to the two main stages of research, where the first stage only the of the area were used in the calculations. The proposed approach will significantly reduce the amount of computations at each step in this complex discrete structure.
A. L. Osipov
Full Text Available The basic statistical methods used in conducting the genetic analysis of human traits. We studied by segregation analysis, linkage analysis and allelic associations. Developed software for the implementation of these methods support.
Lahoz Beltrá, Rafael
Genetic algorithms (GAs) are a class of evolutionary algorithms inspired by Darwinian natural selection. They are popular heuristic optimisation methods based on simulated genetic mechanisms, i.e., mutation, crossover, etc. and population dynamical processes such as reproduction, selection, etc. Over the last decade, the possibility to emulate a quantum computer (a computer using quantum-mechanical phenomena to perform operations on data) has led to a new class of GAs known as “Quantum Geneti...
Full Text Available Genetic algorithms (GAs are a class of evolutionary algorithms inspired by Darwinian natural selection. They are popular heuristic optimisation methods based on simulated genetic mechanisms, i.e., mutation, crossover, etc. and population dynamical processes such as reproduction, selection, etc. Over the last decade, the possibility to emulate a quantum computer (a computer using quantum-mechanical phenomena to perform operations on data has led to a new class of GAs known as “Quantum Genetic Algorithms” (QGAs. In this review, we present a discussion, future potential, pros and cons of this new class of GAs. The review will be oriented towards computer scientists interested in QGAs “avoiding” the possible difficulties of quantum-mechanical phenomena.
Waagepetersen, Rasmus; Ibanez, Noelia; Sorensen, Daniel
Given observations of a trait and a pedigree for a group of animals, the basic model in quantitative genetics is a linear mixed model with genetic random effects. The correlation matrix of the genetic random effects is determined by the pedigree and is typically very highdimensional but with a sp......Given observations of a trait and a pedigree for a group of animals, the basic model in quantitative genetics is a linear mixed model with genetic random effects. The correlation matrix of the genetic random effects is determined by the pedigree and is typically very highdimensional...
Waagepetersen, Rasmus; Ibánez, N.; Sorensen, Daniel
Given observations of a trait and a pedigree for a group of animals, the basic model in quantitative genetics is a linear mixed model with genetic random effects. The correlation matrix of the genetic random effects is determined by the pedigree and is typically very highdimensional but with a sp...
Bryan K. Epperson; Brad H. McRae; Kim Scribner; Samuel A. Cushman; Michael S. Rosenberg; Marie-Josee Fortin; Patrick M. A. James; Melanie Murphy; Stephanie Manel; Pierre Legendre; Mark R. T. Dale
Population genetics theory is primarily based on mathematical models in which spatial complexity and temporal variability are largely ignored. In contrast, the field of landscape genetics expressly focuses on how population genetic processes are affected by complex spatial and temporal environmental heterogeneity. It is spatially explicit and relates patterns to...
Lin, Jerry; Wang, Chia-Hung; Jiang, Xin
This book gathers papers presented at the 10th International Conference on Genetic and Evolutionary Computing (ICGEC 2016). The conference was co-sponsored by Springer, Fujian University of Technology in China, the University of Computer Studies in Yangon, University of Miyazaki in Japan, National Kaohsiung University of Applied Sciences in Taiwan, Taiwan Association for Web Intelligence Consortium, and VSB-Technical University of Ostrava, Czech Republic. The ICGEC 2016, which was held from November 7 to 9, 2016 in Fuzhou City, China, was intended as an international forum for researchers and professionals in all areas of genetic and evolutionary computing.
Full Text Available The task scheduling strategy based on cultural genetic algorithm(CGA is proposed in order to improve the efficiency of task scheduling in the cloud computing platform, which targets at minimizing the total time and cost of task scheduling. The improved genetic algorithm is used to construct the main population space and knowledge space under cultural framework which get independent parallel evolution, forming a mechanism of mutual promotion to dispatch the cloud task. Simultaneously, in order to prevent the defects of the genetic algorithm which is easy to fall into local optimum, the non-uniform mutation operator is introduced to improve the search performance of the algorithm. The experimental results show that CGA reduces the total time and lowers the cost of the scheduling, which is an effective algorithm for the cloud task scheduling.
Lin, Jerry; Pan, Jeng-Shyang; Tin, Pyke; Yokota, Mitsuhiro; Genetic and Evolutionary Computing
This volume of Advances in Intelligent Systems and Computing contains accepted papers presented at ICGEC 2015, the 9th International Conference on Genetic and Evolutionary Computing. The conference this year was technically co-sponsored by Ministry of Science and Technology, Myanmar, University of Computer Studies, Yangon, University of Miyazaki in Japan, Kaohsiung University of Applied Science in Taiwan, Fujian University of Technology in China and VSB-Technical University of Ostrava. ICGEC 2015 is held from 26-28, August, 2015 in Yangon, Myanmar. Yangon, the most multiethnic and cosmopolitan city in Myanmar, is the main gateway to the country. Despite being the commercial capital of Myanmar, Yangon is a city engulfed by its rich history and culture, an integration of ancient traditions and spiritual heritage. The stunning SHWEDAGON Pagoda is the center piece of Yangon city, which itself is famous for the best British colonial era architecture. Of particular interest in many shops of Bogyoke Aung San Market,...
Knabe, Johannes F
Genetic Regulatory Networks (GRNs) in biological organisms are primary engines for cells to enact their engagements with environments, via incessant, continually active coupling. In differentiated multicellular organisms, tremendous complexity has arisen in the course of evolution of life on earth. Engineering and science have so far achieved no working system that can compare with this complexity, depth and scope of organization. Abstracting the dynamics of genetic regulatory control to a computational framework in which artificial GRNs in artificial simulated cells differentiate while connected in a changing topology, it is possible to apply Darwinian evolution in silico to study the capacity of such developmental/differentiated GRNs to evolve. In this volume an evolutionary GRN paradigm is investigated for its evolvability and robustness in models of biological clocks, in simple differentiated multicellularity, and in evolving artificial developing 'organisms' which grow and express an ontogeny starting fr...
Yang, Chin-Yu; Lin, Chun-Wei; Pan, Jeng-Shyang; Snasel, Vaclav; Abraham, Ajith
This volume of Advances in Intelligent Systems and Computing contains accepted papers presented at ICGEC 2014, the 8th International Conference on Genetic and Evolutionary Computing. The conference this year was technically co-sponsored by Nanchang Institute of Technology in China, Kaohsiung University of Applied Science in Taiwan, and VSB-Technical University of Ostrava. ICGEC 2014 is held from 18-20 October 2014 in Nanchang, China. Nanchang is one of is the capital of Jiangxi Province in southeastern China, located in the north-central portion of the province. As it is bounded on the west by the Jiuling Mountains, and on the east by Poyang Lake, it is famous for its scenery, rich history and cultural sites. Because of its central location relative to the Yangtze and Pearl River Delta regions, it is a major railroad hub in Southern China. The conference is intended as an international forum for the researchers and professionals in all areas of genetic and evolutionary computing.
Krömer, Pavel; Snášel, Václav
Genetic and Evolutionary Computing This volume of Advances in Intelligent Systems and Computing contains accepted papers presented at ICGEC 2013, the 7th International Conference on Genetic and Evolutionary Computing. The conference this year was technically co-sponsored by The Waseda University in Japan, Kaohsiung University of Applied Science in Taiwan, and VSB-Technical University of Ostrava. ICGEC 2013 was held in Prague, Czech Republic. Prague is one of the most beautiful cities in the world whose magical atmosphere has been shaped over ten centuries. Places of the greatest tourist interest are on the Royal Route running from the Powder Tower through Celetna Street to Old Town Square, then across Charles Bridge through the Lesser Town up to the Hradcany Castle. One should not miss the Jewish Town, and the National Gallery with its fine collection of Czech Gothic art, collection of old European art, and a beautiful collection of French art. The conference was intended as an international forum for the res...
Kumar, Sudhir; Stecher, Glen; Li, Michael; Knyaz, Christina; Tamura, Koichiro
The Molecular Evolutionary Genetics Analysis (Mega) software implements many analytical methods and tools for phylogenomics and phylomedicine. Here, we report a transformation of Mega to enable cross-platform use on Microsoft Windows and Linux operating systems. Mega X does not require virtualization or emulation software and provides a uniform user experience across platforms. Mega X has additionally been upgraded to use multiple computing cores for many molecular evolutionary analyses. Mega X is available in two interfaces (graphical and command line) and can be downloaded from www.megasoftware.net free of charge.
The concepts of quantum automata and quantum computation are studied in the context of quantum genetics and genetic networks with nonlinear dynamics. In previous publications (Baianu,1971a, b) the formal concept of quantum automaton and quantum computation, respectively, were introduced and their possible implications for genetic processes and metabolic activities in living cells and organisms were considered. This was followed by a report on quantum and abstract, symbolic computation based on the theory of categories, functors and natural transformations (Baianu,1971b; 1977; 1987; 2004; Baianu et al, 2004). The notions of topological semigroup, quantum automaton, or quantum computer, were then suggested with a view to their potential applications to the analogous simulation of biological systems, and especially genetic activities and nonlinear dynamics in genetic networks. Further, detailed studies of nonlinear dynamics in genetic networks were carried out in categories of n-valued, Lukasiewicz Logic Algebra...
Razuky AL-Shaikhly, Mazin H.
An essential change had occurred in the field of Information Technology which represented with cloud computing, cloud giving virtual assets by means of web yet awesome difficulties in the field of information security and security assurance. Currently main problem with cloud computing is how to improve privacy and security for cloud “cloud is critical security”. This paper attempts to solve cloud security by using intelligent system with genetic algorithm as wall to provide cloud data secure, all services provided by cloud must detect who receive and register it to create list of users (trusted or un-trusted) depend on behavior. The execution of present proposal has shown great outcome.
Buckles, Bill P.; Petry, Frederick E.
As a Guest Computational Investigator under the NASA administered component of the High Performance Computing and Communication Program, we implemented a massively parallel genetic algorithm on the MasPar SIMD computer. Experiments were conducted using Earth Science data in the domains of meteorology and oceanography. Results obtained in these domains are competitive with, and in most cases better than, similar problems solved using other methods. In the meteorological domain, we chose to identify clouds using AVHRR spectral data. Four cloud speciations were used although most researchers settle for three. Results were remarkedly consistent across all tests (91% accuracy). Refinements of this method may lead to more timely and complete information for Global Circulation Models (GCMS) that are prevalent in weather forecasting and global environment studies. In the oceanographic domain, we chose to identify ocean currents from a spectrometer having similar characteristics to AVHRR. Here the results were mixed (60% to 80% accuracy). Given that one is willing to run the experiment several times (say 10), then it is acceptable to claim the higher accuracy rating. This problem has never been successfully automated. Therefore, these results are encouraging even though less impressive than the cloud experiment. Successful conclusion of an automated ocean current detection system would impact coastal fishing, naval tactics, and the study of micro-climates. Finally we contributed to the basic knowledge of GA (genetic algorithm) behavior in parallel environments. We developed better knowledge of the use of subpopulations in the context of shared breeding pools and the migration of individuals. Rigorous experiments were conducted based on quantifiable performance criteria. While much of the work confirmed current wisdom, for the first time we were able to submit conclusive evidence. The software developed under this grant was placed in the public domain. An extensive user
Lathrop, G M; Lalouel, J M
A computer program that calculates lod scores and genetic risks for a wide variety of both qualitative and quantitative genetic traits is discussed. An illustration is given of the joint use of a genetic marker, affection status, and quantitative information in counseling situations regarding Duchenne muscular dystrophy. PMID:6585139
Moroz, Z.; Bouzyk, J.; Sowinski, M.; Chmielewski, A.G.
The use of computer genetic algorithm (GA) for driving a controller device for the industrial flue gas purification systems employing the electron beam irradiation, has been studied. As the mathematical model of the installation the properly trained artificial neural net (ANN) was used. Various cost functions and optimising strategies of the genetic code were tested. These computer simulations proved, that ANN + GA controller can be sufficiently precise and fast to be applied in real installations. (author)
Roč. 171, č. 1 (2001), s. 201-223 ISSN 0890-5401 R&D Projects: GA AV ČR IAA1019901 Institutional research plan: CEZ:AV0Z1019905; CEZ:AV0Z1019905 Keywords : complexity * genetics * croning over Subject RIV: BA - General Mathematics Impact factor: 0.571, year: 2001
Full Text Available In the Infrastructure-as-a-Service cloud computing model, virtualized computing resources in the form of virtual machines are provided over the Internet. A user can rent an arbitrary number of computing resources to meet their requirements, making cloud computing an attractive choice for executing real-time tasks. Economical task allocation and scheduling on a set of leased virtual machines is an important problem in the cloud computing environment. This paper proposes a greedy and a genetic algorithm with an adaptive selection of suitable crossover and mutation operations (named as AGA to allocate and schedule real-time tasks with precedence constraint on heterogamous virtual machines. A comprehensive simulation study has been done to evaluate the performance of the proposed algorithms in terms of their solution quality and efficiency. The simulation results show that AGA outperforms the greedy algorithm and non-adaptive genetic algorithm in terms of solution quality.
Pryor, R.J.; Cline, D.D.
A method of solving the two-phase fluid flow equations using a genetic algorithm on a NCUBE multiprocessor computer is presented. The topics discussed are the two-phase flow equations, the genetic representation of the unknowns, the fitness function, the genetic operators, and the implementation of the algorithm on the NCUBE computer. The efficiency of the implementation is investigated using a pipe blowdown problem. Effects of varying the genetic parameters and the number of processors are presented
Pryor, R.J.; Cline, D.D.
A method of solving the two-phases fluid flow equations using a genetic algorithm on a NCUBE multiprocessor computer is presented. The topics discussed are the two-phase flow equations, the genetic representation of the unkowns, the fitness function, the genetic operators, and the implementation of the algorithm on the NCUBE computer. The efficiency of the implementation is investigated using a pipe blowdown problem. Effects of varying the genetic parameters and the number of processors are presented. (orig.)
Gonzalez-Perez, Abel; Mustonen, Ville; Reva, Boris
The International Cancer Genome Consortium (ICGC) aims to catalog genomic abnormalities in tumors from 50 different cancer types. Genome sequencing reveals hundreds to thousands of somatic mutations in each tumor but only a minority of these drive tumor progression. We present the result of discu......The International Cancer Genome Consortium (ICGC) aims to catalog genomic abnormalities in tumors from 50 different cancer types. Genome sequencing reveals hundreds to thousands of somatic mutations in each tumor but only a minority of these drive tumor progression. We present the result...... of discussions within the ICGC on how to address the challenge of identifying mutations that contribute to oncogenesis, tumor maintenance or response to therapy, and recommend computational techniques to annotate somatic variants and predict their impact on cancer phenotype....
Fang, Fang; Chen, Jing; Jiang, Li-Yun; Qu, Yan-Hua; Qiao, Ge-Xia
Biological invasion is considered one of the most important global environmental problems. Knowledge of the source and dispersal routes of invasion could facilitate the eradication and control of invasive species. Soybean aphid, Aphis glycines Matsumura, is one of the most destructive soybean pests. For effective management of this pest, we conducted genetic analyses and approximate Bayesian computation (ABC) analysis to determine the origins and dispersal of the aphid species, as well as the source of its invasion in the USA, using eight microsatellite loci and the mitochondrial cytochrome c oxidase subunit I (COI) gene. We were able to identify a significant isolation by distance (IBD) pattern and three genetic lineages in the microsatellite data but not in the mtDNA dataset. The genetic structure showed that the USA population has the closest relationship with those from Korea and Japan, indicating that the two latter populations might be the sources of the invasion to the USA. Both population genetic analyses and ABC showed that the northeastern populations in China were the possible sources of the further spread of A. glycines to Indonesia. The dispersal history of this aphid can provide useful information for pest management strategies and can further help predict areas at risk of invasion. This article is protected by copyright. All rights reserved.
Benazzo, Andrea; Panziera, Alex; Bertorelle, Giorgio
Massive DNA sequencing has significantly increased the amount of data available for population genetics and molecular ecology studies. However, the parallel computation of simple statistics within and between populations from large panels of polymorphic sites is not yet available, making the exploratory analyses of a set or subset of data a very laborious task. Here, we present 4P (parallel processing of polymorphism panels), a stand-alone software program for the rapid computation of genetic variation statistics (including the joint frequency spectrum) from millions of DNA variants in multiple individuals and multiple populations. It handles a standard input file format commonly used to store DNA variation from empirical or simulation experiments. The computational performance of 4P was evaluated using large SNP (single nucleotide polymorphism) datasets from human genomes or obtained by simulations. 4P was faster or much faster than other comparable programs, and the impact of parallel computing using multicore computers or servers was evident. 4P is a useful tool for biologists who need a simple and rapid computer program to run exploratory population genetics analyses in large panels of genomic data. It is also particularly suitable to analyze multiple data sets produced in simulation studies. Unix, Windows, and MacOs versions are provided, as well as the source code for easier pipeline implementations.
Li, Jin; Wei, Zhi; Hakonarson, Hakon
Genetic factors play an important role in the etiology of inflammatory bowel disease (IBD). The launch of genome-wide association study (GWAS) represents a landmark in the genetic study of human complex disease. Concurrently, computational methods have undergone rapid development during the past a few years, which led to the identification of numerous disease susceptibility loci. IBD is one of the successful examples of GWAS and related analyses. A total of 163 genetic loci and multiple signaling pathways have been identified to be associated with IBD. Pleiotropic effects were found for many of these loci; and risk prediction models were built based on a broad spectrum of genetic variants. Important gene-gene, gene-environment interactions and key contributions of gut microbiome are being discovered. Here we will review the different types of analyses that have been applied to IBD genetic study, discuss the computational methods for each type of analysis, and summarize the discoveries made in IBD research with the application of these methods.
Zhong, Qing; Rüschoff, Jan H; Guo, Tiannan; Gabrani, Maria; Schüffler, Peter J; Rechsteiner, Markus; Liu, Yansheng; Fuchs, Thomas J; Rupp, Niels J; Fankhauser, Christian; Buhmann, Joachim M; Perner, Sven; Poyet, Cédric; Blattner, Miriam; Soldini, Davide; Moch, Holger; Rubin, Mark A; Noske, Aurelia; Rüschoff, Josef; Haffner, Michael C; Jochum, Wolfram; Wild, Peter J
Recent large-scale genome analyses of human tissue samples have uncovered a high degree of genetic alterations and tumour heterogeneity in most tumour entities, independent of morphological phenotypes and histopathological characteristics. Assessment of genetic copy-number variation (CNV) and tumour heterogeneity by fluorescence in situ hybridization (ISH) provides additional tissue morphology at single-cell resolution, but it is labour intensive with limited throughput and high inter-observer variability. We present an integrative method combining bright-field dual-colour chromogenic and silver ISH assays with an image-based computational workflow (ISHProfiler), for accurate detection of molecular signals, high-throughput evaluation of CNV, expressive visualization of multi-level heterogeneity (cellular, inter- and intra-tumour heterogeneity), and objective quantification of heterogeneous genetic deletions (PTEN) and amplifications (19q12, HER2) in diverse human tumours (prostate, endometrial, ovarian and gastric), using various tissue sizes and different scanners, with unprecedented throughput and reproducibility.
Butenko Melinka A
Full Text Available Abstract Background When generating a genetically modified organism (GMO, the primary goal is to give a target organism one or several novel traits by using biotechnology techniques. A GMO will differ from its parental strain in that its pool of transcripts will be altered. Currently, there are no methods that are reliably able to determine if an organism has been genetically altered if the nature of the modification is unknown. Results We show that the concept of computational subtraction can be used to identify transgenic cDNA sequences from genetically modified plants. Our datasets include 454-type sequences from a transgenic line of Arabidopsis thaliana and published EST datasets from commercially relevant species (rice and papaya. Conclusion We believe that computational subtraction represents a powerful new strategy for determining if an organism has been genetically modified as well as to define the nature of the modification. Fewer assumptions have to be made compared to methods currently in use and this is an advantage particularly when working with unknown GMOs.
Full Text Available Abstract Background Single nucleotide polymorphisms (SNPs and genes that exhibit presence/absence variation have provided informative marker sets for bacterial and viral genotyping. Identification of marker sets optimised for these purposes has been based on maximal generalized discriminatory power as measured by Simpson's Index of Diversity, or on the ability to identify specific variants. Here we describe the Not-N algorithm, which is designed to identify small sets of genetic markers diagnostic for user-specified subsets of known genetic variants. The algorithm does not treat the user-specified subset and the remaining genetic variants equally. Rather Not-N analysis is designed to underpin assays that provide 0% false negatives, which is very important for e.g. diagnostic procedures for clinically significant subgroups within microbial species. Results The Not-N algorithm has been incorporated into the "Minimum SNPs" computer program and used to derive genetic markers diagnostic for multilocus sequence typing-defined clonal complexes, hepatitis C virus (HCV subtypes, and phylogenetic clades defined by comparative genome hybridization (CGH data for Campylobacter jejuni, Yersinia enterocolitica and Clostridium difficile. Conclusion Not-N analysis is effective for identifying small sets of genetic markers diagnostic for microbial sub-groups. The best results to date have been obtained with CGH data from several bacterial species, and HCV sequence data.
Tomiuk, Jürgen; Guldbrandtsen, Bernt; Loeschcke, Volker
For evolutionary studies of polyploid species estimates of the genetic identity between species with different degrees of ploidy are particularly required because gene counting in samples of polyploid individuals often cannot be done, e.g., in triploids the phenotype AB can be genotypically either...... ABB or AAB. We recently suggested a genetic distance measure that is based on phenotype counting and made available the computer program POPDIST. The program provides maximum-likelihood estimates of the genetic identities and distances between polyploid populations, but this approach...
Tomazou, Marios; Barahona, Mauricio; Polizzi, Karen M; Stan, Guy-Bart
To perform well in biotechnology applications, synthetic genetic oscillators must be engineered to allow independent modulation of amplitude and period. This need is currently unmet. Here, we demonstrate computationally how two classic genetic oscillators, the dual-feedback oscillator and the repressilator, can be re-designed to provide independent control of amplitude and period and improve tunability-that is, a broad dynamic range of periods and amplitudes accessible through the input "dials." Our approach decouples frequency and amplitude modulation by incorporating an orthogonal "sink module" where the key molecular species are channeled for enzymatic degradation. This sink module maintains fast oscillation cycles while alleviating the translational coupling between the oscillator's transcription factors and output. We characterize the behavior of our re-designed oscillators over a broad range of physiologically reasonable parameters, explain why this facilitates broader function and control, and provide general design principles for building synthetic genetic oscillators that are more precisely controllable. Copyright © 2018 The Authors. Published by Elsevier Inc. All rights reserved.
Luksys, Gediminas; Fastenrath, Matthias; Coynel, David; Freytag, Virginie; Gschwind, Leo; Heck, Angela; Jessen, Frank; Maier, Wolfgang; Milnik, Annette; Riedel-Heller, Steffi G.; Scherer, Martin; Spalek, Klara; Vogler, Christian; Wagner, Michael; Wolfsgruber, Steffen; Papassotiropoulos, Andreas; de Quervain, Dominique J.-F.
Episodic memory performance is the result of distinct mental processes, such as learning, memory maintenance, and emotional modulation of memory strength. Such processes can be effectively dissociated using computational models. Here we performed gene set enrichment analyses of model parameters estimated from the episodic memory performance of 1,765 healthy young adults. We report robust and replicated associations of the amine compound SLC (solute-carrier) transporters gene set with the learning rate, of the collagen formation and transmembrane receptor protein tyrosine kinase activity gene sets with the modulation of memory strength by negative emotional arousal, and of the L1 cell adhesion molecule (L1CAM) interactions gene set with the repetition-based memory improvement. Furthermore, in a large functional MRI sample of 795 subjects we found that the association between L1CAM interactions and memory maintenance revealed large clusters of differences in brain activity in frontal cortical areas. Our findings provide converging evidence that distinct genetic profiles underlie specific mental processes of human episodic memory. They also provide empirical support to previous theoretical and neurobiological studies linking specific neuromodulators to the learning rate and linking neural cell adhesion molecules to memory maintenance. Furthermore, our study suggests additional memory-related genetic pathways, which may contribute to a better understanding of the neurobiology of human memory. PMID:26261317
Luksys, Gediminas; Fastenrath, Matthias; Coynel, David; Freytag, Virginie; Gschwind, Leo; Heck, Angela; Jessen, Frank; Maier, Wolfgang; Milnik, Annette; Riedel-Heller, Steffi G; Scherer, Martin; Spalek, Klara; Vogler, Christian; Wagner, Michael; Wolfsgruber, Steffen; Papassotiropoulos, Andreas; de Quervain, Dominique J-F
Episodic memory performance is the result of distinct mental processes, such as learning, memory maintenance, and emotional modulation of memory strength. Such processes can be effectively dissociated using computational models. Here we performed gene set enrichment analyses of model parameters estimated from the episodic memory performance of 1,765 healthy young adults. We report robust and replicated associations of the amine compound SLC (solute-carrier) transporters gene set with the learning rate, of the collagen formation and transmembrane receptor protein tyrosine kinase activity gene sets with the modulation of memory strength by negative emotional arousal, and of the L1 cell adhesion molecule (L1CAM) interactions gene set with the repetition-based memory improvement. Furthermore, in a large functional MRI sample of 795 subjects we found that the association between L1CAM interactions and memory maintenance revealed large clusters of differences in brain activity in frontal cortical areas. Our findings provide converging evidence that distinct genetic profiles underlie specific mental processes of human episodic memory. They also provide empirical support to previous theoretical and neurobiological studies linking specific neuromodulators to the learning rate and linking neural cell adhesion molecules to memory maintenance. Furthermore, our study suggests additional memory-related genetic pathways, which may contribute to a better understanding of the neurobiology of human memory.
Full Text Available Cancer therapy exerts a strong selection pressure that shapes tumor evolution, yet our knowledge of how tumors change during treatment is limited. Here, we report the analysis of cellular heterogeneity for genetic and phenotypic features and their spatial distribution in breast tumors pre- and post-neoadjuvant chemotherapy. We found that intratumor genetic diversity was tumor-subtype specific, and it did not change during treatment in tumors with partial or no response. However, lower pretreatment genetic diversity was significantly associated with pathologic complete response. In contrast, phenotypic diversity was different between pre- and posttreatment samples. We also observed significant changes in the spatial distribution of cells with distinct genetic and phenotypic features. We used these experimental data to develop a stochastic computational model to infer tumor growth patterns and evolutionary dynamics. Our results highlight the importance of integrated analysis of genotypes and phenotypes of single cells in intact tissues to predict tumor evolution.
Almendro, Vanessa; Cheng, Yu-Kang; Randles, Amanda; Itzkovitz, Shalev; Marusyk, Andriy; Ametller, Elisabet; Gonzalez-Farre, Xavier; Muñoz, Montse; Russnes, Hege G.; Helland, Åslaug; Rye, Inga H.; Borresen-Dale, Anne-Lise; Maruyama, Reo; Van Oudenaarden, Alexander; Dowsett, Mitchell; Jones, Robin L.; Reis-Filho, Jorge; Gascon, Pere; Gönen, Mithat; Michor, Franziska; Polyak, Kornelia
Cancer therapy exerts a strong selection pressure that shapes tumor evolution, yet our knowledge of how tumors change during treatment is limited. Here, we report the analysis of cellular heterogeneity for genetic and phenotypic features and their spatial distribution in breast tumors pre- and post-neoadjuvant chemotherapy. We found that intratumor genetic diversity was tumor-subtype specific, and it did not change during treatment in tumors with partial or no response. However, lower pretreatment genetic diversity was significantly associated with pathologic complete response. In contrast, phenotypic diversity was different between pre- and post-treatment samples. We also observed significant changes in the spatial distribution of cells with distinct genetic and phenotypic features. We used these experimental data to develop a stochastic computational model to infer tumor growth patterns and evolutionary dynamics. Our results highlight the importance of integrated analysis of genotypes and phenotypes of single cells in intact tissues to predict tumor evolution
Corina Teodora Bot
Full Text Available While the mouse presents an invaluable experimental model organism in biology, its usefulness in cardiac arrhythmia research is limited in some aspects due to major electrophysiological differences between murine and human action potentials (APs. As previously described, these species-specific traits can be partly overcome by application of a cell-type transforming clamp (CTC to anthropomorphize the murine cardiac AP. CTC is a hybrid experimental-computational dynamic clamp technique, in which a computationally calculated time-dependent current is inserted into a cell in real time, to compensate for the differences between sarcolemmal currents of that cell (e.g., murine and the desired species (e.g., human. For effective CTC performance, mismatch between the measured cell and a mathematical model used to mimic the measured AP must be minimal. We have developed a genetic algorithm (GA approach that rapidly tunes a mathematical model to reproduce the AP of the murine cardiac myocyte under study. Compared to a prior implementation that used a template-based model selection approach, we show that GA optimization to a cell-specific model results in a much better recapitulation of the desired AP morphology with CTC. This improvement was more pronounced when anthropomorphizing neonatal mouse cardiomyocytes to human-like APs than to guinea pig APs. CTC may be useful for a wide range of applications, from screening effects of pharmaceutical compounds on ion channel activity, to exploring variations in the mouse or human genome. Rapid GA optimization of a cell-specific mathematical model improves CTC performance and may therefore expand the applicability and usage of the CTC technique.
Azadeh, A.; Tarverdian, S.
This study presents an integrated algorithm for forecasting monthly electrical energy consumption based on genetic algorithm (GA), computer simulation and design of experiments using stochastic procedures. First, time-series model is developed as a benchmark for GA and simulation. Computer simulation is developed to generate random variables for monthly electricity consumption. This is achieved to foresee the effects of probabilistic distribution on monthly electricity consumption. The GA and simulated-based GA models are then developed by the selected time-series model. Therefore, there are four treatments to be considered in analysis of variance (ANOVA) which are actual data, time series, GA and simulated-based GA. Furthermore, ANOVA is used to test the null hypothesis of the above four alternatives being equal. If the null hypothesis is accepted, then the lowest mean absolute percentage error (MAPE) value is used to select the best model, otherwise the Duncan Multiple Range Test (DMRT) method of paired comparison is used to select the optimum model, which could be time series, GA or simulated-based GA. In case of ties the lowest MAPE value is considered as the benchmark. The integrated algorithm has several unique features. First, it is flexible and identifies the best model based on the results of ANOVA and MAPE, whereas previous studies consider the best-fit GA model based on MAPE or relative error results. Second, the proposed algorithm may identify conventional time series as the best model for future electricity consumption forecasting because of its dynamic structure, whereas previous studies assume that GA always provide the best solutions and estimation. To show the applicability and superiority of the proposed algorithm, the monthly electricity consumption in Iran from March 1994 to February 2005 (131 months) is used and applied to the proposed algorithm
Progress is reported on the following research projects: genetic effects of high LET radiations; genetic regulation, alteration, and repair; chromosome replication and the division cycle of Escherichia coli; effects of radioisotope decay in the DNA of microorganisms; initiation and termination of DNA replication in Bacillus subtilis; mutagenesis in mouse myeloma cells; lethal and mutagenic effects of near-uv radiation; effect of 8-methoxypsoralen on photodynamic lethality and mutagenicity in Escherichia coli; DNA repair of the lethal effects of far-uv; and near uv irradiation of bacterial cells
There is a need for approaches to efficiently evaluate human genetic variability and susceptibility related to environmental chemical exposure. Direct estimation of the genetic contribution to variability in susceptibility to environmental chemicals is only possible in special ca...
Susceptibility to environmental chemicals can be modulated by genetic differences. Direct estimation of the genetic contribution to variability in susceptibility to environmental chemicals is only possible in special cases where there is an observed association between exposure a...
Węgrzyn, Stefan; Vidal, Pierre
Many facts were at the origin of the present monograph. The ftrst is the beauty of maple leaves in Quebec forests in Fall. It raised the question: how does nature create and reproduce such beautiful patterns? The second was the reading of A. Lindenmayer's works on L systems. Finally came the discovery of "the secrets of DNA" together with many stimulating ex changes with biologists. Looking at such facts from the viewpoint of recursive numerical systems led to devise a simple model based on six elementary operations organized in a generating word, the analog of the program of a computer and of the genetic code of DNA in the cells of a living organism. It turned out that such a model, despite its simplicity, can account for a great number of properties of living organisms, e.g. their hierarchical structure, their ability to regenerate after a trauma, the possibility of cloning, their sensitivity to mutation, their growth, decay and reproduction. The model lends itself to analysis: the knowledge of the genera...
This paper presents the new DNAStat version 2.1 for processing genetic profile databases and biostatistical calculations. The popularization of DNA studies employed in the judicial system has led to the necessity of developing appropriate computer programs. Such programs must, above all, address two critical problems, i.e. the broadly understood data processing and data storage, and biostatistical calculations. Moreover, in case of terrorist attacks and mass natural disasters, the ability to identify victims by searching related individuals is very important. DNAStat version 2.1 is an adequate program for such purposes. The DNAStat version 1.0 was launched in 2005. In 2006, the program was updated to 1.1 and 1.2 versions. There were, however, slight differences between those versions and the original one. The DNAStat version 2.0 was launched in 2007 and the major program improvement was an introduction of the group calculation options with the potential application to personal identification of mass disasters and terrorism victims. The last 2.1 version has the option of language selection--Polish or English, which will enhance the usage and application of the program also in other countries.
Christensen, Kaare; McGue, Matt
The sequenced genomes of individuals aged ≥80 years, who were highly educated, self-referred volunteers and with no self-reported chronic diseases were compared to young controls. In these data, healthy ageing is a distinct phenotype from exceptional longevity and genetic factors that protect...
Mark P Miller
Full Text Available Identifying causal relationships in phylogeographic and landscape genetic investigations is notoriously difficult, but can be facilitated by use of multispecies comparisons.We used data visualizations to identify common spatial patterns within single lineages of four taxa inhabiting Pacific Northwest forests (northern spotted owl: Strix occidentalis caurina; red tree vole: Arborimus longicaudus; southern torrent salamander: Rhyacotriton variegatus; and western white pine: Pinus monticola. Visualizations suggested that, despite occupying the same geographical region and habitats, species responded differently to prevailing historical processes. S. o. caurina and P. monticola demonstrated directional patterns of spatial genetic structure where genetic distances and diversity were greater in southern versus northern locales. A. longicaudus and R. variegatus displayed opposite patterns where genetic distances were greater in northern versus southern regions. Statistical analyses of directional patterns subsequently confirmed observations from visualizations. Based upon regional climatological history, we hypothesized that observed latitudinal patterns may have been produced by range expansions. Subsequent computer simulations confirmed that directional patterns can be produced by expansion events.We discuss phylogeographic hypotheses regarding historical processes that may have produced observed patterns. Inferential methods used here may become increasingly powerful as detailed simulations of organisms and historical scenarios become plausible. We further suggest that inter-specific comparisons of historical patterns take place prior to drawing conclusions regarding effects of current anthropogenic change within landscapes.
Albada, A.; Dulmen, S. van; Otten, R.; Bensing, J.M.; Ausems, M.G.E.M.
This article describes the stepwise development of the website ‘E-info geneca’. The website provides counselees in breast cancer genetic counseling with computer-tailored information and a question prompt prior to their first consultation. Counselees generally do not know what to expect from genetic
Kling, Daniel; Tillmar, Andreas; Egeland, Thore; Mostad, Petter
Several applications necessitate an unbiased determination of relatedness, be it in linkage or association studies or in a forensic setting. An appropriate model to compute the joint probability of some genetic data for a set of persons given some hypothesis about the pedigree structure is then required. The increasing number of markers available through high-density SNP microarray typing and NGS technologies intensifies the demand, where using a large number of markers may lead to biased results due to strong dependencies between closely located loci, both within pedigrees (linkage) and in the population (allelic association or linkage disequilibrium (LD)). We present a new general model, based on a Markov chain for inheritance patterns and another Markov chain for founder allele patterns, the latter allowing us to account for LD. We also demonstrate a specific implementation for X chromosomal markers that allows for computation of likelihoods based on hypotheses of alleged relationships and genetic marker data. The algorithm can simultaneously account for linkage, LD, and mutations. We demonstrate its feasibility using simulated examples. The algorithm is implemented in the software FamLinkX, providing a user-friendly GUI for Windows systems (FamLinkX, as well as further usage instructions, is freely available at www.famlink.se ). Our software provides the necessary means to solve cases where no previous implementation exists. In addition, the software has the possibility to perform simulations in order to further study the impact of linkage and LD on computed likelihoods for an arbitrary set of markers.
Tsakonas, Athanasios; Dounias, Georgios; Jantzen, Jan
The paper suggests the combined use of different computational intelligence (CI) techniques in a hybrid scheme, as an effective approach to medical diagnosis. Getting to know the advantages and disadvantages of each computational intelligence technique in the recent years, the time has come...
Genetic algorithms (GAs) are computational optimisation schemes with an ... The algorithms solve optimisation problems ..... Genetic Algorithms in Search, Optimisation and Machine. Learning, Addison-Wesley Publishing Company, Inc. 1989.
Background / Question / Methods A factor limiting the rate of progress in landscape genetics has been the shortage of spatial models capable of linking life history attributes such as dispersal behavior to complex dynamic landscape features. The recent development of new models...
Pertry, I.; Nothegger, C.; Sweet, J.; Kuiper, H.A.; Davies, H.; Iserentant, D.; Hull, R.; Mezzetti, B.; Messens, K.; Loose, De M.; Oliveira, de D.; Burssens, S.; Gheysen, G.; Tzotzos, G.
Risk assessment of genetically modified organisms (GMOs) remains a contentious area and a major factor influencing the adoption of agricultural biotech. Methodologically, in many countries, risk assessment is conducted by expert committees with little or no recourse to databases and expert systems
Botta, Filippo; Eriksen, Casper; Fontaine, Michael Christophe; Guillot, Gilles
In a recent paper, Bradburd et al. (2013) proposed a model to quantify the relative effect ofgeographic and environmental distance on genetic differentiation. Here, we enhance this method in several ways. 1. We modify the covariance model so as to fit better with mainstream geostatistical models and
Smith, Mike U.
Criticizes an article by Browning and Lehman (1988) for (1) using "gene" instead of allele, (2) misusing the word "misconception," and (3) the possible influences of the computer environment on the results of the study. (PR)
Mahjani, Behrang; Toor, Salman; Nettelblad, Carl; Holmgren, Sverker
In quantitative trait locus (QTL) mapping significance of putative QTL is often determined using permutation testing. The computational needs to calculate the significance level are immense, 10 4 up to 10 8 or even more permutations can be needed. We have previously introduced the PruneDIRECT algorithm for multiple QTL scan with epistatic interactions. This algorithm has specific strengths for permutation testing. Here, we present a flexible, parallel computing framework for identifying multiple interacting QTL using the PruneDIRECT algorithm which uses the map-reduce model as implemented in Hadoop. The framework is implemented in R, a widely used software tool among geneticists. This enables users to rearrange algorithmic steps to adapt genetic models, search algorithms, and parallelization steps to their needs in a flexible way. Our work underlines the maturity of accessing distributed parallel computing for computationally demanding bioinformatics applications through building workflows within existing scientific environments. We investigate the PruneDIRECT algorithm, comparing its performance to exhaustive search and DIRECT algorithm using our framework on a public cloud resource. We find that PruneDIRECT is vastly superior for permutation testing, and perform 2 ×10 5 permutations for a 2D QTL problem in 15 hours, using 100 cloud processes. We show that our framework scales out almost linearly for a 3D QTL search.
Muhammad Rizwan Riaz
Full Text Available Cancer Genetics and Proteomics Database (CGPD is a repository for genetics and proteomics data of those Homo sapiens genes which are involved in Cancer. These genes are categorized in the database on the basis of cancer type. 72 genes of 13 types of cancers are considered in this database yet. Primers, promoters and peptides of these genes are also made available. Primers provided for each gene, with their features and conditions given to facilitate the researchers, are useful in PCR amplification, especially in cloning experiments. CGPD also contains Online Cancer Diagnostic Center (OCDC. It also contains transcription and translation tools to assist research work in progressive manner. The database is publicly available at http://www.cgpd.comyr.com.
Botta, Filippo; Eriksen, Casper; Fontaine, Michaël C.
1. In a recent paper, Bradburd et al. (Evolution, 67, 2013, 3258) proposed a model to quantify the relative effect of geographic and environmental distance on genetic differentiation. Here, we enhance this method in several ways. 2. We modify the covariance model so as to fit better with mainstre...... available as an R package called sunder. It takes as input georeferenced allele counts at the individual or population level for co-dominant markers. Program homepage: http://www2.imm.dtu.dk/~gigu/Sunder/....
Deleflie, Florent; Coulot, David; Vienne, Alain; Decosta, Romain; Richard, Pascal; Lasri, Mohammed Amjad
In order to test a preliminary orbit determination method, we fit an orbit of the geostationary satellite TELECOM-2D, as if we did not know any a priori information on its trajectory. The method is based on a genetic algorithm coupled to an analytical propagator of the trajectory, that is used over a couple of days, and that uses a whole set of altazimutal data that are acquired by the tracking network made up of the two TAROT telescopes. The adjusted orbit is then compared to a numerical reference. The method is described, and the results are analyzed, as a step towards an operational method of preliminary orbit determination for uncatalogued objects.
In order to investigate the complexity of mutations, a computational approach named Genome Evolution by Matrix Algorithms ("GEMA") has been implemented. GEMA models genomic changes, taking into account hundreds of mutations within each individual in a population. By modeling of entire human chromosomes, GEMA precisely mimics real…
Biology is increasingly data driven by virtue of the development of high-throughput technologies, such as DNA and RNA sequencing. Computational biology and bioinformatics are scientific disciplines that cross-over between the disciplines of biology, informatics and statistics; which is clearly
Othman M. K. Alsmadi
Full Text Available A robust computational technique for model order reduction (MOR of multi-time-scale discrete systems (single input single output (SISO and multi-input multioutput (MIMO is presented in this paper. This work is motivated by the singular perturbation of multi-time-scale systems where some specific dynamics may not have significant influence on the overall system behavior. The new approach is proposed using genetic algorithms (GA with the advantage of obtaining a reduced order model, maintaining the exact dominant dynamics in the reduced order, and minimizing the steady state error. The reduction process is performed by obtaining an upper triangular transformed matrix of the system state matrix defined in state space representation along with the elements of B, C, and D matrices. The GA computational procedure is based on maximizing the fitness function corresponding to the response deviation between the full and reduced order models. The proposed computational intelligence MOR method is compared to recently published work on MOR techniques where simulation results show the potential and advantages of the new approach.
Dzwinel, W.; Pepyolyshev, N.
The main goal of this paper is the presentation of our experience in development of the diagnostic system for the IBR-2 (Russia - Dubna) nuclear reactor. The authors show the principal results of the system modifications to make it work more reliable and much faster. The former needs the adaptation of new techniques of data processing, the latter, implementation of the newest computational facilities. The results of application of the clustering techniques and a method of visualization of the multi-dimensional information directly on the operator display are presented. The experiences with neural nets, used for prediction of the reactor operation, are discussed. The genetic algorithms were also tested, to reduce the quantity of data nd extracting the most informative components of the analyzed spectra. (authors)
Computational thermodynamics based on the CALPHAD approach (Thermo-Calc software) are used to design creep-resistant and affordable superalloys for large-scale applications such as power plants. Cost is reduced by the introduction of iron and by avoiding the use of expensive alloying elements such as Nb, Ta, Mo, Co etc. Strengthening is ensured by the addition of W, and of Al and Ti to provoke the precipitation of γ′. However, the addition of iron reduces the maximum possible volume fraction of γ′. The latter is maximized automatically using a genetic algorithm during simulation, while keeping the alloys free of undesirable phases at high temperatures. New superalloys with 20 wt% Cr are designed, with Fe content up to 37 wt%. They should be forgeable, weldable, oxidation resistant and significantly cheaper than existing alloys with equivalent properties. (paper)
Bang, Jeongho [Seoul National University, Seoul (Korea, Republic of); Hanyang University, Seoul (Korea, Republic of); Yoo, Seokwon [Hanyang University, Seoul (Korea, Republic of)
We propose a genetic-algorithm-based method to find the unitary transformations for any desired quantum computation. We formulate a simple genetic algorithm by introducing the 'genetic parameter vector' of the unitary transformations to be found. In the genetic algorithm process, all components of the genetic parameter vectors are supposed to evolve to the solution parameters of the unitary transformations. We apply our method to find the optimal unitary transformations and to generalize the corresponding quantum algorithms for a realistic problem, the one-bit oracle decision problem, or the often-called Deutsch problem. By numerical simulations, we can faithfully find the appropriate unitary transformations to solve the problem by using our method. We analyze the quantum algorithms identified by the found unitary transformations and generalize the variant models of the original Deutsch's algorithm.
Lippert, Christoph; Xiang, Jing; Horta, Danilo; Widmer, Christian; Kadie, Carl; Heckerman, David; Listgarten, Jennifer
Set-based variance component tests have been identified as a way to increase power in association studies by aggregating weak individual effects. However, the choice of test statistic has been largely ignored even though it may play an important role in obtaining optimal power. We compared a standard statistical test-a score test-with a recently developed likelihood ratio (LR) test. Further, when correction for hidden structure is needed, or gene-gene interactions are sought, state-of-the art algorithms for both the score and LR tests can be computationally impractical. Thus we develop new computationally efficient methods. After reviewing theoretical differences in performance between the score and LR tests, we find empirically on real data that the LR test generally has more power. In particular, on 15 of 17 real datasets, the LR test yielded at least as many associations as the score test-up to 23 more associations-whereas the score test yielded at most one more association than the LR test in the two remaining datasets. On synthetic data, we find that the LR test yielded up to 12% more associations, consistent with our results on real data, but also observe a regime of extremely small signal where the score test yielded up to 25% more associations than the LR test, consistent with theory. Finally, our computational speedups now enable (i) efficient LR testing when the background kernel is full rank, and (ii) efficient score testing when the background kernel changes with each test, as for gene-gene interaction tests. The latter yielded a factor of 2000 speedup on a cohort of size 13 500. Software available at http://research.microsoft.com/en-us/um/redmond/projects/MSCompBio/Fastlmm/. email@example.com Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press.
Yusoh, Zeratul Izzah Mohd; Tang, Maolin
Recently, Software as a Service (SaaS) in Cloud computing, has become more and more significant among software users and providers. To offer a SaaS with flexible functions at a low cost, SaaS providers have focused on the decomposition of the SaaS functionalities, or known as composite SaaS. This approach has introduced new challenges in SaaS resource management in data centres. One of the challenges is managing the resources allocated to the composite SaaS. Due to the dynamic environment of ...
Kumar, Sudhir; Stecher, Glen; Peterson, Daniel; Tamura, Koichiro
There is a growing need in the research community to apply the molecular evolutionary genetics analysis (MEGA) software tool for batch processing a large number of datasets and to integrate it into analysis workflows. Therefore, we now make available the computing core of the MEGA software as a stand-alone executable (MEGA-CC), along with an analysis prototyper (MEGA-Proto). MEGA-CC provides users with access to all the computational analyses available through MEGA's graphical user interface version. This includes methods for multiple sequence alignment, substitution model selection, evolutionary distance estimation, phylogeny inference, substitution rate and pattern estimation, tests of natural selection and ancestral sequence inference. Additionally, we have upgraded the source code for phylogenetic analysis using the maximum likelihood methods for parallel execution on multiple processors and cores. Here, we describe MEGA-CC and outline the steps for using MEGA-CC in tandem with MEGA-Proto for iterative and automated data analysis. http://www.megasoftware.net/.
Karl, Stefan; Dandekar, Thomas
Boolean networks capture switching behavior of many naturally occurring regulatory networks. For semi-quantitative modeling, interpolation between ON and OFF states is necessary. The high degree polynomial interpolation of Boolean genetic regulatory networks (GRNs) in cellular processes such as apoptosis or proliferation allows for the modeling of a wider range of node interactions than continuous activator-inhibitor models, but suffers from scaling problems for networks which contain nodes with more than ~10 inputs. Many GRNs from literature or new gene expression experiments exceed those limitations and a new approach was developed. (i) As a part of our new GRN simulation framework Jimena we introduce and setup Boolean-tree-based data structures; (ii) corresponding algorithms greatly expedite the calculation of the polynomial interpolation in almost all cases, thereby expanding the range of networks which can be simulated by this model in reasonable time. (iii) Stable states for discrete models are efficiently counted and identified using binary decision diagrams. As application example, we show how system states can now be sampled efficiently in small up to large scale hormone disease networks (Arabidopsis thaliana development and immunity, pathogen Pseudomonas syringae and modulation by cytokinins and plant hormones). Jimena simulates currently available GRNs about 10-100 times faster than the previous implementation of the polynomial interpolation model and even greater gains are achieved for large scale-free networks. This speed-up also facilitates a much more thorough sampling of continuous state spaces which may lead to the identification of new stable states. Mutants of large networks can be constructed and analyzed very quickly enabling new insights into network robustness and behavior.
Full Text Available Rapid progress in exploring the human and mouse genome has resulted in the generation of a multitude of mouse models to study gene functions in their biological context. However, effective screening methods that allow rapid noninvasive phenotyping of transgenic and knockout mice are still lacking. To identify murine models with bone alterations in vivo, we used flat-panel volume computed tomography (fpVCT for high-resolution 3-D imaging and developed an algorithm with a computational intelligence system. First, we tested the accuracy and reliability of this approach by imaging discoidin domain receptor 2- (DDR2- deficient mice, which display distinct skull abnormalities as shown by comparative landmark-based analysis. High-contrast fpVCT data of the skull with 200 microm isotropic resolution and 8-s scan time allowed segmentation and computation of significant shape features as well as visualization of morphological differences. The application of a trained artificial neuronal network to these datasets permitted a semi-automatic and highly accurate phenotype classification of DDR2-deficient compared to C57BL/6 wild-type mice. Even heterozygous DDR2 mice with only subtle phenotypic alterations were correctly determined by fpVCT imaging and identified as a new class. In addition, we successfully applied the algorithm to classify knockout mice lacking the DDR1 gene with no apparent skull deformities. Thus, this new method seems to be a potential tool to identify novel mouse phenotypes with skull changes from transgenic and knockout mice on the basis of random mutagenesis as well as from genetic models. However for this purpose, new neuronal networks have to be created and trained. In summary, the combination of fpVCT images with artificial neuronal networks provides a reliable, novel method for rapid, cost-effective, and noninvasive primary screening tool to detect skeletal phenotypes in mice.
Molnár-Gábor, Fruzsina; Lueck, Rupert; Yakneen, Sergei; Korbel, Jan O
Biomedical research is becoming increasingly large-scale and international. Cloud computing enables the comprehensive integration of genomic and clinical data, and the global sharing and collaborative processing of these data within a flexibly scalable infrastructure. Clouds offer novel research opportunities in genomics, as they facilitate cohort studies to be carried out at unprecedented scale, and they enable computer processing with superior pace and throughput, allowing researchers to address questions that could not be addressed by studies using limited cohorts. A well-developed example of such research is the Pan-Cancer Analysis of Whole Genomes project, which involves the analysis of petabyte-scale genomic datasets from research centers in different locations or countries and different jurisdictions. Aside from the tremendous opportunities, there are also concerns regarding the utilization of clouds; these concerns pertain to perceived limitations in data security and protection, and the need for due consideration of the rights of patient donors and research participants. Furthermore, the increased outsourcing of information technology impedes the ability of researchers to act within the realm of existing local regulations owing to fundamental differences in the understanding of the right to data protection in various legal systems. In this Opinion article, we address the current opportunities and limitations of cloud computing and highlight the responsible use of federated and hybrid clouds that are set up between public and private partners as an adequate solution for genetics and genomics research in Europe, and under certain conditions between Europe and international partners. This approach could represent a sensible middle ground between fragmented individual solutions and a "one-size-fits-all" approach.
Full Text Available Although various computer tools have been elaborately developed to calculate a series of statistics in molecular population genetics for both small- and large-scale DNA data, there is no efficient and easy-to-use toolkit available yet for exclusively focusing on the steps of mathematical calculation. Here, we present PopSc, a bioinformatic toolkit for calculating 45 basic statistics in molecular population genetics, which could be categorized into three classes, including (i genetic diversity of DNA sequences, (ii statistical tests for neutral evolution, and (iii measures of genetic differentiation among populations. In contrast to the existing computer tools, PopSc was designed to directly accept the intermediate metadata, such as allele frequencies, rather than the raw DNA sequences or genotyping results. PopSc is first implemented as the web-based calculator with user-friendly interface, which greatly facilitates the teaching of population genetics in class and also promotes the convenient and straightforward calculation of statistics in research. Additionally, we also provide the Python library and R package of PopSc, which can be flexibly integrated into other advanced bioinformatic packages of population genetics analysis.
Chen, Shi-Yi; Deng, Feilong; Huang, Ying; Li, Cao; Liu, Linhai; Jia, Xianbo; Lai, Song-Jia
Although various computer tools have been elaborately developed to calculate a series of statistics in molecular population genetics for both small- and large-scale DNA data, there is no efficient and easy-to-use toolkit available yet for exclusively focusing on the steps of mathematical calculation. Here, we present PopSc, a bioinformatic toolkit for calculating 45 basic statistics in molecular population genetics, which could be categorized into three classes, including (i) genetic diversity of DNA sequences, (ii) statistical tests for neutral evolution, and (iii) measures of genetic differentiation among populations. In contrast to the existing computer tools, PopSc was designed to directly accept the intermediate metadata, such as allele frequencies, rather than the raw DNA sequences or genotyping results. PopSc is first implemented as the web-based calculator with user-friendly interface, which greatly facilitates the teaching of population genetics in class and also promotes the convenient and straightforward calculation of statistics in research. Additionally, we also provide the Python library and R package of PopSc, which can be flexibly integrated into other advanced bioinformatic packages of population genetics analysis.
Eighty patients with schizophrenia and 45 medical controls were examined by cranial computed tomography (CT). The schizophrenic group showed statistically significant differences with more severe widening of the lateral ventricles, the third ventricle and Sylvius fissure, and more severe atrophy in the frontal and parietal cortex than controls. The schizophrenic cases were classified into patients with or without any hereditary trait. The former group was further subdivided into 3 groups, i.e., patients with a horizontal transmission who have sibling with schizophrenia, patients with a vertical transmission who have parents or offspring with schizophrenia, and patients with both familial traits. More severe widening of the Sylvius fissure was seen in patients with a hereditary trait than in those without. The schizophrenic group, with a hereditary trait, demonstrated significantly longer duration of hospitalization, more negative symptoms and more severe widening of the lateral ventricles and the third ventricle in comparison with the schizophrenic group without any hereditary trait. The degree of the widening of the lateral ventricles seen in the subgroup with a horizontal transmission did not correlate with any other findings in CT. This means that the change is specific to this group. No significant correlation was seen between such clinical items as age and duration of illness, and widening of the lateral ventricles in the subgroup with a horizontal transmission. All this evidence implies that more advanced widening of the lateral ventricles and more serious negative symptoms exist constantly and not progressively at the point of onset in the subgroup with a horizontal transmission. (J.P.N.)
By 2030, it is predicted that over 14 million people will die of heart disease annually, many of whom will discover their risk when it is too late to seek effective treatment or pursue lifestyle changes. In this research study, I sought to design a robust computational platform to gauge a patient's risk for cardiac diseases (CDs) based on demographics, genotype, and cardiac action potentials through machine learning, statistical analysis, and vectorcardiography. By analyzing previously published data, I discovered that certain polymorphisms in the ACE and MTHFR genes contribute significantly to CD risk. The deletion allele of the ACE insertion/deletion polymorphism increases ACE serum levels, promoting CD phenotypes. A point mutation in the MTHFR gene curbs the metabolism of folic acid, giving rise to CD phenotypes. I analyzed over 9000 British Medical Journal and American Heart Association patients to determine the CD risk associated with each ACE and MTHFR genotype. In the vectorcardiography phase of my study, I investigated trends in the maximal vectors of the QRS loop of the cardiac wave. Using a database with both normal and diseased vectorcardiographic action potentials, I plotted the maximal vectors on a 3D RAS coordinate plane to analyze their magnitude and direction. From the ACE datasets, I discovered that female patients over 45 and of Indian descent with two ACE deletion alleles exhibited the highest CD risk. Using this spectrum, I successfully constructed a neural network with an accuracy score of 0.867 that predicts CD risk based on ACE genotype, gender, region, and age. Investigation of the MTHFR genome showed that those with a homozygous mutated gene had a significantly higher CD risk. In my vectorcardiography study, I found that healthy QRS vectors pointed predominantly to the right-anterior region of the coordinate plane and exhibited short, consistent magnitudes. On the other hand, diseased vectors pointed to the left-posterior region and
Overview In autumn the main focus was to process and handle CRAFT data and to perform the Summer08 MC production. The operational aspects were well covered by regular Computing Shifts, experts on duty and Computing Run Coordination. At the Computing Resource Board (CRB) in October a model to account for service work at Tier 2s was approved. The computing resources for 2009 were reviewed for presentation at the C-RRB. The quarterly resource monitoring is continuing. Facilities/Infrastructure operations Operations during CRAFT data taking ran fine. This proved to be a very valuable experience for T0 workflows and operations. The transfers of custodial data to most T1s went smoothly. A first round of reprocessing started at the Tier-1 centers end of November; it will take about two weeks. The Computing Shifts procedure was tested full scale during this period and proved to be very efficient: 30 Computing Shifts Persons (CSP) and 10 Computing Resources Coordinators (CRC). The shift program for the shut down w...
Introduction CMS distributed computing system performed well during the 2011 start-up. The events in 2011 have more pile-up and are more complex than last year; this results in longer reconstruction times and harder events to simulate. Significant increases in computing capacity were delivered in April for all computing tiers, and the utilisation and load is close to the planning predictions. All computing centre tiers performed their expected functionalities. Heavy-Ion Programme The CMS Heavy-Ion Programme had a very strong showing at the Quark Matter conference. A large number of analyses were shown. The dedicated heavy-ion reconstruction facility at the Vanderbilt Tier-2 is still involved in some commissioning activities, but is available for processing and analysis. Facilities and Infrastructure Operations Facility and Infrastructure operations have been active with operations and several important deployment tasks. Facilities participated in the testing and deployment of WMAgent and WorkQueue+Request...
The Computing Project is preparing for a busy year where the primary emphasis of the project moves towards steady operations. Following the very successful completion of Computing Software and Analysis challenge, CSA06, last fall, we have reorganized and established four groups in computing area: Commissioning, User Support, Facility/Infrastructure Operations and Data Operations. These groups work closely together with groups from the Offline Project in planning for data processing and operations. Monte Carlo production has continued since CSA06, with about 30M events produced each month to be used for HLT studies and physics validation. Monte Carlo production will continue throughout the year in the preparation of large samples for physics and detector studies ramping to 50 M events/month for CSA07. Commissioning of the full CMS computing system is a major goal for 2007. Site monitoring is an important commissioning component and work is ongoing to devise CMS specific tests to be included in Service Availa...
Overview During the past three months activities were focused on data operations, testing and re-enforcing shift and operational procedures for data production and transfer, MC production and on user support. Planning of the computing resources in view of the new LHC calendar in ongoing. Two new task forces were created for supporting the integration work: Site Commissioning, which develops tools helping distributed sites to monitor job and data workflows, and Analysis Support, collecting the user experience and feedback during analysis activities and developing tools to increase efficiency. The development plan for DMWM for 2009/2011 was developed at the beginning of the year, based on the requirements from the Physics, Computing and Offline groups (see Offline section). The Computing management meeting at FermiLab on February 19th and 20th was an excellent opportunity discussing the impact and for addressing issues and solutions to the main challenges facing CMS computing. The lack of manpower is particul...
Mazurowski, Maciej A; Habas, Piotr A; Zurada, Jacek M; Tourassi, Georgia D
This paper presents an optimization framework for improving case-based computer-aided decision (CB-CAD) systems. The underlying hypothesis of the study is that each example in the knowledge database of a medical decision support system has different importance in the decision making process. A new decision algorithm incorporating an importance weight for each example is proposed to account for these differences. The search for the best set of importance weights is defined as an optimization problem and a genetic algorithm is employed to solve it. The optimization process is tailored to maximize the system's performance according to clinically relevant evaluation criteria. The study was performed using a CAD system developed for the classification of regions of interests (ROIs) in mammograms as depicting masses or normal tissue. The system was constructed and evaluated using a dataset of ROIs extracted from the Digital Database for Screening Mammography (DDSM). Experimental results show that, according to receiver operator characteristic (ROC) analysis, the proposed method significantly improves the overall performance of the CAD system as well as its average specificity for high breast mass detection rates
Computing activity had ramped down after the completion of the reprocessing of the 2012 data and parked data, but is increasing with new simulation samples for analysis and upgrade studies. Much of the Computing effort is currently involved in activities to improve the computing system in preparation for 2015. Operations Office Since the beginning of 2013, the Computing Operations team successfully re-processed the 2012 data in record time, not only by using opportunistic resources like the San Diego Supercomputer Center which was accessible, to re-process the primary datasets HTMHT and MultiJet in Run2012D much earlier than planned. The Heavy-Ion data-taking period was successfully concluded in February collecting almost 500 T. Figure 3: Number of events per month (data) In LS1, our emphasis is to increase efficiency and flexibility of the infrastructure and operation. Computing Operations is working on separating disk and tape at the Tier-1 sites and the full implementation of the xrootd federation ...
Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion An activity that is still in progress is computing for the heavy-ion program. The heavy-ion events are collected without zero suppression, so the event size is much large at roughly 11 MB per event of RAW. The central collisions are more complex and...
M. Kasemann P. McBride Edited by M-C. Sawley with contributions from: P. Kreuzer D. Bonacorsi S. Belforte F. Wuerthwein L. Bauerdick K. Lassila-Perini M-C. Sawley
Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the comput...
It has been a very active year for the computing project with strong contributions from members of the global community. The project has focused on site preparation and Monte Carlo production. The operations group has begun processing data from P5 as part of the global data commissioning. Improvements in transfer rates and site availability have been seen as computing sites across the globe prepare for large scale production and analysis as part of CSA07. Preparations for the upcoming Computing Software and Analysis Challenge CSA07 are progressing. Ian Fisk and Neil Geddes have been appointed as coordinators for the challenge. CSA07 will include production tests of the Tier-0 production system, reprocessing at the Tier-1 sites and Monte Carlo production at the Tier-2 sites. At the same time there will be a large analysis exercise at the Tier-2 centres. Pre-production simulation of the Monte Carlo events for the challenge is beginning. Scale tests of the Tier-0 will begin in mid-July and the challenge it...
Introduction During the past six months, Computing participated in the STEP09 exercise, had a major involvement in the October exercise and has been working with CMS sites on improving open issues relevant for data taking. At the same time operations for MC production, real data reconstruction and re-reconstructions and data transfers at large scales were performed. STEP09 was successfully conducted in June as a joint exercise with ATLAS and the other experiments. It gave good indication about the readiness of the WLCG infrastructure with the two major LHC experiments stressing the reading, writing and processing of physics data. The October Exercise, in contrast, was conducted as an all-CMS exercise, where Physics, Computing and Offline worked on a common plan to exercise all steps to efficiently access and analyze data. As one of the major results, the CMS Tier-2s demonstrated to be fully capable for performing data analysis. In recent weeks, efforts were devoted to CMS Computing readiness. All th...
Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion The Tier 0 infrastructure was able to repack and promptly reconstruct heavy-ion collision data. Two copies were made of the data at CERN using a large CASTOR disk pool, and the core physics sample was replicated ...
Introduction Computing continued with a high level of activity over the winter in preparation for conferences and the start of the 2012 run. 2012 brings new challenges with a new energy, more complex events, and the need to make the best use of the available time before the Long Shutdown. We expect to be resource constrained on all tiers of the computing system in 2012 and are working to ensure the high-priority goals of CMS are not impacted. Heavy ions After a successful 2011 heavy-ion run, the programme is moving to analysis. During the run, the CAF resources were well used for prompt analysis. Since then in 2012 on average 200 job slots have been used continuously at Vanderbilt for analysis workflows. Operations Office As of 2012, the Computing Project emphasis has moved from commissioning to operation of the various systems. This is reflected in the new organisation structure where the Facilities and Data Operations tasks have been merged into a common Operations Office, which now covers everything ...
CCRC’08 challenges and CSA08 During the February campaign of the Common Computing readiness challenges (CCRC’08), the CMS computing team had achieved very good results. The link between the detector site and the Tier0 was tested by gradually increasing the number of parallel transfer streams well beyond the target. Tests covered the global robustness at the Tier0, processing a massive number of very large files and with a high writing speed to tapes. Other tests covered the links between the different Tiers of the distributed infrastructure and the pre-staging and reprocessing capacity of the Tier1’s: response time, data transfer rate and success rate for Tape to Buffer staging of files kept exclusively on Tape were measured. In all cases, coordination with the sites was efficient and no serious problem was found. These successful preparations prepared the ground for the second phase of the CCRC’08 campaign, in May. The Computing Software and Analysis challen...
Introduction The first data taking period of November produced a first scientific paper, and this is a very satisfactory step for Computing. It also gave the invaluable opportunity to learn and debrief from this first, intense period, and make the necessary adaptations. The alarm procedures between different groups (DAQ, Physics, T0 processing, Alignment/calibration, T1 and T2 communications) have been reinforced. A major effort has also been invested into remodeling and optimizing operator tasks in all activities in Computing, in parallel with the recruitment of new Cat A operators. The teams are being completed and by mid year the new tasks will have been assigned. CRB (Computing Resource Board) The Board met twice since last CMS week. In December it reviewed the experience of the November data-taking period and could measure the positive improvements made for the site readiness. It also reviewed the policy under which Tier-2 are associated with Physics Groups. Such associations are decided twice per ye...
Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the co...
Introduction Just two months after the “LHC First Physics” event of 30th March, the analysis of the O(200) million 7 TeV collision events in CMS accumulated during the first 60 days is well under way. The consistency of the CMS computing model has been confirmed during these first weeks of data taking. This model is based on a hierarchy of use-cases deployed between the different tiers and, in particular, the distribution of RECO data to T1s, who then serve data on request to T2s, along a topology known as “fat tree”. Indeed, during this period this model was further extended by almost full “mesh” commissioning, meaning that RECO data were shipped to T2s whenever possible, enabling additional physics analyses compared with the “fat tree” model. Computing activities at the CMS Analysis Facility (CAF) have been marked by a good time response for a load almost evenly shared between ALCA (Alignment and Calibration tasks - highest p...
Contributions from I. Fisk
Introduction The start of the 2012 run has been busy for Computing. We have reconstructed, archived, and served a larger sample of new data than in 2011, and we are in the process of producing an even larger new sample of simulations at 8 TeV. The running conditions and system performance are largely what was anticipated in the plan, thanks to the hard work and preparation of many people. Heavy ions Heavy Ions has been actively analysing data and preparing for conferences. Operations Office Figure 6: Transfers from all sites in the last 90 days For ICHEP and the Upgrade efforts, we needed to produce and process record amounts of MC samples while supporting the very successful data-taking. This was a large burden, especially on the team members. Nevertheless the last three months were very successful and the total output was phenomenal, thanks to our dedicated site admins who keep the sites operational and the computing project members who spend countless hours nursing the...
Introduction A large fraction of the effort was focused during the last period into the preparation and monitoring of the February tests of Common VO Computing Readiness Challenge 08. CCRC08 is being run by the WLCG collaboration in two phases, between the centres and all experiments. The February test is dedicated to functionality tests, while the May challenge will consist of running at all centres and with full workflows. For this first period, a number of functionality checks of the computing power, data repositories and archives as well as network links are planned. This will help assess the reliability of the systems under a variety of loads, and identifying possible bottlenecks. Many tests are scheduled together with other VOs, allowing the full scale stress test. The data rates (writing, accessing and transfer¬ring) are being checked under a variety of loads and operating conditions, as well as the reliability and transfer rates of the links between Tier-0 and Tier-1s. In addition, the capa...
Overview The main focus during the summer was to handle data coming from the detector and to perform Monte Carlo production. The lessons learned during the CCRC and CSA08 challenges in May were addressed by dedicated PADA campaigns lead by the Integration team. Big improvements were achieved in the stability and reliability of the CMS Tier1 and Tier2 centres by regular and systematic follow-up of faults and errors with the help of the Savannah bug tracking system. In preparation for data taking the roles of a Computing Run Coordinator and regular computing shifts monitoring the services and infrastructure as well as interfacing to the data operations tasks are being defined. The shift plan until the end of 2008 is being put together. User support worked on documentation and organized several training sessions. The ECoM task force delivered the report on “Use Cases for Start-up of pp Data-Taking” with recommendations and a set of tests to be performed for trigger rates much higher than the ...
The Computing Software and Analysis Challenge CSA07 has been the main focus of the Computing Project for the past few months. Activities began over the summer with the preparation of the Monte Carlo data sets for the challenge and tests of the new production system at the Tier-0 at CERN. The pre-challenge Monte Carlo production was done in several steps: physics generation, detector simulation, digitization, conversion to RAW format and the samples were run through the High Level Trigger (HLT). The data was then merged into three "Soups": Chowder (ALPGEN), Stew (Filtered Pythia) and Gumbo (Pythia). The challenge officially started when the first Chowder events were reconstructed on the Tier-0 on October 3rd. The data operations teams were very busy during the the challenge period. The MC production teams continued with signal production and processing while the Tier-0 and Tier-1 teams worked on splitting the Soups into Primary Data Sets (PDS), reconstruction and skimming. The storage sys...
Computing operation has been lower as the Run 1 samples are completing and smaller samples for upgrades and preparations are ramping up. Much of the computing activity is focusing on preparations for Run 2 and improvements in data access and flexibility of using resources. Operations Office Data processing was slow in the second half of 2013 with only the legacy re-reconstruction pass of 2011 data being processed at the sites. Figure 1: MC production and processing was more in demand with a peak of over 750 Million GEN-SIM events in a single month. Figure 2: The transfer system worked reliably and efficiently and transferred on average close to 520 TB per week with peaks at close to 1.2 PB. Figure 3: The volume of data moved between CMS sites in the last six months The tape utilisation was a focus for the operation teams with frequent deletion campaigns from deprecated 7 TeV MC GEN-SIM samples to INVALID datasets, which could be cleaned up...
Introduction Computing activity has been running at a sustained, high rate as we collect data at high luminosity, process simulation, and begin to process the parked data. The system is functional, though a number of improvements are planned during LS1. Many of the changes will impact users, we hope only in positive ways. We are trying to improve the distributed analysis tools as well as the ability to access more data samples more transparently. Operations Office Figure 2: Number of events per month, for 2012 Since the June CMS Week, Computing Operations teams successfully completed data re-reconstruction passes and finished the CMSSW_53X MC campaign with over three billion events available in AOD format. Recorded data was successfully processed in parallel, exceeding 1.2 billion raw physics events per month for the first time in October 2012 due to the increase in data-parking rate. In parallel, large efforts were dedicated to WMAgent development and integrati...
Introduction The Computing Team successfully completed the storage, initial processing, and distribution for analysis of proton-proton data in 2011. There are still a variety of activities ongoing to support winter conference activities and preparations for 2012. Heavy ions The heavy-ion run for 2011 started in early November and has already demonstrated good machine performance and success of some of the more advanced workflows planned for 2011. Data collection will continue until early December. Facilities and Infrastructure Operations Operational and deployment support for WMAgent and WorkQueue+Request Manager components, routinely used in production by Data Operations, are provided. The GlideInWMS and components installation are now deployed at CERN, which is added to the GlideInWMS factory placed in the US. There has been new operational collaboration between the CERN team and the UCSD GlideIn factory operators, covering each others time zones by monitoring/debugging pilot jobs sent from the facto...
Sahiner, Berkman; Chan, Heang-Ping; Petrick, Nicholas; Helvie, Mark A.; Goodsitt, Mitchell M.
A genetic algorithm (GA) based feature selection method was developed for the design of high-sensitivity classifiers, which were tailored to yield high sensitivity with high specificity. The fitness function of the GA was based on the receiver operating characteristic (ROC) partial area index, which is defined as the average specificity above a given sensitivity threshold. The designed GA evolved towards the selection of feature combinations which yielded high specificity in the high-sensitivity region of the ROC curve, regardless of the performance at low sensitivity. This is a desirable quality of a classifier used for breast lesion characterization, since the focus in breast lesion characterization is to diagnose correctly as many benign lesions as possible without missing malignancies. The high-sensitivity classifier, formulated as the Fisher's linear discriminant using GA-selected feature variables, was employed to classify 255 biopsy-proven mammographic masses as malignant or benign. The mammograms were digitized at a pixel size of 0.1mmx0.1mm, and regions of interest (ROIs) containing the biopsied masses were extracted by an experienced radiologist. A recently developed image transformation technique, referred to as the rubber-band straightening transform, was applied to the ROIs. Texture features extracted from the spatial grey-level dependence and run-length statistics matrices of the transformed ROIs were used to distinguish malignant and benign masses. The classification accuracy of the high-sensitivity classifier was compared with that of linear discriminant analysis with stepwise feature selection (LDA sfs ). With proper GA training, the ROC partial area of the high-sensitivity classifier above a true-positive fraction of 0.95 was significantly larger than that of LDA sfs , although the latter provided a higher total area (A z ) under the ROC curve. By setting an appropriate decision threshold, the high-sensitivity classifier and LDA sfs correctly
CMS relies on a well functioning, distributed computing infrastructure. The Site Availability Monitoring (SAM) and the Job Robot submission have been very instrumental for site commissioning in order to increase availability of more sites such that they are available to participate in CSA07 and are ready to be used for analysis. The commissioning process has been further developed, including "lessons learned" documentation via the CMS twiki. Recently the visualization, presentation and summarizing of SAM tests for sites has been redesigned, it is now developed by the central ARDA project of WLCG. Work to test the new gLite Workload Management System was performed; a 4 times increase in throughput with respect to LCG Resource Broker is observed. CMS has designed and launched a new-generation traffic load generator called "LoadTest" to commission and to keep exercised all data transfer routes in the CMS PhE-DEx topology. Since mid-February, a transfer volume of about 12 P...
Lin, J.; Bartal, Y.; Uhrig, R.E.
The importance of automatic diagnostic systems for nuclear power plants (NPPs) has been discussed in numerous studies, and various such systems have been proposed. None of those systems were designed to predict the severity of the diagnosed scenario. A classification and severity prediction system for NPP transients is developed. The system is based on nearest neighbors modeling, which is optimized using genetic algorithms. The optimization process is used to determine the most important variables for each of the transient types analyzed. An enhanced version of the genetic algorithms is used in which a local downhill search is performed to further increase the accuracy achieved. The genetic algorithms search was implemented on a massively parallel supercomputer, the KSR1-64, to perform the analysis in a reasonable time. The data for this study were supplied by the high-fidelity simulator of the San Onofre unit 1 pressurized water reactor
Campbell, M. T.; Walia, H.; Grondin, A.; Knecht, A.
The development of abiotic stress tolerant crops (i.e. drought, salinity, or heat stress) requires the discovery of DNA sequence variants associated with stress tolerance-related traits. However, many traits underlying adaptation to abiotic stress involve a suite of physiological pathways that may be induced at different times throughout the duration of stress. Conventional single-point phenotyping approaches fail to fully capture these temporal responses, and thus downstream genetic analysis may only identify a subset of the genetic variants that are important for adaptation to sub-optimal environments. Although genomic resources for crops have advanced tremendously, the collection of phenotypic data for morphological and physiological traits is laborious and remains a significant bottleneck in bridging the phenotype-genotype gap. In recent years, the availability of automated, image-based phenotyping platforms has provided researchers with an opportunity to collect morphological and physiological traits non-destructively in a highly controlled environment. Moreover, these platforms allow abiotic stress responses to be recorded throughout the duration of the experiment, and have facilitated the use of function-valued traits for genetic analyses in major crops. We will present our approaches for addressing abiotic stress tolerance in cereals. This talk will focus on novel open-source software to process and extract biological meaningful data from images generated from these phenomics platforms. In addition, we will discuss the statistical approaches to model longitudinal phenotypes and dissect the genetic basis of dynamic responses to these abiotic stresses throughout development.
Braak, ter C.J.F.
Differential Evolution (DE) is a simple genetic algorithm for numerical optimization in real parameter spaces. In a statistical context one would not just want the optimum but also its uncertainty. The uncertainty distribution can be obtained by a Bayesian analysis (after specifying prior and
Browning, Mark; Lehman, James D.
Authors respond to criticisms by Smith in the same issue and defend their use of the term "gene" and "misconception." Authors indicate that they did not believe that the use of computers significantly skewed their data concerning student errors. (PR)
Trezza, Alfonso; Bernini, Andrea; Langella, Andrea; Ascher, David B; Pires, Douglas E V; Sodi, Andrea; Passerini, Ilaria; Pelo, Elisabetta; Rizzo, Stanislao; Niccolai, Neri; Spiga, Ottavia
The aim of this article is to report the investigation of the structural features of ABCA4, a protein associated with a genetic retinal disease. A new database collecting knowledge of ABCA4 structure may facilitate predictions about the possible functional consequences of gene mutations observed in clinical practice. In order to correlate structural and functional effects of the observed mutations, the structure of mouse P-glycoprotein was used as a template for homology modeling. The obtained structural information and genetic data are the basis of our relational database (ABCA4Database). Sequence variability among all ABCA4-deposited entries was calculated and reported as Shannon entropy score at the residue level. The three-dimensional model of ABCA4 structure was used to locate the spatial distribution of the observed variable regions. Our predictions from structural in silico tools were able to accurately link the functional effects of mutations to phenotype. The development of the ABCA4Database gathers all the available genetic and structural information, yielding a global view of the molecular basis of some retinal diseases. ABCA4 modeled structure provides a molecular basis on which to analyze protein sequence mutations related to genetic retinal disease in order to predict the risk of retinal disease across all possible ABCA4 mutations. Additionally, our ABCA4 predicted structure is a good starting point for the creation of a new data analysis model, appropriate for precision medicine, in order to develop a deeper knowledge network of the disease and to improve the management of patients.
Julie B. Fiévet
Full Text Available Heterosis, the superiority of hybrids over their parents for quantitative traits, represents a crucial issue in plant and animal breeding as well as evolutionary biology. Heterosis has given rise to countless genetic, genomic and molecular studies, but has rarely been investigated from the point of view of systems biology. We hypothesized that heterosis is an emergent property of living systems resulting from frequent concave relationships between genotypic variables and phenotypes, or between different phenotypic levels. We chose the enzyme-flux relationship as a model of the concave genotype-phenotype (GP relationship, and showed that heterosis can be easily created in the laboratory. First, we reconstituted in vitro the upper part of glycolysis. We simulated genetic variability of enzyme activity by varying enzyme concentrations in test tubes. Mixing the content of “parental” tubes resulted in “hybrids,” whose fluxes were compared to the parental fluxes. Frequent heterotic fluxes were observed, under conditions that were determined analytically and confirmed by computer simulation. Second, to test this model in a more realistic situation, we modeled the glycolysis/fermentation network in yeast by considering one input flux, glucose, and two output fluxes, glycerol and acetaldehyde. We simulated genetic variability by randomly drawing parental enzyme concentrations under various conditions, and computed the parental and hybrid fluxes using a system of differential equations. Again we found that a majority of hybrids exhibited positive heterosis for metabolic fluxes. Cases of negative heterosis were due to local convexity between certain enzyme concentrations and fluxes. In both approaches, heterosis was maximized when the parents were phenotypically close and when the distributions of parental enzyme concentrations were contrasted and constrained. These conclusions are not restricted to metabolic systems: they only depend on the
Kooths, Stefan; Mitze, Timo Friedel; Ringhut, Eric
This paper compares the predictive power of linear econometric and non-linear computational models for forecasting the inflation rate in the European Monetary Union (EMU). Various models of both types are developed using different monetary and real activity indicators. They are compared according...
Yamamoto, Akio; Hashimoto, Hiroshi
The distributed genetic algorithm (DGA) is applied for loading pattern optimization problems of the pressurized water reactors. A basic concept of DGA follows that of the conventional genetic algorithm (GA). However, DGA equally distributes candidates of solutions (i.e. loading patterns) to several independent ''islands'' and evolves them in each island. Communications between islands, i.e. migrations of some candidates between islands are performed with a certain period. Since candidates of solutions independently evolve in each island while accepting different genes of migrants, premature convergence in the conventional GA can be prevented. Because many candidate loading patterns should be evaluated in GA or DGA, the parallelization is efficient to reduce turn around time. Parallel efficiency of DGA was measured using our optimization code and good efficiency was attained even in a heterogeneous cluster environment due to dynamic distribution of the calculation load. The optimization code is based on the client/server architecture with the TCP/IP native socket and a client (optimization) module and calculation server modules communicate the objects of loading patterns each other. Throughout the sensitivity study on optimization parameters of DGA, a suitable set of the parameters for a test problem was identified. Finally, optimization capability of DGA and the conventional GA was compared in the test problem and DGA provided better optimization results than the conventional GA. (author)
Lewis, Benjamin C; Nair, Pramod C; Heran, Subash S; Somogyi, Andrew A; Bowden, Jeffrey J; Doogue, Matthew P; Miners, John O
The variable response to warfarin treatment often has a genetic basis. A protein homology model of human vitamin K epoxide reductase, subunit 1 (VKORC1), was generated to elucidate the mechanism of warfarin resistance observed in a patient with the Val66Met mutation. The VKORC1 homology model comprises four transmembrane (TM) helical domains and a half helical lid domain. Cys132 and Cys135, located in the N-terminal end of TM-4, are linked through a disulfide bond. Two distinct binding sites for warfarin were identified. Site-1, which binds vitamin K epoxide (KO) in a catalytically favorable orientation, shows higher affinity for S-warfarin compared with R-warfarin. Site-2, positioned in the domain occupied by the hydrophobic tail of KO, binds both warfarin enantiomers with similar affinity. Displacement of Arg37 occurs in the Val66Met mutant, blocking access of warfarin (but not KO) to Site-1, consistent with clinical observation of warfarin resistance.
Scott, W.A.; Turner, R.M.; McCammon, R.B.
Integrated logic circuits were described as a means of formally representing genetic-geologic models for estimating undiscovered uranium resources. The logic circuits are logical combinations of selected geologic characteristics judged to be associated with particular types of uranium deposits. Each combination takes on a value which corresponds to the combined presence, absence, or don't know states of the selected characteristic within a specified geographic cell. Within each cell, the output of the logic circuit is taken as a measure of the favorability of occurrence of an undiscovered deposit of the type being considered. In this way, geological, geochemical, and geophysical data are incorporated explicitly into potential uranium resource estimates. The present report describes how integrated logic circuits are constructed by use of a computer graphics program. A user's guide is also included
Mih, Nathan; Brunk, Elizabeth; Bordbar, Aarash; Palsson, Bernhard O
Progress in systems medicine brings promise to addressing patient heterogeneity and individualized therapies. Recently, genome-scale models of metabolism have been shown to provide insight into the mechanistic link between drug therapies and systems-level off-target effects while being expanded to explicitly include the three-dimensional structure of proteins. The integration of these molecular-level details, such as the physical, structural, and dynamical properties of proteins, notably expands the computational description of biochemical network-level properties and the possibility of understanding and predicting whole cell phenotypes. In this study, we present a multi-scale modeling framework that describes biological processes which range in scale from atomistic details to an entire metabolic network. Using this approach, we can understand how genetic variation, which impacts the structure and reactivity of a protein, influences both native and drug-induced metabolic states. As a proof-of-concept, we study three enzymes (catechol-O-methyltransferase, glucose-6-phosphate dehydrogenase, and glyceraldehyde-3-phosphate dehydrogenase) and their respective genetic variants which have clinically relevant associations. Using all-atom molecular dynamic simulations enables the sampling of long timescale conformational dynamics of the proteins (and their mutant variants) in complex with their respective native metabolites or drug molecules. We find that changes in a protein's structure due to a mutation influences protein binding affinity to metabolites and/or drug molecules, and inflicts large-scale changes in metabolism.
Full Text Available Progress in systems medicine brings promise to addressing patient heterogeneity and individualized therapies. Recently, genome-scale models of metabolism have been shown to provide insight into the mechanistic link between drug therapies and systems-level off-target effects while being expanded to explicitly include the three-dimensional structure of proteins. The integration of these molecular-level details, such as the physical, structural, and dynamical properties of proteins, notably expands the computational description of biochemical network-level properties and the possibility of understanding and predicting whole cell phenotypes. In this study, we present a multi-scale modeling framework that describes biological processes which range in scale from atomistic details to an entire metabolic network. Using this approach, we can understand how genetic variation, which impacts the structure and reactivity of a protein, influences both native and drug-induced metabolic states. As a proof-of-concept, we study three enzymes (catechol-O-methyltransferase, glucose-6-phosphate dehydrogenase, and glyceraldehyde-3-phosphate dehydrogenase and their respective genetic variants which have clinically relevant associations. Using all-atom molecular dynamic simulations enables the sampling of long timescale conformational dynamics of the proteins (and their mutant variants in complex with their respective native metabolites or drug molecules. We find that changes in a protein's structure due to a mutation influences protein binding affinity to metabolites and/or drug molecules, and inflicts large-scale changes in metabolism.
Gason, Alexandra A; Aitken, MaryAnne; Delatycki, Martin B; Sheffield, Edith; Metcalfe, Sylvia A
Tay Sachs disease is a recessively inherited neurodegenerative disorder, for which carrier screening programs exist worldwide. Education for those offered a screening test is essential in facilitating informed decision-making. In Melbourne, Australia, we have designed, developed, and evaluated a computer-based instructional resource for use in the Tay Sachs disease carrier screening program for secondary school students attending Jewish schools. The resource entitled "Genetics in the Community: Tay Sachs disease" was designed on a platform of educational learning theory. The development of the resource included formative evaluation using qualitative data analysis supported by descriptive quantitative data. The final resource was evaluated within the screening program and compared with the standard oral presentation using a questionnaire. Knowledge outcomes were measured both before and after either of the educational formats. Data from the formative evaluation were used to refine the content and functionality of the final resource. The questionnaire evaluation of 302 students over two years showed the multimedia resource to be equally effective as an oral educational presentation in facilitating participants' knowledge construction. The resource offers a large number of potential benefits, which are not limited to the Tay Sachs disease carrier screening program setting, such as delivery of a consistent educational message, short delivery time, and minimum financial and resource commitment. This article outlines the value of considering educational theory and describes the process of multimedia development providing a framework that may be of value when designing genetics multimedia resources in general.
Full Text Available APOBEC3 family DNA cytosine deaminases provide overlapping defenses against pathogen infections. However, most viruses have elaborate evasion mechanisms such as the HIV-1 Vif protein, which subverts cellular CBF-β and a polyubiquitin ligase complex to neutralize these enzymes. Despite advances in APOBEC3 and Vif biology, a full understanding of this direct host-pathogen conflict has been elusive. We combine virus adaptation and computational studies to interrogate the APOBEC3F-Vif interface and build a robust structural model. A recurring compensatory amino acid substitution from adaptation experiments provided an initial docking constraint, and microsecond molecular dynamic simulations optimized interface contacts. Virus infectivity experiments validated a long-lasting electrostatic interaction between APOBEC3F E289 and HIV-1 Vif R15. Taken together with mutagenesis results, we propose a wobble model to explain how HIV-1 Vif has evolved to bind different APOBEC3 enzymes and, more generally, how pathogens may evolve to escape innate host defenses.
Gordon, Derek; Londono, Douglas; Patel, Payal; Kim, Wonkuk; Finch, Stephen J; Heiman, Gary A
Our motivation here is to calculate the power of 3 statistical tests used when there are genetic traits that operate under a pleiotropic mode of inheritance and when qualitative phenotypes are defined by use of thresholds for the multiple quantitative phenotypes. Specifically, we formulate a multivariate function that provides the probability that an individual has a vector of specific quantitative trait values conditional on having a risk locus genotype, and we apply thresholds to define qualitative phenotypes (affected, unaffected) and compute penetrances and conditional genotype frequencies based on the multivariate function. We extend the analytic power and minimum-sample-size-necessary (MSSN) formulas for 2 categorical data-based tests (genotype, linear trend test [LTT]) of genetic association to the pleiotropic model. We further compare the MSSN of the genotype test and the LTT with that of a multivariate ANOVA (Pillai). We approximate the MSSN for statistics by linear models using a factorial design and ANOVA. With ANOVA decomposition, we determine which factors most significantly change the power/MSSN for all statistics. Finally, we determine which test statistics have the smallest MSSN. In this work, MSSN calculations are for 2 traits (bivariate distributions) only (for illustrative purposes). We note that the calculations may be extended to address any number of traits. Our key findings are that the genotype test usually has lower MSSN requirements than the LTT. More inclusive thresholds (top/bottom 25% vs. top/bottom 10%) have higher sample size requirements. The Pillai test has a much larger MSSN than both the genotype test and the LTT, as a result of sample selection. With these formulas, researchers can specify how many subjects they must collect to localize genes for pleiotropic phenotypes. © 2017 S. Karger AG, Basel.
Full Text Available Abstract Background Although oxidative stress by accumulation of reactive oxygen species (ROS in diabetes has become evident, it remains unclear what genes, involved in redox balance, would determine susceptibility for development of atherosclerosis in diabetes. This study evaluated the effect of genetic polymorphism of enzymes producing or responsible for reducing ROS on coronary artery calcification in type 2 diabetes (T2D. Methods An index for coronary-arteriosclerosis, coronary artery calcium score (CACS was evaluated in 91 T2D patients using a multi-slice computed tomography. Patients were genotyped for ROS-scavenging enzymes, Glutathione peroxidase-1 (GPx-1, Catalase, Mn-SOD, Cu/Zn-SOD, as well as SNPs of NADPH oxidase as ROS-promoting elements, genes related to onset of T2D (CAPN10, ADRB3, PPAR gamma, FATP4. Age, blood pressure, BMI, HbA1c, lipid and duration of diabetes were evaluated for a multivariate regression analysis. Results CACS with Pro/Leu genotype of the GPx-1 gene was significantly higher than in those with Pro/Pro (744 ± 1,291 vs. 245 ± 399, respectively, p = 0.006. In addition, genotype frequency of Pro/Leu in those with CACS ≥ 1000 was significantly higher than in those with CACS OR = 3.61, CI = 0.97–13.42; p = 0.045 when tested for deviation from Hardy-Weinberg's equilibrium. Multivariate regression analyses revealed that CACS significantly correlated with GPx-1 genotypes and age. Conclusion The presence of Pro197Leu substitution of the GPx-1 gene may play a crucial role in determining genetic susceptibility to coronary-arteriosclerosis in T2D. The mechanism may be associated with a decreased ability to scavenge ROS with the variant GPx-1.
Full Text Available Abstract Background Computer-based teaching (CBT is a well-known educational device, but it has never been applied systematically to the teaching of a complex, rare, genetic disease, such as Hunter disease (MPS II. Aim To develop interactive teaching software functioning as a virtual clinic for the management of MPS II. Implementation and Results The Hunter disease eClinic, a self-training, user-friendly educational software program, available at the Lysosomal Storage Research Group (http://www.lysosomalstorageresearch.ca, was developed using the Adobe Flash multimedia platform. It was designed to function both to provide a realistic, interactive virtual clinic and instantaneous access to supporting literature on Hunter disease. The Hunter disease eClinic consists of an eBook and an eClinic. The eClinic is the interactive virtual clinic component of the software. Within an environment resembling a real clinic, the trainee is instructed to perform a medical history, to examine the patient, and to order appropriate investigation. The program provides clinical data derived from the management of actual patients with Hunter disease. The eBook provides instantaneous, electronic access to a vast collection of reference information to provide detailed background clinical and basic science, including relevant biochemistry, physiology, and genetics. In the eClinic, the trainee is presented with quizzes designed to provide immediate feedback on both trainee effectiveness and efficiency. User feedback on the merits of the program was collected at several seminars and formal clinical rounds at several medical centres, primarily in Canada. In addition, online usage statistics were documented for a 2-year period. Feedback was consistently positive and confirmed the practical benefit of the program. The online English-language version is accessed daily by users from all over the world; a Japanese translation of the program is also available. Conclusions The
Al-Jasmi, Fatma; Moldovan, Laura; Clarke, Joe T R
Computer-based teaching (CBT) is a well-known educational device, but it has never been applied systematically to the teaching of a complex, rare, genetic disease, such as Hunter disease (MPS II). To develop interactive teaching software functioning as a virtual clinic for the management of MPS II. The Hunter disease eClinic, a self-training, user-friendly educational software program, available at the Lysosomal Storage Research Group (http://www.lysosomalstorageresearch.ca), was developed using the Adobe Flash multimedia platform. It was designed to function both to provide a realistic, interactive virtual clinic and instantaneous access to supporting literature on Hunter disease. The Hunter disease eClinic consists of an eBook and an eClinic. The eClinic is the interactive virtual clinic component of the software. Within an environment resembling a real clinic, the trainee is instructed to perform a medical history, to examine the patient, and to order appropriate investigation. The program provides clinical data derived from the management of actual patients with Hunter disease. The eBook provides instantaneous, electronic access to a vast collection of reference information to provide detailed background clinical and basic science, including relevant biochemistry, physiology, and genetics. In the eClinic, the trainee is presented with quizzes designed to provide immediate feedback on both trainee effectiveness and efficiency. User feedback on the merits of the program was collected at several seminars and formal clinical rounds at several medical centres, primarily in Canada. In addition, online usage statistics were documented for a 2-year period. Feedback was consistently positive and confirmed the practical benefit of the program. The online English-language version is accessed daily by users from all over the world; a Japanese translation of the program is also available. The Hunter disease eClinic employs a CBT model providing the trainee with realistic
Home; Journals; Journal of Genetics; Volume 80; Issue 1. Testing quantum dynamics in genetic information processing ... Keywords. assembly; computation; database search; DNA replication; genetic information; nucleotide base; polymerase enzyme; quantum coherence; quantum mechanics; quantum superposition.
Yaacoub, Charles; Mhanna, Georges; Rihana, Sandy
Electroencephalography is a non-invasive measure of the brain electrical activity generated by millions of neurons. Feature extraction in electroencephalography analysis is a core issue that may lead to accurate brain mental state classification. This paper presents a new feature selection method that improves left/right hand movement identification of a motor imagery brain-computer interface, based on genetic algorithms and artificial neural networks used as classifiers. Raw electroencephalography signals are first preprocessed using appropriate filtering. Feature extraction is carried out afterwards, based on spectral and temporal signal components, and thus a feature vector is constructed. As various features might be inaccurate and mislead the classifier, thus degrading the overall system performance, the proposed approach identifies a subset of features from a large feature space, such that the classifier error rate is reduced. Experimental results show that the proposed method is able to reduce the number of features to as low as 0.5% (i.e., the number of ignored features can reach 99.5%) while improving the accuracy, sensitivity, specificity, and precision of the classifier.
Full Text Available Electroencephalography is a non-invasive measure of the brain electrical activity generated by millions of neurons. Feature extraction in electroencephalography analysis is a core issue that may lead to accurate brain mental state classification. This paper presents a new feature selection method that improves left/right hand movement identification of a motor imagery brain-computer interface, based on genetic algorithms and artificial neural networks used as classifiers. Raw electroencephalography signals are first preprocessed using appropriate filtering. Feature extraction is carried out afterwards, based on spectral and temporal signal components, and thus a feature vector is constructed. As various features might be inaccurate and mislead the classifier, thus degrading the overall system performance, the proposed approach identifies a subset of features from a large feature space, such that the classifier error rate is reduced. Experimental results show that the proposed method is able to reduce the number of features to as low as 0.5% (i.e., the number of ignored features can reach 99.5% while improving the accuracy, sensitivity, specificity, and precision of the classifier.
Computer program for allocation of generators in isolated systems of direct current using genetic algorithm; Programa computacional para alocacao de geradores em sistemas isolados de corrente continua utilizando algoritmo genetico
Gewehr, Diego N.; Vargas, Ricardo B.; Melo, Eduardo D. de; Paschoareli Junior, Dionizio [Universidade Estadual Paulista (DEE/UNESP), Ilha Solteira, SP (Brazil). Dept. de Engenharia Eletrica. Grupo de Pesquisa em Fontes Alternativas e Aproveitamento de Energia
This paper presents a methodology for electric power sources location in isolated direct current micro grids, using genetic algorithm. In this work, photovoltaic panels are considered, although the methodology can be extended for any kind of DC sources. A computational tool is developed using the Matlab simulator, to obtain the best dc system configuration for reduction of panels quantity and costs, and to improve the system performance. (author)
Wang, Lui; Bayer, Steven E.
Genetic algorithms are mathematical, highly parallel, adaptive search procedures (i.e., problem solving methods) based loosely on the processes of natural genetics and Darwinian survival of the fittest. Basic genetic algorithms concepts are introduced, genetic algorithm applications are introduced, and results are presented from a project to develop a software tool that will enable the widespread use of genetic algorithm technology.
... greatly advanced genetics research. The improved quality of genetic data has reduced the time required to identify a ... cases, a matter of months or even weeks. Genetic mapping data generated by the HGP's laboratories is freely accessible ...
The book presents state-of-the-art works in computational engineering. Focus is on mathematical modeling, numerical simulation, experimental validation and visualization in engineering sciences. In particular, the following topics are presented: constitutive models and their implementation into finite element codes, numerical models in nonlinear elasto-dynamics including seismic excitations, multiphase models in structural engineering and multiscale models of materials systems, sensitivity and reliability analysis of engineering structures, the application of scientific computing in urban water management and hydraulic engineering, and the application of genetic algorithms for the registration of laser scanner point clouds.
During the past 10 years, the number of genetic tests performed more than tripled, and public concern about genetic privacy emerged. The majority of states and the U.S. government have passed regulations protecting genetic information. However, research has shown that concerns about genetic privacy are disproportionate to known instances of information misuse. Beliefs in genetic determinacy explain some of the heightened concern about genetic privacy. Discussion of the debate over genetic testing within families illustrates the most recent response to genetic privacy concerns.
Resconi, Germano; Xu, Guanglin
This book offers a concise introduction to morphogenetic computing, showing that its use makes global and local relations, defects in crystal non-Euclidean geometry databases with source and sink, genetic algorithms, and neural networks more stable and efficient. It also presents applications to database, language, nanotechnology with defects, biological genetic structure, electrical circuit, and big data structure. In Turing machines, input and output states form a system – when the system is in one state, the input is transformed into output. This computation is always deterministic and without any possible contradiction or defects. In natural computation there are defects and contradictions that have to be solved to give a coherent and effective computation. The new computation generates the morphology of the system that assumes different forms in time. Genetic process is the prototype of the morphogenetic computing. At the Boolean logic truth value, we substitute a set of truth (active sets) values with...
Lauritzen, Steffen Lilholt; Sheehan, Nuala A.
This paper introduces graphical models as a natural environment in which to formulate and solve problems in genetics and related areas. Particular emphasis is given to the relationships among various local computation algorithms which have been developed within the hitherto mostly separate areas...... of graphical models and genetics. The potential of graphical models is explored and illustrated through a number of example applications where the genetic element is substantial or dominating....
Presents a variant of Genetic Programming that evolves imperative computer programs as linear sequences of instructions, in contrast to the more traditional functional expressions or syntax trees. This book serves as a reference for researchers, but also contains sufficient introduction for students and those who are new to the field
Resnik, David B; Vorhaus, Daniel B
In this article we examine four objections to the genetic modification of human beings: the freedom argument, the giftedness argument, the authenticity argument, and the uniqueness argument. We then demonstrate that each of these arguments against genetic modification assumes a strong version of genetic determinism. Since these strong deterministic assumptions are false, the arguments against genetic modification, which assume and depend upon these assumptions, are therefore unsound. Serious discussion of the morality of genetic modification, and the development of sound science policy, should be driven by arguments that address the actual consequences of genetic modification for individuals and society, not by ones propped up by false or misleading biological assumptions. PMID:16800884
Fernández, E N; Legarra, A; Martínez, R; Sánchez, J P; Baselga, M
Inbreeding generates covariances between additive and dominance effects (breeding values and dominance deviations). In this work, we developed and applied models for estimation of dominance and additive genetic variances and their covariance, a model that we call "full dominance," from pedigree and phenotypic data. Estimates with this model such as presented here are very scarce both in livestock and in wild genetics. First, we estimated pedigree-based condensed probabilities of identity using recursion. Second, we developed an equivalent linear model in which variance components can be estimated using closed-form algorithms such as REML or Gibbs sampling and existing software. Third, we present a new method to refer the estimated variance components to meaningful parameters in a particular population, i.e., final partially inbred generations as opposed to outbred base populations. We applied these developments to three closed rabbit lines (A, V and H) selected for number of weaned at the Polytechnic University of Valencia. Pedigree and phenotypes are complete and span 43, 39 and 14 generations, respectively. Estimates of broad-sense heritability are 0.07, 0.07 and 0.05 at the base versus 0.07, 0.07 and 0.09 in the final generations. Narrow-sense heritability estimates are 0.06, 0.06 and 0.02 at the base versus 0.04, 0.04 and 0.01 at the final generations. There is also a reduction in the genotypic variance due to the negative additive-dominance correlation. Thus, the contribution of dominance variation is fairly large and increases with inbreeding and (over)compensates for the loss in additive variation. In addition, estimates of the additive-dominance correlation are -0.37, -0.31 and 0.00, in agreement with the few published estimates and theoretical considerations. © 2017 Blackwell Verlag GmbH.
artificial genetic system) string feature or ... called the genotype whereas it is called a structure in artificial genetic ... assigned a fitness value based on the cost function. Better ..... way it has produced complex, intelligent living organisms capable of ...
Genetic algorithms mimic biological evolution by natural selection in their search for better individuals within a changing population. they can be used as efficient optimizers. This report discusses the developing field of genetic algorithms. It gives a simple example of the search process and introduces the concept of schema. It also discusses modifications to the basic genetic algorithm that result in species and niche formation, in machine learning and artificial evolution of computer programs, and in the streamlining of human-computer interaction. (author). 3 refs., 1 tab., 2 figs
... clinical care in many areas of medicine. Assisted Reproductive Technology/Infertility Genetics Cancer Genetics Cardiovascular Genetics Cystic Fibrosis Genetics Fetal Intervention and Therapy Genetics Hematology Genetics Metabolic Genetics ...
programming paradigm. The induction of mathematical expressions based on data is called symbolic regression. In this work, genetic programming is extended to not just fit the data i.e., get the numbers right, but also to get the dimensions right. For this units of measurement are used. The main contribution......Genetic Programming is capable of automatically inducing symbolic computer programs on the basis of a set of examples or their performance in a simulation. Mathematical expressions are a well-defined subset of symbolic computer programs and are also suitable for optimization using the genetic...... in this work can be summarized as: The symbolic expressions produced by genetic programming can be made suitable for analysis and interpretation by using units of measurements to guide or restrict the search. To achieve this, the following has been accomplished: A standard genetic programming system...
Computational Neurogenetic Modeling is a student text, introducing the scope and problems of a new scientific discipline - Computational Neurogenetic Modeling (CNGM). CNGM is concerned with the study and development of dynamic neuronal models for modeling brain functions with respect to genes and dynamic interactions between genes. These include neural network models and their integration with gene network models. This new area brings together knowledge from various scientific disciplines, such as computer and information science, neuroscience and cognitive science, genetics and molecular biol
Moor, J H
Computing plays an important role in genetics (and vice versa). Theoretically, computing provides a conceptual model for the function and malfunction of our genetic machinery. Practically, contemporary computers and robots equipped with advanced algorithms make the revelation of the complete human genome imminent--computers are about to reveal our genetic souls for the first time. Ethically, computers help protect privacy by restricting access in sophisticated ways to genetic information. But the inexorable fact that computers will increasingly collect, analyze, and disseminate abundant amounts of genetic information made available through the genetic revolution, not to mention that inexpensive computing devices will make genetic information gathering easier, underscores the need for strong and immediate privacy legislation.
Vorhaus Daniel B
Full Text Available Abstract In this article we examine four objections to the genetic modification of human beings: the freedom argument, the giftedness argument, the authenticity argument, and the uniqueness argument. We then demonstrate that each of these arguments against genetic modification assumes a strong version of genetic determinism. Since these strong deterministic assumptions are false, the arguments against genetic modification, which assume and depend upon these assumptions, are therefore unsound. Serious discussion of the morality of genetic modification, and the development of sound science policy, should be driven by arguments that address the actual consequences of genetic modification for individuals and society, not by ones propped up by false or misleading biological assumptions.
Presents a review of genetic engineering, in which the genotypes of plants and animals (including human genotypes) may be manipulated for the benefit of the human species. Discusses associated problems and solutions and provides an extensive bibliography of literature relating to genetic engineering. (JR)
inheritance as a way to unify populations within politically and geographically bounded areas. Thus, new genetics have contributed to the development of genetic romanticisms, whereby populations (human, plant, and animal) can be delineated and mobilized through scientific and medical practices to represent...
systems that would fall under the Exascale rubric . In this chapter, we first discuss the attributes by which achievement of the label “Exascale” may be...Carrington, and E. Strohmaier. A Genetic Algorithms Approach to Modeling the Performance of Memory-bound Computations. Reno, NV, November 2007. ACM/IEEE... genetic stochasticity (random mating, mutation, etc). Outcomes are thus stochastic as well, and ecologists wish to ask questions like, “What is the
Tetushkin, E Iu
The supplementary historical discipline genealogy is also a supplementary genetic discipline. In its formation, genetics borrowed from genealogy some methods of pedigree analysis. In the 21th century, it started receiving contribution from computer-aided genealogy and genetic (molecular) genealogy. The former provides novel tools for genetics, while the latter, which employing genetic methods, enriches genetics with new evidence. Genealogists formulated three main laws ofgenealogy: the law of three generations, the law of doubling the ancestry number, and the law of declining ancestry. The significance and meaning of these laws can be fully understood only in light of genetics. For instance, a controversy between the exponential growth of the number of ancestors of an individual, i.e., the law of doubling the ancestry number, and the limited number of the humankind is explained by the presence of weak inbreeding because of sibs' interference; the latter causes the pedigrees' collapse, i.e., explains also the law of diminishing ancestry number. Mathematic modeling of pedigrees' collapse presented in a number of studies showed that the number of ancestors of each individual attains maximum in a particular generation termed ancestry saturated generation. All representatives of this and preceding generation that left progeny are common ancestors of all current members of the population. In subdivided populations, these generations are more ancient than in panmictic ones, whereas in small isolates and social strata with limited numbers of partners, they are younger. The genealogical law of three generations, according to which each hundred years contain on average three generation intervals, holds for generation lengths for Y-chromosomal DNA, typically equal to 31-32 years; for autosomal and mtDNA, this time is somewhat shorter. Moving along ascending lineas, the number of genetically effective ancestors transmitting their DNA fragment to descendants increases far
Handbook of technology law. General funamentals, environment law, genetic engineering act, energy act, telecommunication act and media act, patent act, computer act. 2. ed.; Handbuch des Technikrechts. Allgemeine Grundlagen Umweltrecht, Gentechnikrecht, Energierecht, Telekommunikations- und Medienrecht, Patentrecht, Computerrecht
Schulte, Martin; Schroeder, Rainer (eds.) [Technische Univ. Dresden (Germany). Juristische Fakultaet
On the boundaries between technology sciences, jurisprudence, social sciences and economic science the technology law proves as a cross-sectional area par excellence. The bases of the technology law are presented: individual, particularly important scopes of the technology law (appliance safety regulations, technology law and environment law, genetic engineering act, energy right, telecommunications law and media law, patent law, computer law, data security, legally binding telecooperation) are analyzed in detail. The manual contacts all lawyers who want to provide a first in-depth insight of this new field of law. [German] Im Grenzbereich von Technik-, Rechts-, Sozial- und Wirtschaftswissenschaften erweist sich das Technikrecht als Querschnittsmaterie par excellence. Die Grundlagen des Technikrechts werden dargestellt; einzelne, besonders wichtige Bereiche des Technikrechts (Geraetesicherheitsrecht, Technik und Umweltrecht, Gentechnikrecht, Energierecht, Telekommunikations- und Medienrecht, Patentrecht, Computerrecht, Datensicherheit, Rechtsverbindliche Telekooperation) werden eingehend analysiert. Das Handbuch wendet sich an alle in Wissenschaft und Praxis mit dem Technikrecht befassten Juristen, die sich einen ersten vertieften Einblick in dieses neue Rechtsgebiet verschaffen wollen. (orig.)
Maynard Smith, John
.... It differs from other textbooks of population genetics in applying the basic theory to topics, such as social behaviour, molecular evolution, reiterated DNA, and sex, which are the main subjects...
... Care Genomic Medicine Working Group New Horizons and Research Patient Management Policy and Ethics Issues Quick Links for Patient Care Education All About the Human Genome Project Fact Sheets Genetic Education Resources for ...
Introduces an activity on arthropod genetics that involves phenotype and genotype identification of the creature and the construction process. Includes a list of required materials and directions to build a model arthropod. (YDS)
Hough, Soren H; Ajetunmobi, Ayokunmi; Brody, Leigh; Humphryes-Kirilov, Neil; Perello, Edward
Desktop Genetics is a bioinformatics company building a gene-editing platform for personalized medicine. The company works with scientists around the world to design and execute state-of-the-art clustered regularly interspaced short palindromic repeats (CRISPR) experiments. Desktop Genetics feeds the lessons learned about experimental intent, single-guide RNA design and data from international genomics projects into a novel CRISPR artificial intelligence system. We believe that machine learni...
Kia, Behnam; Murali, K.; Jahed Motlagh, Mohammad-Reza; Sinha, Sudeshna; Ditto, William L.
Nonlinearity and chaos can illustrate numerous behaviors and patterns, and one can select different patterns from this rich library of patterns. In this paper we focus on synthetic computing, a field that engineers and synthesizes nonlinear systems to obtain computation. We explain the importance of nonlinearity, and describe how nonlinear systems can be engineered to perform computation. More specifically, we provide an overview of chaos computing, a field that manually programs chaotic systems to build different types of digital functions. Also we briefly describe logical stochastic resonance (LSR), and then extend the approach of LSR to realize combinational digital logic systems via suitable concatenation of existing logical stochastic resonance blocks. Finally we demonstrate how a chaotic system can be engineered and mated with different machine learning techniques, such as artificial neural networks, random searching, and genetic algorithm, to design different autonomous systems that can adapt and respond to environmental conditions.
Journal of Genetics, Vol. 85, No. 2, August 2006. 101. Page 2. J. Genet. classic. 102. Journal of Genetics, Vol. 85, No. 2, August 2006. Page 3. J. Genet. classic. Journal of Genetics, Vol. 85, No. 2, August 2006. 103. Page 4. J. Genet. classic. 104. Journal of Genetics, Vol. 85, No. 2, August 2006. Page 5. J. Genet. classic.
Journal of Genetics, Vol. 84, No. 1, April 2005. 37. Page 2. J. Genet. classic. Journal of Genetics, Vol. 84, No. 1, April 2005. 38. Page 3. J. Genet. classic. Journal of Genetics, Vol. 84, No. 1, April 2005. 39. Page 4. J. Genet. classic. Journal of Genetics, Vol. 84, No. 1, April 2005. 40. Page 5. J. Genet. classic. Journal of ...
Mora, Antonio M.; Squillero, Giovanni; Di Chio, C; Agapitos, Alexandros; Cagnoni, Stefano; Cotta, Carlos; Fernández De Vega, F; Di Caro, G A; Drechsler, R.; Ekárt, A; Esparcia-Alcázar, Anna I.; Farooq, M; Langdon, W B; Merelo-Guervós, J.J.; Preuss, M; Richter, O.-M.H.; Silva, Sara; Sim$\\$~oes, A; Squillero, Giovanni; Tarantino, Ernesto; Tettamanzi, Andrea G B; Togelius, J; Urquhart, Neil; Uyar, A S; Yannakakis, G N; Smith, Stephen L; Caserta, Marco; Ramirez, Adriana; Voß, Stefan; Squillero, Giovanni; Burelli, Paolo; Mora, Antonio M.; Squillero, Giovanni; Jan, Mathieu; Matthias, M; Di Chio, C; Agapitos, Alexandros; Cagnoni, Stefano; Cotta, Carlos; Fernández De Vega, F; Di Caro, G A; Drechsler, R.; Ekárt, A; Esparcia-Alcázar, Anna I.; Farooq, M; Langdon, W B; Merelo-Guervós, J.J.; Preuss, M; Richter, O.-M.H.; Silva, Sara; Sim$\\$~oes, A; Squillero, Giovanni; Tarantino, Ernesto; Tettamanzi, Andrea G B; Togelius, J; Urquhart, Neil; Uyar, A S; Yannakakis, G N; Caserta, Marco; Ramirez, Adriana; Voß, Stefan; Squillero, Giovanni; Burelli, Paolo; Esparcia-Alcazar, Anna I; Silva, Sara; Agapitos, Alexandros; Cotta, Carlos; De Falco, Ivanoe; Cioppa, Antonio Della; Diwold, Konrad; Ekart, Aniko; Tarantino, Ernesto; Vega, Francisco Fernandez De; Burelli, Paolo; Sim, Kevin; Cagnoni, Stefano; Simoes, Anabela; Merelo, J.J.; Urquhart, Neil; Haasdijk, Evert; Zhang, Mengjie; Squillero, Giovanni; Eiben, A E; Tettamanzi, Andrea G B; Glette, Kyrre; Rohlfshagen, Philipp; Schaefer, Robert; Caserta, Marco; Ramirez, Adriana; Voß, Stefan
The application of genetic and evolutionary computation to problems in medicine has increased rapidly over the past five years, but there are specific issues and challenges that distinguish it from other real-world applications. Obtaining reliable and coherent patient data, establishing the clinical
Jacquez, Geoffrey; Sabel, Clive E; Shi, Chen
The exposome, defined as the totality of an individual's exposures over the life course, is a seminal concept in the environmental health sciences. Although inherently geographic, the exposome as yet is unfamiliar to many geographers. This article proposes a place-based synthesis, genetic...... geographic information science (genetic GIScience), that is founded on the exposome, genome+, and behavome. It provides an improved understanding of human health in relation to biology (the genome+), environmental exposures (the exposome), and their social, societal, and behavioral determinants (the behavome......). Genetic GIScience poses three key needs: first, a mathematical foundation for emergent theory; second, process-based models that bridge biological and geographic scales; third, biologically plausible estimates of space?time disease lags. Compartmental models are a possible solution; this article develops...
Hough, Soren H; Ajetunmobi, Ayokunmi; Brody, Leigh; Humphryes-Kirilov, Neil; Perello, Edward
Desktop Genetics is a bioinformatics company building a gene-editing platform for personalized medicine. The company works with scientists around the world to design and execute state-of-the-art clustered regularly interspaced short palindromic repeats (CRISPR) experiments. Desktop Genetics feeds the lessons learned about experimental intent, single-guide RNA design and data from international genomics projects into a novel CRISPR artificial intelligence system. We believe that machine learning techniques can transform this information into a cognitive therapeutic development tool that will revolutionize medicine.
Baum, E B; Boneh, D; Garrett, C
We analyze the performance of a genetic algorithm (GA) we call Culling, and a variety of other algorithms, on a problem we refer to as the Additive Search Problem (ASP). We show that the problem of learning the Ising perceptron is reducible to a noisy version of ASP. Noisy ASP is the first problem we are aware of where a genetic-type algorithm bests all known competitors. We generalize ASP to k-ASP to study whether GAs will achieve "implicit parallelism" in a problem with many more schemata. GAs fail to achieve this implicit parallelism, but we describe an algorithm we call Explicitly Parallel Search that succeeds. We also compute the optimal culling point for selective breeding, which turns out to be independent of the fitness function or the population distribution. We also analyze a mean field theoretic algorithm performing similarly to Culling on many problems. These results provide insight into when and how GAs can beat competing methods.
... of the booklet. » more Chapter 1: How Genes Work Covers DNA, RNA, transcription, RNA splicing, translation, ribosomes, antibiotics, genetic diseases, gene chips. » more Chapter 2: RNA and DNA Revealed: New Roles, New Rules Covers microRNAs, RNAi, epigenetics, telomeres, mtDNA, recombinant DNA. » ...
In 1948-1953 a large scale field survey was conducted to investigate the possible genetic effects of A-bomb radiation on over 70,000 pregnancy terminations in the cities of Hiroshima and Nagasaki. The indices of possible genetic effect including sex ratio, birth weight, frequency of malformation, stillbirth, neonatal death, deaths within 9 months and anthropometric measurements at 9 months of age for these children were investigated in relation to their parent's exposure status to the A-bomb. There were no detectable genetic effects in this sample, except for a slight change in sex ratio which was in the direction to be expected if exposure had induced sex-linked lethal mutations. However, continued study of the sex ratio, based upon birth certificates in Hiroshima and Nagasaki for 1954-1962, did not confirm the earlier trend. Mortality in these children of A-bomb survivors is being followed using a cohort of 54,000 subjects. No clearly significant effect of parental exposure on survival of the children has been demonstrated up to 1972 (age 17 on the average). On the basis of the regression data, the minimal genetic doubling dose of this type of radiation for mutations resulting in death is estimated at 46 rem for the father and 125 rem for the mother. (auth.)
Read, Jazlyn; Wadt, Karin A W; Hayward, Nicholas K
Approximately 10% of melanoma cases report a relative affected with melanoma, and a positive family history is associated with an increased risk of developing melanoma. Although the majority of genetic alterations associated with melanoma development are somatic, the underlying presence of herita......Approximately 10% of melanoma cases report a relative affected with melanoma, and a positive family history is associated with an increased risk of developing melanoma. Although the majority of genetic alterations associated with melanoma development are somatic, the underlying presence...... in a combined total of approximately 50% of familial melanoma cases, the underlying genetic basis is unexplained for the remainder of high-density melanoma families. Aside from the possibility of extremely rare mutations in a few additional high penetrance genes yet to be discovered, this suggests a likely...... polygenic component to susceptibility, and a unique level of personal melanoma risk influenced by multiple low-risk alleles and genetic modifiers. In addition to conferring a risk of cutaneous melanoma, some 'melanoma' predisposition genes have been linked to other cancers, with cancer clustering observed...
Whitehouse, H. L. K.
Discusses the mechanisms of genetic recombination with particular emphasis on the study of the fungus Sordaria brevicollis. The study of recombination is facilitated by the use of mutants of this fungus in which the color of the ascospores is affected. (JR)
Koornneef, M.; Alonso-Blanco, C.; Stam, P.
The Mendelian analysis of genetic variation, available as induced mutants or as natural variation, requires a number of steps that are described in this chapter. These include the determination of the number of genes involved in the observed trait's variation, the determination of dominance
Parkinson, D.R.; Krontiris, T.G.
In this chapter the authors review new findings concerning the molecular genetics of malignant melanoma in the context of other information obtained from clinical, epidemiologic, and cytogenetic studies in this malignancy. These new molecular approaches promise to provide a more complete understanding of the mechanisms involved in the development of melanoma, thereby suggesting new methods for its treatment and prevention
Journal of Genetics, Vol. 83, No. 3, December 2004. 235. Page 2. J. Genet. classic. Journal of Genetics, Vol. 83, No. 3, December 2004. 236. Page 3. J. Genet. classic. Journal of Genetics, Vol. 83, No. 3, December 2004. 237. Page 4. J. Genet. classic. Journal of Genetics, Vol. 83, No. 3, December 2004. 238. Page 5 ...
Bender, M.A.; Abrahamson, S.; Denniston, C.; Schull, W.J.
In this chapter, we present a comprehensive analysis of the major classes of genetic diseases that would be increased as a result of an increased gonadal radiation exposure to a human population. The risk analysis takes on two major forms: the increase in genetic disease that would be observed in the immediate offspring of the exposed population, and the subsequent transmission of the newly induced mutations through future generations. The major classes of genetic disease will be induced at different frequencies, and will also impact differentially in terms of survivability and fertility on the affected individuals and their descendants. Some classes of disease will be expected to persist for only a few generations at most. Other types of genetic disease will persist through a longer period. The classes of genetic diseases studied are: dominant gene mutation, X-linked gene mutation, chromosome disorders and multifactorial disorders which involve the interaction of many mutant genes and environmental factors. For each of these classes we have derived the general equations of mutation induction for the male and female germ cells of critical importance in the mutation process. The frequency of induced mutations will be determined initially by the dose received, the type of radiation and, to some extent at high dose, by the manner in which the dose is received. We have used the modeling analyses to predict the outcomes for two nuclear power plant accident scenarios, the first in which the population receives a chronic dose of 0.1 Gy (10 rad) over a 50-year period, the second in which an equivalent population receives an acute dose of 2 Gy. In both cases the analyses are projected over a period of five generations
Abrahamson, S.; Bender, M.; Denniston, C.; Schull, W.
Modeling analyses are used to predict the outcomes for two nuclear power plant accident scenarios, the first in which the population received a chronic dose of 0.1 Gy (10 rad) over a 50 year period, the second in which an equivalent population receives acute dose of 2 Gy. In both cases the analyses are projected over a period of five generations. The risk analysis takes on two major forms: the increase in genetic disease that would be observed in the immediate offspring of the exposed population, and the subsequent transmission of the newly induced mutations through future generations. The classes of genetic diseases studied are: dominant gene mutation, X-linked gene mutation, chromosome disorders and multifactorial disorders which involve the interaction of many mutant genes and environmental factors. 28 references, 3 figures, 5 tables
... Services Directory Cancer Prevention Overview Research NCI Cancer Genetics Services Directory This directory lists professionals who provide services related to cancer genetics (cancer risk assessment, genetic counseling, genetic susceptibility testing, ...
Woods, Damien; Naughton, Thomas J.
We consider optical computers that encode data using images and compute by transforming such images. We give an overview of a number of such optical computing architectures, including descriptions of the type of hardware commonly used in optical computing, as well as some of the computational efficiencies of optical devices. We go on to discuss optical computing from the point of view of computational complexity theory, with the aim of putting some old, and some very recent, re...
Bauer, H.; Black, I.; Heusler, A.; Hoeptner, G.; Krafft, F.; Lang, R.; Moellenkamp, R.; Mueller, W.; Mueller, W.F.; Schati, C.; Schmidt, A.; Schwind, D.; Weber, G.
The computer groups has been reorganized to take charge for the general purpose computers DEC10 and VAX and the computer network (Dataswitch, DECnet, IBM - connections to GSI and IPP, preparation for Datex-P). (orig.)
Looks at computer engineers and describes their job, employment outlook, earnings, and training and qualifications. Provides a list of resources related to computer engineering careers and the computer industry. (JOW)
Cook, Perry R.
This chapter covers algorithms, technologies, computer languages, and systems for computer music. Computer music involves the application of computers and other digital/electronic technologies to music composition, performance, theory, history, and the study of perception. The field combines digital signal processing, computational algorithms, computer languages, hardware and software systems, acoustics, psychoacoustics (low-level perception of sounds from the raw acoustic signal), and music cognition (higher-level perception of musical style, form, emotion, etc.).
Posthuma, Daniëlle; Beem, A Leo; de Geus, Eco J C
With the rapid advances in molecular biology, the near completion of the human genome, the development of appropriate statistical genetic methods and the availability of the necessary computing power, the identification of quantitative trait loci has now become a realistic prospect for quantitative...... geneticists. We briefly describe the theoretical biometrical foundations underlying quantitative genetics. These theoretical underpinnings are translated into mathematical equations that allow the assessment of the contribution of observed (using DNA samples) and unobserved (using known genetic relationships......) genetic variation to population variance in quantitative traits. Several statistical models for quantitative genetic analyses are described, such as models for the classical twin design, multivariate and longitudinal genetic analyses, extended twin analyses, and linkage and association analyses. For each...
of genetic means and variances, models for the analysis of categorical and count data, the statistical genetics of a model postulating that environmental variance is partly under genetic control, and a short discussion of models that incorporate massive genetic marker information. We provide an overview......A remarkable research impetus has taken place in statistical genetics since the last World Conference. This has been stimulated by breakthroughs in molecular genetics, automated data-recording devices and computer-intensive statistical methods. The latter were revolutionized by the bootstrap...... and by Markov chain Monte Carlo (McMC). In this overview a number of specific areas are chosen to illustrate the enormous flexibility that McMC has provided for fitting models and exploring features of data that were previously inaccessible. The selected areas are inferences of the trajectories over time...
Bosworth, J. L.; Foo, N. Y.; Zeigler, B. P.
Genetic algorithms for mathematical function optimization are modeled on search strategies employed in natural adaptation. Comparisons of genetic algorithms with conjugate gradient methods, which were made on an IBM 1800 digital computer, show that genetic algorithms display superior performance over gradient methods for functions which are poorly behaved mathematically, for multimodal functions, and for functions obscured by additive random noise. Genetic methods offer performance comparable to gradient methods for many of the standard functions.
Hanks, Ephraim M.; Hooten, Mevin B.; Knick, Steven T.; Oyler-McCance, Sara J.; Fike, Jennifer A.; Cross, Todd B.; Schwartz, Michael K.
We propose a spatially-explicit approach for modeling genetic variation across space and illustrate how this approach can be used to optimize spatial prediction and sampling design for landscape genetic data. We propose a multinomial data model for categorical microsatellite allele data commonly used in landscape genetic studies, and introduce a latent spatial random effect to allow for spatial correlation between genetic observations. We illustrate how modern dimension reduction approaches to spatial statistics can allow for efficient computation in landscape genetic statistical models covering large spatial domains. We apply our approach to propose a retrospective spatial sampling design for greater sage-grouse (Centrocercus urophasianus) population genetics in the western United States.
Progress is reported on studies on the nature and action of lethal and mutagenic lesions in DNA and the mechanisms by which these are produced in bacteria by ionizing radiation or by decay of radioisotopes incorporated in DNA. Studies of radioisotope decay provide the advantages that the original lesion is localized in the genetic material and the immediate physical and chemical changes that occur at decay are known. Specific types of DNA damage were related to characteristic decay properties of several radioisotopes. Incorporated 125 I, for example, induces a double-stranded break in DNA with almost every decay, but causes remarkably little damage of any other kind to the DNA. (U.S.)
Tvedebrink, Torben; Morling, Niels
The increase in the number of forensic genetic loci used for identification purposes results in infinitesimal random match probabilities. These probabilities are computed under assumptions made for rather simple population genetic models. Often, the forensic expert reports likelihood ratios, where...... published results accounting for close familial relationships. However, we revisit the discussion to increase the awareness among forensic genetic practitioners and include new information on medical and societal factors to assess the risk of not considering a monozygotic twin as the true perpetrator......, then data relevant for the Danish society suggests that the threshold of likelihood ratios should approximately be between 150,000 and 2,000,000 in order to take the risk of an unrecognised identical, monozygotic twin into consideration. In other societies, the threshold of the likelihood ratio in crime...
Full Text Available The paper is focused on numerical studies of electromagnetic properties of composite materials used for the construction of small airplanes. Discussions concentrate on the genetic homogenization of composite layers and composite layers with a slot. The homogenization is aimed to reduce CPU-time demands of EMC computational models of electrically large airplanes. First, a methodology of creating a 3-dimensional numerical model of a composite material in CST Microwave Studio is proposed focusing on a sufficient accuracy of the model. Second, a proper implementation of a genetic optimization in Matlab is discussed. Third, an association of the optimization script and a simplified 2-dimensional model of the homogeneous equivalent model in Comsol Multiphysics is proposed considering EMC issues. Results of computations are experimentally verified.
This book is a comprehensive introduction to analog computing. As most textbooks about this powerful computing paradigm date back to the 1960s and 1970s, it fills a void and forges a bridge from the early days of analog computing to future applications. The idea of analog computing is not new. In fact, this computing paradigm is nearly forgotten, although it offers a path to both high-speed and low-power computing, which are in even more demand now than they were back in the heyday of electronic analog computers.
Vallgårda, Anna K. A.; Redström, Johan
Computational composite is introduced as a new type of composite material. Arguing that this is not just a metaphorical maneuver, we provide an analysis of computational technology as material in design, which shows how computers share important characteristics with other materials used in design...... and architecture. We argue that the notion of computational composites provides a precise understanding of the computer as material, and of how computations need to be combined with other materials to come to expression as material. Besides working as an analysis of computers from a designer’s point of view......, the notion of computational composites may also provide a link for computer science and human-computer interaction to an increasingly rapid development and use of new materials in design and architecture....
The aim of this thesis was to explain what quantum computing is. The information for the thesis was gathered from books, scientific publications, and news articles. The analysis of the information revealed that quantum computing can be broken down to three areas: theories behind quantum computing explaining the structure of a quantum computer, known quantum algorithms, and the actual physical realizations of a quantum computer. The thesis reveals that moving from classical memor...
... Well with Rheumatic Disease Genetics and Rheumatic Disease Genetics and Rheumatic Disease Fast Facts Studying twins has ... 70%, and for non-identical pairs, even lower. Genetics and ankylosing spondylitis Each rheumatic disease has its ...
Hopwood, D A
Formerly, when microbiologists had only existing organisms at their disposal whose characteristics could only be changed randomly by genetic experiments, they used to dream of programmed genetic changes. This dream has come true with modern genetic engineering.
Full Text Available Approximate Bayesian computation techniques, also called likelihood-free methods, are one of the most satisfactory approach to intractable likelihood problems. This overview presents recent results since its introduction about ten years ago in population genetics.
Monendra Grover; Rajesh Kumar; Tapan Kumar Mondal; S. Rajkumar
Genetic erosion is a serious problem and computational models have been developed to prevent it. The computational modeling in this field not only includes (terrestrial) reserve design, but also decision modeling for related problems such as habitat restoration, marine reserve design, and nonreserve approaches to conservation management. Models have been formulated for evaluating tradeoffs between socioeconomic, biophysical, and spatial criteria in establishing marine reserves. The percolatio...
Libbrecht, Maxwell W.; Noble, William Stafford
The field of machine learning promises to enable computers to assist humans in making sense of large, complex data sets. In this review, we outline some of the main applications of machine learning to genetic and genomic data. In the process, we identify some recurrent challenges associated with this type of analysis and provide general guidelines to assist in the practical application of machine learning to real genetic and genomic data. PMID:25948244
Single and Multi-Objective Evolutionary Computation (MOEA), Genetic Algorithms (GAs), Artificial Neural Networks (ANNs), Fuzzy Controllers (FCs), Particle Swarm Optimization (PSO) and Ant colony Optimization (ACO) are becoming omnipresent in almost every intelligent system design. Unfortunately, the application of the majority of these techniques is complex and so requires a huge computational effort to yield useful and practical results. Therefore, dedicated hardware for evolutionary, neural and fuzzy computation is a key issue for designers. With the spread of reconfigurable hardware such as FPGAs, digital as well as analog hardware implementations of such computation become cost-effective. The idea behind this book is to offer a variety of hardware designs for soft computing techniques that can be embedded in any final product. Also, to introduce the successful application of soft computing technique to solve many hard problem encountered during the design of embedded hardware designs. Reconfigurable em...
Nygaard, Jens Vinge
The Health Technology Program at Aarhus University applies computational biology to investigate the heterogeneity of tumours......The Health Technology Program at Aarhus University applies computational biology to investigate the heterogeneity of tumours...
A computing grid interconnects resources such as high performancecomputers, scientific databases, and computercontrolledscientific instruments of cooperating organizationseach of which is autonomous. It precedes and is quitedifferent from cloud computing, which provides computingresources by vendors to customers ...
Full Text Available Green computing is all about using computers in a smarter and eco-friendly way. It is the environmentally responsible use of computers and related resources which includes the implementation of energy-efficient central processing units, servers and peripherals as well as reduced resource consumption and proper disposal of electronic waste .Computers certainly make up a large part of many people lives and traditionally are extremely damaging to the environment. Manufacturers of computer and its parts have been espousing the green cause to help protect environment from computers and electronic waste in any way.Research continues into key areas such as making the use of computers as energy-efficient as Possible, and designing algorithms and systems for efficiency-related computer technologies.
Valiev, Kamil' A
This review outlines the principles of operation of quantum computers and their elements. The theory of ideal computers that do not interact with the environment and are immune to quantum decohering processes is presented. Decohering processes in quantum computers are investigated. The review considers methods for correcting quantum computing errors arising from the decoherence of the state of the quantum computer, as well as possible methods for the suppression of the decohering processes. A brief enumeration of proposed quantum computer realizations concludes the review. (reviews of topical problems)
Quantum computers can (in theory) solve certain problems far faster than a classical computer running any known classical algorithm. While existing technologies for building quantum computers are in their infancy, it is not too early to consider their scalability and reliability in the context of the design of large-scale quantum computers. To architect such systems, one must understand what it takes to design and model a balanced, fault-tolerant quantum computer architecture. The goal of this lecture is to provide architectural abstractions for the design of a quantum computer and to explore
This book provides a concise introduction to Pervasive Computing, otherwise known as Internet of Things (IoT) and Ubiquitous Computing (Ubicomp) which addresses the seamless integration of computing systems within everyday objects. By introducing the core topics and exploring assistive pervasive
The book is suitable for advanced courses in computer vision and image processing. In addition to providing an overall view of computational vision, it contains extensive material on topics that are not usually covered in computer vision texts (including parallel distributed processing and neural networks) and considers many real applications.
Computation and today’s microprocessors with the approach to operating system architecture, and the controversy between microkernels and monolithic kernels...Both Spatial Computation and microkernels break away a relatively monolithic architecture into in- dividual lightweight pieces, well specialized...for their particular functionality. Spatial Computation removes global signals and control, in the same way microkernels remove the global address
Salem, A M [Computer Science Dept, Faculty of Computer and Information Sciences, Ain Shams University, Cairo (Egypt); Mohamed, A H [Solid State Dept., (NCRRT), Cairo (Egypt)
Case based reasoning (CBR) paradigm has been widely used to provide computer support for recalling and adapting known cases to novel situations. Case adaptation algorithms generally rely on knowledge based and heuristics in order to change the past solutions to solve new problems. However, case adaptation has always been a difficult process to engineers within (CBR) cycle. Its difficulties can be referred to its domain dependency; and computational cost. In an effort to solve this problem, this research explores a general-purpose method that applying a genetic algorithm (GA) to CBR adaptation. Therefore, it can decrease the computational complexity of the search space in the problems having a great dependency on their domain knowledge. The proposed model can be used to perform a variety of design tasks on a broad set of application domains. However, it has been implemented for the tablet formulation as a domain of application. The proposed system has improved the performance of the CBR design systems.
Salem, A.M.; Mohamed, A.H.
Case based reasoning (CBR) paradigm has been widely used to provide computer support for recalling and adapting known cases to novel situations. Case adaptation algorithms generally rely on knowledge based and heuristics in order to change the past solutions to solve new problems. However, case adaptation has always been a difficult process to engineers within (CBR) cycle. Its difficulties can be referred to its domain dependency; and computational cost. In an effort to solve this problem, this research explores a general-purpose method that applying a genetic algorithm (GA) to CBR adaptation. Therefore, it can decrease the computational complexity of the search space in the problems having a great dependency on their domain knowledge. The proposed model can be used to perform a variety of design tasks on a broad set of application domains. However, it has been implemented for the tablet formulation as a domain of application. The proposed system has improved the performance of the CBR design systems
There are abundant instances in the history of genetics and medical genetics to illustrate how curiosity, charisma of mentors, nature, art, the saving of lives and many other matters have inspired great discoveries. These achievements from deciphering genetic concepts to characterizing genetic disorders have been crucial for management of the patients. There remains, however, a long pathway ahead. © The Author(s) 2014.
Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn
What if people could play computer games and accomplish work without even realizing it? What if billions of people collaborated to solve important problems for humanity or generate training data for computers? My work aims at a general paradigm for doing exactly that: utilizing human processing power to solve computational problems in a distributed manner. In particular, I focus on harnessing human time and energy for addressing problems that computers cannot yet solve. Although computers have advanced dramatically in many respects over the last 50 years, they still do not possess the basic conceptual intelligence or perceptual capabilities...
As computers become ever more complex, they inevitably become smaller. This leads to a need for components which are fabricated and operate on increasingly smaller size scales. Quantum theory is already taken into account in microelectronics design. This article explores how quantum theory will need to be incorporated into computers in future in order to give them their components functionality. Computation tasks which depend on quantum effects will become possible. Physicists may have to reconsider their perspective on computation in the light of understanding developed in connection with universal quantum computers. (UK)
Rosenthal, L E
Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips.
Smith, Paul H.
The Computer Science Program provides advanced concepts, techniques, system architectures, algorithms, and software for both space and aeronautics information sciences and computer systems. The overall goal is to provide the technical foundation within NASA for the advancement of computing technology in aerospace applications. The research program is improving the state of knowledge of fundamental aerospace computing principles and advancing computing technology in space applications such as software engineering and information extraction from data collected by scientific instruments in space. The program includes the development of special algorithms and techniques to exploit the computing power provided by high performance parallel processors and special purpose architectures. Research is being conducted in the fundamentals of data base logic and improvement techniques for producing reliable computing systems.
of Computational Intelligence. First, comprehensive surveys of genetic algorithms, genetic programming, evolution strategies, parallel evolutionary algorithms are presented, which are readable and constructive so that a large audience might find them useful and – to some extent – ready to use. Some more general...... kinds of evolutionary algorithms, have been prudently analyzed. This analysis was followed by a thorough analysis of various issues involved in stochastic local search algorithms. An interesting survey of various technological and industrial applications in mechanical engineering and design has been...... topics like the estimation of distribution algorithms, indicator-based selection, etc., are also discussed. An important problem, from a theoretical and practical point of view, of learning classifier systems is presented in depth. Multiobjective evolutionary algorithms, which constitute one of the most...
... What is genetic ancestry testing? What is genetic ancestry testing? Genetic ancestry testing, or genetic genealogy, is ... with other groups. For more information about genetic ancestry testing: The University of Utah provides video tutorials ...
Computer Programming and Computer Systems imparts a "reading knowledge? of computer systems.This book describes the aspects of machine-language programming, monitor systems, computer hardware, and advanced programming that every thorough programmer should be acquainted with. This text discusses the automatic electronic digital computers, symbolic language, Reverse Polish Notation, and Fortran into assembly language. The routine for reading blocked tapes, dimension statements in subroutines, general-purpose input routine, and efficient use of memory are also elaborated.This publication is inten
Würtz, Rolf P
Organic Computing is a research field emerging around the conviction that problems of organization in complex systems in computer science, telecommunications, neurobiology, molecular biology, ethology, and possibly even sociology can be tackled scientifically in a unified way. From the computer science point of view, the apparent ease in which living systems solve computationally difficult problems makes it inevitable to adopt strategies observed in nature for creating information processing machinery. In this book, the major ideas behind Organic Computing are delineated, together with a sparse sample of computational projects undertaken in this new field. Biological metaphors include evolution, neural networks, gene-regulatory networks, networks of brain modules, hormone system, insect swarms, and ant colonies. Applications are as diverse as system design, optimization, artificial growth, task allocation, clustering, routing, face recognition, and sign language understanding.
Pataskar, Abhijeet; Tiwari, Vijay K
Cellular transcriptional programs driven by genetic and epigenetic mechanisms could be better understood by integrating "omics" data and subsequently modeling the gene-regulatory events. Toward this end, computational biology should keep pace with evolving experimental procedures and data availability. This article gives an exemplified account of the current computational challenges in molecular biology.
Computational biomechanics is a fast-growing field that integrates modern biological techniques and computer modelling to solve problems of medical and biological interest. Modelling of blood flow in the large arteries is the best-known application of computational biomechanics, but there are many others. Described here is work being carried out in the laboratory on the modelling of blood flow in the coronary arteries and on the transport of viral particles in the eye. (author)
Vallgårda, Anna K. A.
to understand the computer as a material like any other material we would use for design, like wood, aluminum, or plastic. That as soon as the computer forms a composition with other materials it becomes just as approachable and inspiring as other smart materials. I present a series of investigations of what...... Computational Composite, and Telltale). Through the investigations, I show how the computer can be understood as a material and how it partakes in a new strand of materials whose expressions come to be in context. I uncover some of their essential material properties and potential expressions. I develop a way...
Thorne, Jeffrey L; Choi, Sang Chul; Yu, Jiaye
A central goal of computational biology is the prediction of phenotype from DNA and protein sequence data. Recent models of sequence change use in silico prediction systems to incorporate the effects of phenotype on evolutionary rates. These models have been designed for analyzing sequence data...... populations, and parameters of interspecific models should have population genetic interpretations. We show, with two examples, how population genetic interpretations can be assigned to evolutionary models. The first example considers the impact of RNA secondary structure on sequence change, and the second...... reflects the tendency for protein tertiary structure to influence nonsynonymous substitution rates. We argue that statistical fit to data should not be the sole criterion for assessing models of sequence change. A good interspecific model should also yield a clear and biologically plausible population...
Jul 6, 2011 ... relationships was estimated through computer simulation and was compared with the accuracy of ... programs, detect animals with superior genetic and select ... genomic matrices in the mixed model equations of BLUP.
Jan 18, 2010 ... Key words: Genetic diversity, microsatellite markers, Caspian horse breed. INTRODUCTION ... heterozygosity, observed and effective number of alleles at each ... computer program version 1.31 (Yeh et al., 1999). Based on ...
Waters, M D; Stack, H F; Garrett, N E; Jackson, M A
A graphic approach termed a Genetic Activity Profile (GAP) has been developed to display a matrix of data on the genetic and related effects of selected chemical agents. The profiles provide a visual overview of the quantitative (doses) and qualitative (test results) data for each chemical. Either the lowest effective dose (LED) or highest ineffective dose (HID) is recorded for each agent and bioassay. Up to 200 different test systems are represented across the GAP. Bioassay systems are organized according to the phylogeny of the test organisms and the end points of genetic activity. The methodology for the production and evaluation of GAPs has been developed in collaboration with the International Agency for Research on Cancer. Data on individual chemicals have been compiled by IARC and by the U.S. Environmental Protection Agency. Data are available on 299 compounds selected from volumes 1-50 of the IARC Monographs and on 115 compounds identified as Superfund Priority Substances. Software to display the GAPs on an IBM-compatible personal computer is available from the authors. Structurally similar compounds frequently display qualitatively and quantitatively similar GAPs. By examining the patterns of GAPs of pairs and groups of chemicals, it is possible to make more informed decisions regarding the selection of test batteries to be used in evaluating chemical analogs. GAPs have provided useful data for the development of weight-of-evidence hazard ranking schemes. Also, some knowledge of the potential genetic activity of complex environmental mixtures may be gained from assessing the GAPs of component chemicals. The fundamental techniques and computer programs devised for the GAP database may be used to develop similar databases in other disciplines.
Wynne, Cynthia F.; Stewart, Jim; Passmore, Cindy
Paints a different picture of students' reasoning with meiosis as they solved complex, computer-generated genetics problems, some of which required them to revise their understanding of meiosis in response to anomalous data. Students were able to develop a rich understanding of meiosis and can utilize that knowledge to solve genetics problems.…
Full Text Available Since the first idea of using GPU to general purpose computing, things have evolved over the years and now there are several approaches to GPU programming. GPU computing practically began with the introduction of CUDA (Compute Unified Device Architecture by NVIDIA and Stream by AMD. These are APIs designed by the GPU vendors to be used together with the hardware that they provide. A new emerging standard, OpenCL (Open Computing Language tries to unify different GPU general computing API implementations and provides a framework for writing programs executed across heterogeneous platforms consisting of both CPUs and GPUs. OpenCL provides parallel computing using task-based and data-based parallelism. In this paper we will focus on the CUDA parallel computing architecture and programming model introduced by NVIDIA. We will present the benefits of the CUDA programming model. We will also compare the two main approaches, CUDA and AMD APP (STREAM and the new framwork, OpenCL that tries to unify the GPGPU computing models.
Home; Journals; Resonance – Journal of Science Education; Volume 5; Issue 9. Quantum Computing - Building Blocks of a Quantum Computer. C S Vijay Vishal Gupta. General Article Volume 5 Issue 9 September 2000 pp 69-81. Fulltext. Click here to view fulltext PDF. Permanent link:
"Platform Computing releases first grid-enabled workload management solution for IBM eServer Intel and UNIX high performance computing clusters. This Out-of-the-box solution maximizes the performance and capability of applications on IBM HPC clusters" (1/2 page) .
In the first part of this article, we had looked at how quantum physics can be harnessed to make the building blocks of a quantum computer. In this concluding part, we look at algorithms which can exploit the power of this computational device, and some practical difficulties in building such a device. Quantum Algorithms.
Burba, M.; Lapitskaya, T.
This article gives an elementary introduction to quantum computing. It is a draft for a book chapter of the "Handbook of Nature-Inspired and Innovative Computing", Eds. A. Zomaya, G.J. Milburn, J. Dongarra, D. Bader, R. Brent, M. Eshaghian-Wilner, F. Seredynski (Springer, Berlin Heidelberg New York, 2006).
Louis, David N.; Feldman, Michael; Carter, Alexis B.; Dighe, Anand S.; Pfeifer, John D.; Bry, Lynn; Almeida, Jonas S.; Saltz, Joel; Braun, Jonathan; Tomaszewski, John E.; Gilbertson, John R.; Sinard, John H.; Gerber, Georg K.; Galli, Stephen J.; Golden, Jeffrey A.; Becich, Michael J.
Context We define the scope and needs within the new discipline of computational pathology, a discipline critical to the future of both the practice of pathology and, more broadly, medical practice in general. Objective To define the scope and needs of computational pathology. Data Sources A meeting was convened in Boston, Massachusetts, in July 2014 prior to the annual Association of Pathology Chairs meeting, and it was attended by a variety of pathologists, including individuals highly invested in pathology informatics as well as chairs of pathology departments. Conclusions The meeting made recommendations to promote computational pathology, including clearly defining the field and articulating its value propositions; asserting that the value propositions for health care systems must include means to incorporate robust computational approaches to implement data-driven methods that aid in guiding individual and population health care; leveraging computational pathology as a center for data interpretation in modern health care systems; stating that realizing the value proposition will require working with institutional administrations, other departments, and pathology colleagues; declaring that a robust pipeline should be fostered that trains and develops future computational pathologists, for those with both pathology and non-pathology backgrounds; and deciding that computational pathology should serve as a hub for data-related research in health care systems. The dissemination of these recommendations to pathology and bioinformatics departments should help facilitate the development of computational pathology. PMID:26098131
with technological changes, the paradigmatic pendulum has swung between increased centralization on one side and a focus on distributed computing that pushes IT power out to end users on the other. With the introduction of outsourcing and cloud computing, centralization in large data centers is again dominating...... the IT scene. In line with the views presented by Nicolas Carr in 2003 (Carr, 2003), it is a popular assumption that cloud computing will be the next utility (like water, electricity and gas) (Buyya, Yeo, Venugopal, Broberg, & Brandic, 2009). However, this assumption disregards the fact that most IT production......), for instance, in establishing and maintaining trust between the involved parties (Sabherwal, 1999). So far, research in cloud computing has neglected this perspective and focused entirely on aspects relating to technology, economy, security and legal questions. While the core technologies of cloud computing (e...
What can we compute--even with unlimited resources? Is everything within reach? Or are computations necessarily drastically limited, not just in practice, but theoretically? These questions are at the heart of computability theory. The goal of this book is to give the reader a firm grounding in the fundamentals of computability theory and an overview of currently active areas of research, such as reverse mathematics and algorithmic randomness. Turing machines and partial recursive functions are explored in detail, and vital tools and concepts including coding, uniformity, and diagonalization are described explicitly. From there the material continues with universal machines, the halting problem, parametrization and the recursion theorem, and thence to computability for sets, enumerability, and Turing reduction and degrees. A few more advanced topics round out the book before the chapter on areas of research. The text is designed to be self-contained, with an entire chapter of preliminary material including re...
Paul M. Torrens
Full Text Available Streetscapes have presented a long-standing interest in many fields. Recently, there has been a resurgence of attention on streetscape issues, catalyzed in large part by computing. Because of computing, there is more understanding, vistas, data, and analysis of and on streetscape phenomena than ever before. This diversity of lenses trained on streetscapes permits us to address long-standing questions, such as how people use information while mobile, how interactions with people and things occur on streets, how we might safeguard crowds, how we can design services to assist pedestrians, and how we could better support special populations as they traverse cities. Amid each of these avenues of inquiry, computing is facilitating new ways of posing these questions, particularly by expanding the scope of what-if exploration that is possible. With assistance from computing, consideration of streetscapes now reaches across scales, from the neurological interactions that form among place cells in the brain up to informatics that afford real-time views of activity over whole urban spaces. For some streetscape phenomena, computing allows us to build realistic but synthetic facsimiles in computation, which can function as artificial laboratories for testing ideas. In this paper, I review the domain science for studying streetscapes from vantages in physics, urban studies, animation and the visual arts, psychology, biology, and behavioral geography. I also review the computational developments shaping streetscape science, with particular emphasis on modeling and simulation as informed by data acquisition and generation, data models, path-planning heuristics, artificial intelligence for navigation and way-finding, timing, synthetic vision, steering routines, kinematics, and geometrical treatment of collision detection and avoidance. I also discuss the implications that the advances in computing streetscapes might have on emerging developments in cyber
Liu, Ren-Hu; Meng, Jin-Ling
MAPMAKER is one of the most widely used computer software package for constructing genetic linkage maps.However, the PC version, MAPMAKER 3.0 for PC, could not draw the genetic linkage maps that its Macintosh version, MAPMAKER 3.0 for Macintosh,was able to do. Especially in recent years, Macintosh computer is much less popular than PC. Most of the geneticists use PC to analyze their genetic linkage data. So a new computer software to draw the same genetic linkage maps on PC as the MAPMAKER for Macintosh to do on Macintosh has been crying for. Microsoft Excel,one component of Microsoft Office package, is one of the most popular software in laboratory data processing. Microsoft Visual Basic for Applications (VBA) is one of the most powerful functions of Microsoft Excel. Using this program language, we can take creative control of Excel, including genetic linkage map construction, automatic data processing and more. In this paper, a Microsoft Excel macro called MapDraw is constructed to draw genetic linkage maps on PC computer based on given genetic linkage data. Use this software,you can freely construct beautiful genetic linkage map in Excel and freely edit and copy it to Word or other application. This software is just an Excel format file. You can freely copy it from ftp://184.108.40.206 or ftp://brassica.hzau.edu.cn and the source code can be found in Excel's Visual Basic Editor.
... view the expand/collapse boxes. Description SADDAN (severe achondroplasia with developmental delay and acanthosis nigricans) is a ... Genetic Testing (1 link) Genetic Testing Registry: Severe achondroplasia with developmental delay and acanthosis nigricans Other Diagnosis ...
A genetic brain disorder is caused by a variation or a mutation in a gene. A variation is a different form ... mutation is a change in a gene. Genetic brain disorders affect the development and function of the ...
... genetic counselor can help you work through the pros and cons of genetic testing based on your ... showing symptoms or what their progression will be. Technology is changing rapidly and costs of testing are ...
Bioengineered foods; GMOs; Genetically modified foods ... helps speed up the process of creating new foods with desired traits. The possible benefits of genetic engineering include: More nutritious food Tastier food Disease- and ...
Alderson, P; Aro, A R; Dragonas, T
Although the term 'genetic screening' has been used for decades, this paper discusses how, in its most precise meaning, genetic screening has not yet been widely introduced. 'Prenatal screening' is often confused with 'genetic screening'. As we show, these terms have different meanings, and we...... examine definitions of the relevant concepts in order to illustrate this point. The concepts are i) prenatal, ii) genetic screening, iii) screening, scanning and testing, iv) maternal and foetal tests, v) test techniques and vi) genetic conditions. So far, prenatal screening has little connection...... with precisely defined genetics. There are benefits but also disadvantages in overstating current links between them in the term genetic screening. Policy making and professional and public understandings about screening could be clarified if the distinct meanings of prenatal screening and genetic screening were...
This book assesses the scientific value and merit of research on human genetic differences--including a collection of DNA samples that represents the whole of human genetic diversity--and the ethical...
... A characteristic of X-linked inheritance is that fathers cannot pass X-linked traits to their sons. ... infantile neuroaxonal dystrophy Genetic Testing Registry: Osteopetrosis autosomal dominant type 1 Genetic Testing Registry: Osteopetrosis autosomal dominant ...
Carter, C. O.
Can genetic evolution be controlled by man in a manner which does not violate a civilized, humane, and democratic ethos? The genetics of health and illhealth and of normal variation are discussed with respect to this question. (PEB)
Genetic Science Learning Center Making science and health easy for everyone to understand Home News Our Team What We Do ... Collaboration Conferences Current Projects Publications Contact The Genetic Science Learning Center at The University of Utah is a ...
... an increased risk of abnormal blood clotting, and brittle bones that are prone to fracture ( osteoporosis ) or other ... information about a genetic condition can statistics provide? Why are some genetic conditions more common in particular ...
Evgeniy K. Khenner
Full Text Available Abstract. The aim of the research is to draw attention of the educational community to the phenomenon of computational thinking which actively discussed in the last decade in the foreign scientific and educational literature, to substantiate of its importance, practical utility and the right on affirmation in Russian education.Methods. The research is based on the analysis of foreign studies of the phenomenon of computational thinking and the ways of its formation in the process of education; on comparing the notion of «computational thinking» with related concepts used in the Russian scientific and pedagogical literature.Results. The concept «computational thinking» is analyzed from the point of view of intuitive understanding and scientific and applied aspects. It is shown as computational thinking has evolved in the process of development of computers hardware and software. The practice-oriented interpretation of computational thinking which dominant among educators is described along with some ways of its formation. It is shown that computational thinking is a metasubject result of general education as well as its tool. From the point of view of the author, purposeful development of computational thinking should be one of the tasks of the Russian education.Scientific novelty. The author gives a theoretical justification of the role of computational thinking schemes as metasubject results of learning. The dynamics of the development of this concept is described. This process is connected with the evolution of computer and information technologies as well as increase of number of the tasks for effective solutions of which computational thinking is required. Author substantiated the affirmation that including «computational thinking » in the set of pedagogical concepts which are used in the national education system fills an existing gap.Practical significance. New metasubject result of education associated with
Roche, P A; Annas, G J
This article outlines the arguments for and against new rules to protect genetic privacy. We explain why genetic information is different to other sensitive medical information, why researchers and biotechnology companies have opposed new rules to protect genetic privacy (and favour anti-discrimination laws instead), and discuss what can be done to protect privacy in relation to genetic-sequence information and to DNA samples themselves.
Mackenzie J. Lind; Philip R. Gehrman
This review summarizes current research on the genetics of insomnia, as genetic contributions are thought to be important for insomnia etiology. We begin by providing an overview of genetic methods (both quantitative and measured gene), followed by a discussion of the insomnia genetics literature with regard to each of the following common methodologies: twin and family studies, candidate gene studies, and genome-wide association studies (GWAS). Next, we summarize the most recent gene identif...
In the newly emerging debates about genetics and justice three distinct principles have begun to emerge concerning what the distributive aim of genetic interventions should be. These principles are: genetic equality, a genetic decent minimum, and the genetic difference principle. In this paper, I examine the rationale of each of these principles and argue that genetic equality and a genetic decent minimum are ill-equipped to tackle what I call the currency problem and the problem of weight. The genetic difference principle is the most promising of the three principles and I develop this principle so that it takes seriously the concerns of just health care and distributive justice in general. Given the strains on public funds for other important social programmes, the costs of pursuing genetic interventions and the nature of genetic interventions, I conclude that a more lax interpretation of the genetic difference principle is appropriate. This interpretation stipulates that genetic inequalities should be arranged so that they are to the greatest reasonable benefit of the least advantaged. Such a proposal is consistent with prioritarianism and provides some practical guidance for non-ideal societies--that is, societies that do not have the endless amount of resources needed to satisfy every requirement of justice.
After agreeing to host over 200 students on a daylong genetics field trip, the author needed an easy-to-prepare genetics experiment to accompany the DNA-necklace and gel-electrophoresis activities already planned. One of the student's mothers is a pediatric physician at the local hospital, and she suggested exploring genetic-disease screening…
... Page Search Home Health Conditions Genes Chromosomes & mtDNA Resources Help Me Understand Genetics Share: Email Facebook Twitter Genetics Home Reference provides consumer-friendly information about the effects of genetic variation on human health. Health Conditions More than 1,200 health ...
Full Text Available Coupling dense genotype data with new computational methods offers unprecedented opportunities for individual-level ancestry estimation once geographically precisely defined reference data sets become available. We study such a reference data set for Finland containing 2376 such individuals from the FINRISK Study survey of 1997 both of whose parents were born close to each other. This sampling strategy focuses on the population structure present in Finland before the 1950s. By using the recent haplotype-based methods ChromoPainter (CP and FineSTRUCTURE (FS we reveal a highly geographically clustered genetic structure in Finland and report its connections to the settlement history as well as to the current dialectal regions of the Finnish language. The main genetic division within Finland shows striking concordance with the 1323 borderline of the treaty of Nöteborg. In general, we detect genetic substructure throughout the country, which reflects stronger regional genetic differences in Finland compared to, for example, the UK, which in a similar analysis was dominated by a single unstructured population. We expect that similar population genetic reference data sets will become available for many more populations in the near future with important applications, for example, in forensic genetics and in genetic association studies. With this in mind, we report those extensions of the CP + FS approach that we found most useful in our analyses of the Finnish data.
Lyons, Leslie A
DNA testing for domestic cat diseases and appearance traits is a rapidly growing asset for veterinary medicine. Approximately 33 genes contain 50 mutations that cause feline health problems or alterations in the cat's appearance. A variety of commercial laboratories can now perform cat genetic diagnostics, allowing both the veterinary clinician and the private owner to obtain DNA test results. DNA is easily obtained from a cat via a buccal swab with a standard cotton bud or cytological brush, allowing DNA samples to be easily sent to any laboratory in the world. The DNA test results identify carriers of the traits, predict the incidence of traits from breeding programs, and influence medical prognoses and treatments. An overall goal of identifying these genetic mutations is the correction of the defect via gene therapies and designer drug therapies. Thus, genetic testing is an effective preventative medicine and a potential ultimate cure. However, genetic diagnostic tests may still be novel for many veterinary practitioners and their application in the clinical setting needs to have the same scrutiny as any other diagnostic procedure. This article will review the genetic tests for the domestic cat, potential sources of error for genetic testing, and the pros and cons of DNA results in veterinary medicine. Highlighted are genetic tests specific to the individual cat, which are a part of the cat's internal genome. Copyright © 2010 Elsevier Inc. All rights reserved.
This book explains how computers interact with the world around them and therefore how to make them a useful tool. Topics covered include descriptions of all the components that make up a computer, principles of data exchange, interaction with peripherals, serial communication, input devices, recording methods, computer-controlled motors, and printers.In an informative and straightforward manner, Graham Dixey describes how to turn what might seem an incomprehensible 'black box' PC into a powerful and enjoyable tool that can help you in all areas of your work and leisure. With plenty of handy
A complete introduction to the field of computational physics, with examples and exercises in the Python programming language. Computers play a central role in virtually every major physics discovery today, from astrophysics and particle physics to biophysics and condensed matter. This book explains the fundamentals of computational physics and describes in simple terms the techniques that every physicist should know, such as finite difference methods, numerical quadrature, and the fast Fourier transform. The book offers a complete introduction to the topic at the undergraduate level, and is also suitable for the advanced student or researcher who wants to learn the foundational elements of this important field.
Computers have for many years played a vital role in the acquisition and treatment of experimental data, but they have more recently taken up a much more extended role in physics research. The numerical and algebraic calculations now performed on modern computers make it possible to explore consequences of basic theories in a way which goes beyond the limits of both analytic insight and experimental investigation. This was brought out clearly at the Conference on Perspectives in Computational Physics, held at the International Centre for Theoretical Physics, Trieste, Italy, from 29-31 October.
Baun, Christian; Nimis, Jens; Tai, Stefan
Cloud computing is a buzz-word in today's information technology (IT) that nobody can escape. But what is really behind it? There are many interpretations of this term, but no standardized or even uniform definition. Instead, as a result of the multi-faceted viewpoints and the diverse interests expressed by the various stakeholders, cloud computing is perceived as a rather fuzzy concept. With this book, the authors deliver an overview of cloud computing architecture, services, and applications. Their aim is to bring readers up to date on this technology and thus to provide a common basis for d
Marques, Severino P C
This text is a guide how to solve problems in which viscoelasticity is present using existing commercial computational codes. The book gives information on codes’ structure and use, data preparation and output interpretation and verification. The first part of the book introduces the reader to the subject, and to provide the models, equations and notation to be used in the computational applications. The second part shows the most important Computational techniques: Finite elements formulation, Boundary elements formulation, and presents the solutions of Viscoelastic problems with Abaqus.
Stroke, G. W.
Applications of the optical computer include an approach for increasing the sharpness of images obtained from the most powerful electron microscopes and fingerprint/credit card identification. The information-handling capability of the various optical computing processes is very great. Modern synthetic-aperture radars scan upward of 100,000 resolvable elements per second. Fields which have assumed major importance on the basis of optical computing principles are optical image deblurring, coherent side-looking synthetic-aperture radar, and correlative pattern recognition. Some examples of the most dramatic image deblurring results are shown.
Computers have for many years played a vital role in the acquisition and treatment of experimental data, but they have more recently taken up a much more extended role in physics research. The numerical and algebraic calculations now performed on modern computers make it possible to explore consequences of basic theories in a way which goes beyond the limits of both analytic insight and experimental investigation. This was brought out clearly at the Conference on Perspectives in Computational Physics, held at the International Centre for Theoretical Physics, Trieste, Italy, from 29-31 October
Open peer commentary on the article “Info-computational Constructivism and Cognition” by Gordana Dodig-Crnkovic. Upshot: The main problems with info-computationalism are: (1) Its basic concept of natural computing has neither been defined theoretically or implemented practically. (2. It cannot...... encompass human concepts of subjective experience and intersubjective meaningful communication, which prevents it from being genuinely transdisciplinary. (3) Philosophically, it does not sufficiently accept the deep ontological differences between various paradigms such as von Foerster’s second- order...
Sullivan-Pyke, Chantae; Dokras, Anuja
Preimplantation genetic testing encompasses preimplantation genetic screening (PGS) and preimplantation genetic diagnosis (PGD). PGS improves success rates of in vitro fertilization by ensuring the transfer of euploid embryos that have a higher chance of implantation and resulting in a live birth. PGD enables the identification of embryos with specific disease-causing mutations and transfer of unaffected embryos. The development of whole genome amplification and genomic tools, including single nucleotide polymorphism microarrays, comparative genomic hybridization microarrays, and next-generation sequencing, has led to faster, more accurate diagnoses that translate to improved pregnancy and live birth rates. Copyright © 2017 Elsevier Inc. All rights reserved.
ForewordPrefaceComputing ParadigmsLearning ObjectivesPreambleHigh-Performance ComputingParallel ComputingDistributed ComputingCluster ComputingGrid ComputingCloud ComputingBiocomputingMobile ComputingQuantum ComputingOptical ComputingNanocomputingNetwork ComputingSummaryReview PointsReview QuestionsFurther ReadingCloud Computing FundamentalsLearning ObjectivesPreambleMotivation for Cloud ComputingThe Need for Cloud ComputingDefining Cloud ComputingNIST Definition of Cloud ComputingCloud Computing Is a ServiceCloud Computing Is a Platform5-4-3 Principles of Cloud computingFive Essential Charact
Toong, Hoo-min D.; Gupta, Amar
Describes the hardware, software, applications, and current proliferation of personal computers (microcomputers). Includes discussions of microprocessors, memory, output (including printers), application programs, the microcomputer industry, and major microcomputer manufacturers (Apple, Radio Shack, Commodore, and IBM). (JN)
Chongtay, Rocio; Robering, Klaus
In recent years, there has been a growing interest in and recognition of the importance of Computational Literacy, a skill generally considered to be necessary for success in the 21st century. While much research has concentrated on requirements, tools, and teaching methodologies for the acquisit......In recent years, there has been a growing interest in and recognition of the importance of Computational Literacy, a skill generally considered to be necessary for success in the 21st century. While much research has concentrated on requirements, tools, and teaching methodologies...... for the acquisition of Computational Literacy at basic educational levels, focus on higher levels of education has been much less prominent. The present paper considers the case of courses for higher education programs within the Humanities. A model is proposed which conceives of Computational Literacy as a layered...
Nielbo, Kristoffer Laigaard; Braxton, Donald M.; Upal, Afzal
The computational approach has become an invaluable tool in many fields that are directly relevant to research in religious phenomena. Yet the use of computational tools is almost absent in the study of religion. Given that religion is a cluster of interrelated phenomena and that research...... concerning these phenomena should strive for multilevel analysis, this article argues that the computational approach offers new methodological and theoretical opportunities to the study of religion. We argue that the computational approach offers 1.) an intermediary step between any theoretical construct...... and its targeted empirical space and 2.) a new kind of data which allows the researcher to observe abstract constructs, estimate likely outcomes, and optimize empirical designs. Because sophisticated mulitilevel research is a collaborative project we also seek to introduce to scholars of religion some...
Timmermans, Benjamin; Kuhn, Tobias; Beelen, Kaspar; Aroyo, Lora
Climate change, vaccination, abortion, Trump: Many topics are surrounded by fierce controversies. The nature of such heated debates and their elements have been studied extensively in the social science literature. More recently, various computational approaches to controversy analysis have
emergence of supercomputers led to the use of computer simula- tion as an .... Scientific and engineering applications (e.g., Tera grid secure gate way). Collaborative ... Encryption, privacy, protection from malicious software. Physical Layer.
... consumer genetic testing? What kinds of direct-to-consumer genetic tests are available? What is genetic ancestry testing? What are the benefits and risks of direct-to-consumer genetic testing? ...
... Videos for Educators Search English Español Prenatal Genetic Counseling KidsHealth / For Parents / Prenatal Genetic Counseling What's in ... can they help your family? What Is Genetic Counseling? Genetic counseling is the process of: evaluating family ...
... Videos for Educators Search English Español All About Genetics KidsHealth / For Parents / All About Genetics What's in ... the way they pick up special laboratory dyes. Genetic Problems Errors in the genetic code or "gene ...
Physical foundations and the developments in the transmission and emission computer tomography are presented. On the basis of the available literature and private communications a comparison is made of the various transmission tomographs. A new technique of computer emission tomography ECT, unknown in Poland, is described. The evaluation of two methods of ECT, namely those of positron and single photon emission tomography is made. (author)
Kersting, Kristian; Morik, Katharina
The book at hand gives an overview of the state of the art research in Computational Sustainability as well as case studies of different application scenarios. This covers topics such as renewable energy supply, energy storage and e-mobility, efficiency in data centers and networks, sustainable food and water supply, sustainable health, industrial production and quality, etc. The book describes computational methods and possible application scenarios.
High-energy physics, nuclear physics, space sciences, and many other fields have large challenges in computing. In recent years, PCs have achieved performance comparable to the high-end UNIX workstations, at a small fraction of the price. We review the development and broad applications of commodity PCs as the solution to CPU needs, and look forward to the important and exciting future of large-scale PC computing
This report describes the methodology for using a genetic programming model to develop tracking behaviors for autonomous, microscale robotic vehicles. The use of such vehicles for surveillance and detection operations has become increasingly important in defense and humanitarian applications. Through an evolutionary process similar to that found in nature, the genetic programming model generates a computer program that when downloaded onto a robotic vehicle's on-board computer will guide the robot to successfully accomplish its task. Simulations of multiple robots engaged in problem-solving tasks have demonstrated cooperative behaviors. This report also discusses the behavior model produced by genetic programming and presents some results achieved during the study
Arnold, J. O.
With the advent of supercomputers, modern computational chemistry algorithms and codes, a powerful tool was created to help fill NASA's continuing need for information on the properties of matter in hostile or unusual environments. Computational resources provided under the National Aerodynamics Simulator (NAS) program were a cornerstone for recent advancements in this field. Properties of gases, materials, and their interactions can be determined from solutions of the governing equations. In the case of gases, for example, radiative transition probabilites per particle, bond-dissociation energies, and rates of simple chemical reactions can be determined computationally as reliably as from experiment. The data are proving to be quite valuable in providing inputs to real-gas flow simulation codes used to compute aerothermodynamic loads on NASA's aeroassist orbital transfer vehicles and a host of problems related to the National Aerospace Plane Program. Although more approximate, similar solutions can be obtained for ensembles of atoms simulating small particles of materials with and without the presence of gases. Computational chemistry has application in studying catalysis, properties of polymers, all of interest to various NASA missions, including those previously mentioned. In addition to discussing these applications of computational chemistry within NASA, the governing equations and the need for supercomputers for their solution is outlined.
Heba Sh. Kassem
Full Text Available Genetics have undoubtedly become an integral part of biomedical science and clinical practice, with important implications in deciphering disease pathogenesis and progression, identifying diagnostic and prognostic markers, as well as designing better targeted treatments. The exponential growth of our understanding of different genetic concepts is paralleled by a growing list of genetic terminology that can easily intimidate the unfamiliar reader. Rendering genetics incomprehensible to the clinician however, defeats the very essence of genetic research: its utilization for combating disease and improving quality of life. Herein we attempt to correct this notion by presenting the basic genetic concepts along with their usefulness in the cardiology clinic. Bringing genetics closer to the clinician will enable its harmonious incorporation into clinical care, thus not only restoring our perception of its simple and elegant nature, but importantly ensuring the maximal benefit for our patients.
Kassem, Heba Sh.; Girolami, Francesca; Sanoudou, Despina
Abstract Genetics have undoubtedly become an integral part of biomedical science and clinical practice, with important implications in deciphering disease pathogenesis and progression, identifying diagnostic and prognostic markers, as well as designing better targeted treatments. The exponential growth of our understanding of different genetic concepts is paralleled by a growing list of genetic terminology that can easily intimidate the unfamiliar reader. Rendering genetics incomprehensible to the clinician however, defeats the very essence of genetic research: its utilization for combating disease and improving quality of life. Herein we attempt to correct this notion by presenting the basic genetic concepts along with their usefulness in the cardiology clinic. Bringing genetics closer to the clinician will enable its harmonious incorporation into clinical care, thus not only restoring our perception of its simple and elegant nature, but importantly ensuring the maximal benefit for our patients. PMID:25610837
National Oceanic and Atmospheric Administration, Department of Commerce — Initiated in 1989, this study monitors genetic changes associated with hatchery propagation in multiple Snake River sub-basins for Chinook salmon and steelhead. We...
Lee, Anthony J; Mitchem, Dorian G; Wright, Margaret J; Martin, Nicholas G; Keller, Matthew C; Zietsch, Brendan P
Popular theory suggests that facial averageness is preferred in a partner for genetic benefits to offspring. However, whether facial averageness is associated with genetic quality is yet to be established. Here, we computed an objective measure of facial averageness for a large sample ( N = 1,823) of identical and nonidentical twins and their siblings to test two predictions from the theory that facial averageness reflects genetic quality. First, we use biometrical modelling to estimate the heritability of facial averageness, which is necessary if it reflects genetic quality. We also test for a genetic association between facial averageness and facial attractiveness. Second, we assess whether paternal age at conception (a proxy of mutation load) is associated with facial averageness and facial attractiveness. Our findings are mixed with respect to our hypotheses. While we found that facial averageness does have a genetic component, and a significant phenotypic correlation exists between facial averageness and attractiveness, we did not find a genetic correlation between facial averageness and attractiveness (therefore, we cannot say that the genes that affect facial averageness also affect facial attractiveness) and paternal age at conception was not negatively associated with facial averageness. These findings support some of the previously untested assumptions of the 'genetic benefits' account of facial averageness, but cast doubt on others.
Full Text Available This paper discusses the application of Robust Hybrid Genetic Algorithm to solve a flow-shop scheduling problem. The proposed algorithm attempted to reach minimum makespan. PT. FSCM Manufacturing Indonesia Plant 4's case was used as a test case to evaluate the performance of the proposed algorithm. The proposed algorithm was compared to Ant Colony, Genetic-Tabu, Hybrid Genetic Algorithm, and the company's algorithm. We found that Robust Hybrid Genetic produces statistically better result than the company's, but the same as Ant Colony, Genetic-Tabu, and Hybrid Genetic. In addition, Robust Hybrid Genetic Algorithm required less computational time than Hybrid Genetic Algorithm
Pereira, C.M.N.A.; Schirru, R.; Martinez, A.S.
A genetic algorithm is a powerful search technique that simulates natural evolution in order to fit a population of computational structures to the solution of an optimization problem. This technique presents several advantages over classical ones such as linear programming based techniques, often used in nuclear engineering optimization problems. However, genetic algorithms demand some extra computational cost. Nowadays, due to the fast computers available, the use of genetic algorithms has increased and its practical application has become a reality. In nuclear engineering there are many difficult optimization problems related to nuclear reactor design. Genetic algorithm is a suitable technique to face such kind of problems. This chapter presents applications of genetic algorithms for nuclear reactor core design optimization. A genetic algorithm has been designed to optimize the nuclear reactor cell parameters, such as array pitch, isotopic enrichment, dimensions and cells materials. Some advantages of this genetic algorithm implementation over a classical method based on linear programming are revealed through the application of both techniques to a simple optimization problem. In order to emphasize the suitability of genetic algorithms for design optimization, the technique was successfully applied to a more complex problem, where the classical method is not suitable. Results and comments about the applications are also presented. (orig.)
López de Mántaras Badia, Ramon
Full Text Available New technologies, and in particular artificial intelligence, are drastically changing the nature of creative processes. Computers are playing very significant roles in creative activities such as music, architecture, fine arts, and science. Indeed, the computer is already a canvas, a brush, a musical instrument, and so on. However, we believe that we must aim at more ambitious relations between computers and creativity. Rather than just seeing the computer as a tool to help human creators, we could see it as a creative entity in its own right. This view has triggered a new subfield of Artificial Intelligence called Computational Creativity. This article addresses the question of the possibility of achieving computational creativity through some examples of computer programs capable of replicating some aspects of creative behavior in the fields of music and science.Las nuevas tecnologías y en particular la Inteligencia Artificial están cambiando de forma importante la naturaleza del proceso creativo. Los ordenadores están jugando un papel muy significativo en actividades artísticas tales como la música, la arquitectura, las bellas artes y la ciencia. Efectivamente, el ordenador ya es el lienzo, el pincel, el instrumento musical, etc. Sin embargo creemos que debemos aspirar a relaciones más ambiciosas entre los ordenadores y la creatividad. En lugar de verlos solamente como herramientas de ayuda a la creación, los ordenadores podrían ser considerados agentes creativos. Este punto de vista ha dado lugar a un nuevo subcampo de la Inteligencia Artificial denominado Creatividad Computacional. En este artículo abordamos la cuestión de la posibilidad de alcanzar dicha creatividad computacional mediante algunos ejemplos de programas de ordenador capaces de replicar algunos aspectos relacionados con el comportamiento creativo en los ámbitos de la música y la ciencia.
Full Text Available Abstract Background While the ultimate causes of most species extinctions are environmental, environmental constraints have various secondary consequences on evolutionary and ecological processes. The roles of demographic, genetic mechanisms and their interactions in limiting the viabilities of species or populations have stirred much debate and remain difficult to evaluate in the absence of demography-genetics conceptual and technical framework. Here, I computed projected times to metapopulation extinction using (1 a model focusing on the effects of species properties, habitat quality, quantity and temporal variability on the time to demographic extinction; (2 a genetic model focusing on the dynamics of the drift and inbreeding loads under the same species and habitat constraints; (3 a demo-genetic model accounting for demographic-genetic processes and feedbacks. Results Results indicate that a given population may have a high demographic, but low genetic viability or vice versa; and whether genetic or demographic aspects will be the most limiting to overall viability depends on the constraints faced by the species (e.g., reduction of habitat quantity or quality. As a consequence, depending on metapopulation or species characteristics, incorporating genetic considerations to demographically-based viability assessments may either moderately or severely reduce the persistence time. On the other hand, purely genetically-based estimates of species viability may either underestimate (by neglecting demo-genetic interactions or overestimate (by neglecting the demographic resilience true viability. Conclusion Unbiased assessments of the viabilities of species may only be obtained by identifying and considering the most limiting processes (i.e., demography or genetics, or, preferentially, by integrating them.
The interface of Operation Research and Computer Science - although elusive to a precise definition - has been a fertile area of both methodological and applied research. The papers in this book, written by experts in their respective fields, convey the current state-of-the-art in this interface across a broad spectrum of research domains which include optimization techniques, linear programming, interior point algorithms, networks, computer graphics in operations research, parallel algorithms and implementations, planning and scheduling, genetic algorithms, heuristic search techniques and dat
Don E. Riemenschneider
The genetic and economic effects of two stages of truncation selection in a white spruce seedling orchard were investigated by computer simulation. Genetic effects were computed by assuming a bivariate distribution of juvenile and mature traits and volume was used as the selection criterion. Seed production was assumed to rise in a linear fashion to maturity and then...
Casillas, Sònia; Barbadilla, Antonio
Molecular population genetics aims to explain genetic variation and molecular evolution from population genetics principles. The field was born 50 years ago with the first measures of genetic variation in allozyme loci, continued with the nucleotide sequencing era, and is currently in the era of population genomics. During this period, molecular population genetics has been revolutionized by progress in data acquisition and theoretical developments. The conceptual elegance of the neutral theory of molecular evolution or the footprint carved by natural selection on the patterns of genetic variation are two examples of the vast number of inspiring findings of population genetics research. Since the inception of the field, Drosophila has been the prominent model species: molecular variation in populations was first described in Drosophila and most of the population genetics hypotheses were tested in Drosophila species. In this review, we describe the main concepts, methods, and landmarks of molecular population genetics, using the Drosophila model as a reference. We describe the different genetic data sets made available by advances in molecular technologies, and the theoretical developments fostered by these data. Finally, we review the results and new insights provided by the population genomics approach, and conclude by enumerating challenges and new lines of inquiry posed by increasingly large population scale sequence data. Copyright © 2017 Casillas and Barbadilla.
Cios, K. J.; Berke, L.; Vary, A.; Sharma, S.
Soft computing techniques of neural networks and genetic algorithms are used in the design of superalloys. The cyclic oxidation attack parameter K(sub a), generated from tests at NASA Lewis Research Center, is modelled as a function of the superalloy chemistry and test temperature using a neural network. This model is then used in conjunction with a genetic algorithm to obtain an optimized superalloy composition resulting in low K(sub a) values.
Chuang, Chia-Hua; Lin, Chun-Liang
Rhythmic clock widely occurs in biological systems which controls several aspects of cell physiology. For the different cell types, it is supplied with various rhythmic frequencies. How to synthesize a specific clock signal is a preliminary but a necessary step to further development of a biological computer in the future. This paper presents a genetic sequential logic circuit with a clock pulse generator based on a synthesized genetic oscillator, which generates a consecutive clock signal whose frequency is an inverse integer multiple to that of the genetic oscillator. An analogous electronic waveform-shaping circuit is constructed by a series of genetic buffers to shape logic high/low levels of an oscillation input in a basic sinusoidal cycle and generate a pulse-width-modulated (PWM) output with various duty cycles. By controlling the threshold level of the genetic buffer, a genetic clock pulse signal with its frequency consistent to the genetic oscillator is synthesized. A synchronous genetic counter circuit based on the topology of the digital sequential logic circuit is triggered by the clock pulse to synthesize the clock signal with an inverse multiple frequency to the genetic oscillator. The function acts like a frequency divider in electronic circuits which plays a key role in the sequential logic circuit with specific operational frequency. A cascaded genetic logic circuit generating clock pulse signals is proposed. Based on analogous implement of digital sequential logic circuits, genetic sequential logic circuits can be constructed by the proposed approach to generate various clock signals from an oscillation signal.
The subject of quantum computing brings together ideas from classical information theory, computer science, and quantum physics. This review aims to summarize not just quantum computing, but the whole subject of quantum information theory. Information can be identified as the most general thing which must propagate from a cause to an effect. It therefore has a fundamentally important role in the science of physics. However, the mathematical treatment of information, especially information processing, is quite recent, dating from the mid-20th century. This has meant that the full significance of information as a basic concept in physics is only now being discovered. This is especially true in quantum mechanics. The theory of quantum information and computing puts this significance on a firm footing, and has led to some profound and exciting new insights into the natural world. Among these are the use of quantum states to permit the secure transmission of classical information (quantum cryptography), the use of quantum entanglement to permit reliable transmission of quantum states (teleportation), the possibility of preserving quantum coherence in the presence of irreversible noise processes (quantum error correction), and the use of controlled quantum evolution for efficient computation (quantum computation). The common theme of all these insights is the use of quantum entanglement as a computational resource. It turns out that information theory and quantum mechanics fit together very well. In order to explain their relationship, this review begins with an introduction to classical information theory and computer science, including Shannon's theorem, error correcting codes, Turing machines and computational complexity. The principles of quantum mechanics are then outlined, and the Einstein, Podolsky and Rosen (EPR) experiment described. The EPR-Bell correlations, and quantum entanglement in general, form the essential new ingredient which distinguishes quantum from
Steane, Andrew [Department of Atomic and Laser Physics, University of Oxford, Clarendon Laboratory, Oxford (United Kingdom)
The subject of quantum computing brings together ideas from classical information theory, computer science, and quantum physics. This review aims to summarize not just quantum computing, but the whole subject of quantum information theory. Information can be identified as the most general thing which must propagate from a cause to an effect. It therefore has a fundamentally important role in the science of physics. However, the mathematical treatment of information, especially information processing, is quite recent, dating from the mid-20th century. This has meant that the full significance of information as a basic concept in physics is only now being discovered. This is especially true in quantum mechanics. The theory of quantum information and computing puts this significance on a firm footing, and has led to some profound and exciting new insights into the natural world. Among these are the use of quantum states to permit the secure transmission of classical information (quantum cryptography), the use of quantum entanglement to permit reliable transmission of quantum states (teleportation), the possibility of preserving quantum coherence in the presence of irreversible noise processes (quantum error correction), and the use of controlled quantum evolution for efficient computation (quantum computation). The common theme of all these insights is the use of quantum entanglement as a computational resource. It turns out that information theory and quantum mechanics fit together very well. In order to explain their relationship, this review begins with an introduction to classical information theory and computer science, including Shannon's theorem, error correcting codes, Turing machines and computational complexity. The principles of quantum mechanics are then outlined, and the Einstein, Podolsky and Rosen (EPR) experiment described. The EPR-Bell correlations, and quantum entanglement in general, form the essential new ingredient which distinguishes quantum from
here and discuss other problems caused by the adaptiveness. All protocols in the thesis are formally specified and the proofs of their security are given. Ronald Cramer, Ivan Damgård, Stefan Dziembowski, Martin Hirt, and Tal Rabin. Efficient multiparty computations with dishonest minority......In this thesis we study a problem of doing Verifiable Secret Sharing (VSS) and Multiparty Computations in a model where private channels between the players and a broadcast channel is available. The adversary is active, adaptive and has an unbounded computing power. The thesis is based on two...... to a polynomial time black-box reduction, the complexity of adaptively secure VSS is the same as that of ordinary secret sharing (SS), where security is only required against a passive, static adversary. Previously, such a connection was only known for linear secret sharing and VSS schemes. We then show...
Trangenstein, John A
Wang, Xiao-Jing; Krystal, John H.
Psychiatric disorders such as autism and schizophrenia arise from abnormalities in brain systems that underlie cognitive, emotional and social functions. The brain is enormously complex and its abundant feedback loops on multiple scales preclude intuitive explication of circuit functions. In close interplay with experiments, theory and computational modeling are essential for understanding how, precisely, neural circuits generate flexible behaviors and their impairments give rise to psychiatric symptoms. This Perspective highlights recent progress in applying computational neuroscience to the study of mental disorders. We outline basic approaches, including identification of core deficits that cut across disease categories, biologically-realistic modeling bridging cellular and synaptic mechanisms with behavior, model-aided diagnosis. The need for new research strategies in psychiatry is urgent. Computational psychiatry potentially provides powerful tools for elucidating pathophysiology that may inform both diagnosis and treatment. To achieve this promise will require investment in cross-disciplinary training and research in this nascent field. PMID:25442941
Full Text Available Approximate Bayesian computation (ABC constitutes a class of computational methods rooted in Bayesian statistics. In all model-based statistical inference, the likelihood function is of central importance, since it expresses the probability of the observed data under a particular statistical model, and thus quantifies the support data lend to particular values of parameters and to choices among different models. For simple models, an analytical formula for the likelihood function can typically be derived. However, for more complex models, an analytical formula might be elusive or the likelihood function might be computationally very costly to evaluate. ABC methods bypass the evaluation of the likelihood function. In this way, ABC methods widen the realm of models for which statistical inference can be considered. ABC methods are mathematically well-founded, but they inevitably make assumptions and approximations whose impact needs to be carefully assessed. Furthermore, the wider application domain of ABC exacerbates the challenges of parameter estimation and model selection. ABC has rapidly gained popularity over the last years and in particular for the analysis of complex problems arising in biological sciences (e.g., in population genetics, ecology, epidemiology, and systems biology.
Manning, Timmy; Sleator, Roy D; Walsh, Paul
For decades, computer scientists have looked to nature for biologically inspired solutions to computational problems; ranging from robotic control to scheduling optimization. Paradoxically, as we move deeper into the post-genomics era, the reverse is occurring, as biologists and bioinformaticians look to computational techniques, to solve a variety of biological problems. One of the most common biologically inspired techniques are genetic algorithms (GAs), which take the Darwinian concept of natural selection as the driving force behind systems for solving real world problems, including those in the bioinformatics domain. Herein, we provide an overview of genetic algorithms and survey some of the most recent applications of this approach to bioinformatics based problems.
Schmidt, Kjeld; Bansler, Jørgen P.
The key concern of CSCW research is that of understanding computing technologies in the social context of their use, that is, as integral features of our practices and our lives, and to think of their design and implementation under that perspective. However, the question of the nature...... of that which is actually integrated in our practices is often discussed in confusing ways, if at all. The article aims to try to clarify the issue and in doing so revisits and reconsiders the notion of ‘computational artifact’....
Cloud computing has recently emerged as a subject of substantial industrial and academic interest, though its meaning and scope is hotly debated. For some researchers, clouds are a natural evolution towards the full commercialisation of grid systems, while others dismiss the term as a mere re-branding of existing pay-per-use technologies. From either perspective, 'cloud' is now the label of choice for accountable pay-per-use access to third party applications and computational resources on a massive scale. Clouds support patterns of less predictable resource use for applications and services a
Pacino, Dario; Voss, Stefan; Jensen, Rune Møller
This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized in to...... in topical sections named: maritime shipping, road transport, vehicle routing problems, aviation applications, and logistics and supply chain management.......This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized...
This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized in to...... in topical sections named: maritime shipping, road transport, vehicle routing problems, aviation applications, and logistics and supply chain management.......This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized...
As more and more equipment is interface or'bus' driven, either by the use of controllers or directly from PCs, the question of which bus to use is becoming increasingly important both in industry and in the office. 'Computer Busses' has been designed to help choose the best type of bus for the particular application.There are several books which cover individual busses, but none which provide a complete guide to computer busses. The author provides a basic theory of busses and draws examples and applications from real bus case studies. Busses are analysed using from a top-down approach, helpin
Cardoso, Joao MP
As the complexity of modern embedded systems increases, it becomes less practical to design monolithic processing platforms. As a result, reconfigurable computing is being adopted widely for more flexible design. Reconfigurable Computers offer the spatial parallelism and fine-grained customizability of application-specific circuits with the postfabrication programmability of software. To make the most of this unique combination of performance and flexibility, designers need to be aware of both hardware and software issues. FPGA users must think not only about the gates needed to perform a comp
This book presents a comprehensive treatise on Riemannian geometric computations and related statistical inferences in several computer vision problems. This edited volume includes chapter contributions from leading figures in the field of computer vision who are applying Riemannian geometric approaches in problems such as face recognition, activity recognition, object detection, biomedical image analysis, and structure-from-motion. Some of the mathematical entities that necessitate a geometric analysis include rotation matrices (e.g. in modeling camera motion), stick figures (e.g. for activity recognition), subspace comparisons (e.g. in face recognition), symmetric positive-definite matrices (e.g. in diffusion tensor imaging), and function-spaces (e.g. in studying shapes of closed contours). · Illustrates Riemannian computing theory on applications in computer vision, machine learning, and robotics · Emphasis on algorithmic advances that will allow re-application in other...
Lee, Yung Seng
Common obesity is widely regarded as a complex, multifactorial trait influenced by the 'obesogenic' environment, sedentary behavior, and genetic susceptibility contributed by common and rare genetic variants. This review describes the recent advances in understanding the role of genetics in obesity. New susceptibility loci and genetic variants are being uncovered, but the collective effect is relatively small and could not explain most of the BMI heritability. Yet-to-be identified common and rare variants, epistasis, and heritable epigenetic changes may account for part of the 'missing heritability'. Evidence is emerging about the role of epigenetics in determining obesity susceptibility, mediating developmental plasticity, which confers obesity risk from early life experiences. Genetic prediction scores derived from selected genetic variants, and also differential DNA methylation levels and methylation scores, have been shown to correlate with measures of obesity and response to weight loss intervention. Genetic variants, which confer susceptibility to obesity-related morbidities like nonalcoholic fatty liver disease, were also discovered recently. We can expect discovery of more rare genetic variants with the advent of whole exome and genome sequencing, and also greater understanding of epigenetic mechanisms by which environment influences genetic expression and which mediate the gene-environment interaction.
Harper, Joyce C
Preimplantation genetic diagnosis was first successfully performed in 1989 as an alternative to prenatal diagnosis for couples at risk of transmitting a genetic or chromosomal abnormality, such as cystic fibrosis, to their child. From embryos generated in vitro, biopsied cells are genetically tested. From the mid-1990s, this technology has been employed as an embryo selection tool for patients undergoing in vitro fertilisation, screening as many chromosomes as possible, in the hope that selecting chromosomally normal embryos will lead to higher implantation and decreased miscarriage rates. This procedure, preimplantation genetic screening, was initially performed using fluorescent in situ hybridisation, but 11 randomised controlled trials of screening using this technique showed no improvement in in vitro fertilisation delivery rates. Progress in genetic testing has led to the introduction of array comparative genomic hybridisation, quantitative polymerase chain reaction, and next generation sequencing for preimplantation genetic screening, and three small randomised controlled trials of preimplantation genetic screening using these new techniques indicate a modest benefit. Other trials are still in progress but, regardless of their results, preimplantation genetic screening is now being offered globally. In the near future, it is likely that sequencing will be used to screen the full genetic code of the embryo.
inference and finite population sampling. Sudhakar Kunte. Elements of statistical computing are discussed in this series. ... which captain gets an option to decide whether to field first or bat first ... may of course not be fair, in the sense that the team which wins ... describe two methods of drawing a random number between 0.
Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue
Computation via biological devices has been the subject of close scrutiny since von Neumann’s early work some 60 years ago. In spite of the many relevant works in this field, the notion of programming biological devices seems to be, at best, ill-defined. While many devices are claimed or proved t...
We are still five years from the first LHC data, so we have plenty of time to get the computing into shape, don't we? Well, yes and no: there is time, but there's an awful lot to do! The recently-completed CERN Review of LHC Computing gives the flavour of the LHC computing challenge. The hardware scale for each of the LHC experiments is millions of 'SpecInt95' (SI95) units of cpu power and tens of PetaBytes of data storage. PCs today are about 20-30SI95, and expected to be about 100 SI95 by 2005, so it's a lot of PCs. This hardware will be distributed across several 'Regional Centres' of various sizes, connected by high-speed networks. How to realise this in an orderly and timely fashion is now being discussed in earnest by CERN, Funding Agencies, and the LHC experiments. Mixed in with this is, of course, the GRID concept...but that's a topic for another day! Of course hardware, networks and the GRID constitute just one part of the computing. Most of the ATLAS effort is spent on software development. What we ...
Home; Journals; Resonance – Journal of Science Education; Volume 16; Issue 9. Quantum Computation - Particle and Wave Aspects of Algorithms. Apoorva Patel. General Article Volume 16 Issue 9 September 2011 pp 821-835. Fulltext. Click here to view fulltext PDF. Permanent link:
Wink, Diane M
In this bimonthly series, the author examines how nurse educators can use Internet and Web-based technologies such as search, communication, and collaborative writing tools; social networking and social bookmarking sites; virtual worlds; and Web-based teaching and learning programs. This article describes how cloud computing can be used in nursing education.
Dewdney, A. K.
Describes the creation of the computer program "BOUNCE," designed to simulate a weighted piston coming into equilibrium with a cloud of bouncing balls. The model follows the ideal gas law. Utilizes the critical event technique to create the model. Discusses another program, "BOOM," which simulates a chain reaction. (CW)
"Turn on a water spigot, and it's like tapping a bottomless barrel of water. Ditto for electricity: Flip the switch, and the supply is endless. But computing is another matter. Even with the Internet revolution enabling us to connect in new ways, we are still limited to self-contained systems running locally stored software, limited by corporate, institutional and geographic boundaries" (1 page).
One of the major challenges in todays post-crisis finance environment is calculating the sensitivities of complex products for hedging and risk management. Historically, these derivatives have been determined using bump-and-revalue, but due to the increasing magnitude of these computations does...
Optical computing technology is, in general, developing in two directions. One approach is ... current support in many places, with private companies as well as governments in several countries encouraging such research work. For example, much ... which enables more information to be carried and data to be processed.
A genetic algorithm was used to optimize the power output of multi-junction solar cells. Solar cell operation was modeled using the Silvaco ATLASTM software. The output of the ATLASTM simulation runs served as the input to the genetic algorithm. The genetic algorithm was run as a diffusing computation on a network of eighteen dual processor nodes. Results showed that the genetic algorithm produced better power output optimizations when compared with the results obtained using the hill cli...
Home; Journals; Journal of Genetics; Volume 84; Issue 3. Towards a genetic architecture of cryptic genetic variation and genetic assimilation: the contribution of K. G. Bateman. Ian Dworkin. Commentary on J. Genet. Classic Volume 84 Issue 3 December 2005 pp 223-226 ...
Full Text Available We develop some parts of the frame theory in Banach spaces from the point of view of Computable Analysis. We define computable M-basis and use it to construct a computable Banach space of scalar valued sequences. Computable Xd frames and computable Banach frames are also defined and computable versions of sufficient conditions for their existence are obtained.
... RefSeqGene UniGene All Genes & Expression Resources... Genetics & Medicine Bookshelf Database of Genotypes and Phenotypes (dbGaP) Genetic Testing ... ProtMap HomoloGene Protein Clusters All Homology Resources... Literature Bookshelf E-Utilities Journals in NCBI Databases MeSH Database ...
Coyle, Heather; Drell, Dan
Various: (1)TriState 2000 Genetics in the Courts (2) Growing impact of the new genetics on the courts (3)Human testing (4) Legal analysis - in re G.C. (5) Legal analysis - GM ''peanots'', and (6) Legal analysis for State vs Miller
Czeizel, Andrew E.
The beginning of human genetics and its medical part: medical genetics was promising in the early decades of this century. Many genetic diseases and defects with Mendelian origin were identified and it helped families with significant genetic burden to limit their child number. Unfortunately this good start was shadowed by two tragic events. On the one hand, in the 1930s and early 1940s the German fascism brought about the dominance of an unscientific eugenics to mask vile political crimes. People with genetic diseases-defects were forced to sterilisation and several of them were killed. On the other hand, in the 1950s lysenkoism inhibitied the evolution of genetics in the Soviet Union and their satelite countries. Lysenko's doctrine declared genetics as a product of imperialism and a guilty science, therefore leading geneticists were ousted form their posts and some of them were executed or put in prison. Past decades genetics has resulted fantastic new results and achieved a leading position within the natural sciences. To my mind, however, the expected wider use of new eugenics indicates a new tragedy and this Cassandra's prediction is the topic of this presentation.
Mohammad Saad Zaghloul Salem
Dec 24, 2014 ... ome/transcriptome/proteome, experimental induced maps that are intentionally designed and con- ... genetic maps imposed their application in nearly all fields of medical genetics including ..... or genes located adjacent to, or near, them. ...... types of markers, e.g., clinical markers (eye color), genomic.
The implementation of a new computer algebra system is time consuming: designers of general purpose algebra systems usually say it takes about 50 man-years to create a mature and fully functional system. Hence the range of available systems and their capabilities changes little between one general relativity meeting and the next, despite which there have been significant changes in the period since the last report. The introductory remarks aim to give a brief survey of capabilities of the principal available systems and highlight one or two trends. The reference to the most recent full survey of computer algebra in relativity and brief descriptions of the Maple, REDUCE and SHEEP and other applications are given. (author)
Worzel, William P; Yu, Jianjun; Almal, Arpit A; Chinnaiyan, Arul M
The theory of Darwinian evolution is the fundamental keystones of modern biology. Late in the last century, computer scientists began adapting its principles, in particular natural selection, to complex computational challenges, leading to the emergence of evolutionary algorithms. The conceptual model of selective pressure and recombination in evolutionary algorithms allow scientists to efficiently search high dimensional space for solutions to complex problems. In the last decade, genetic programming has been developed and extensively applied for analysis of molecular data to classify cancer subtypes and characterize the mechanisms of cancer pathogenesis and development. This article reviews current successes using genetic programming and discusses its potential impact in cancer research and treatment in the near future.
Timmermans, Benjamin; Kuhn, Tobias; Beelen, Kaspar; Aroyo, Lora
Climate change, vaccination, abortion, Trump: Many topics are surrounded by fierce controversies. The nature of such heated debates and their elements have been studied extensively in the social science literature. More recently, various computational approaches to controversy analysis have appeared, using new data sources such as Wikipedia, which help us now better understand these phenomena. However, compared to what social sciences have discovered about such debates, the existing computati...
Andre, M.; Resnick, D.
Computed tomography (CT) has matured into a reliable and prominent tool for study of the muscoloskeletal system. When it was introduced in 1973, it was unique in many ways and posed a challenge to interpretation. It is in these unique features, however, that its advantages lie in comparison with conventional techniques. These advantages will be described in a spectrum of important applications in orthopedics and rheumatology
Computed radiography (CR) is an image acquisition process that is used to create digital, 2-dimensional radiographs. CR employs a photostimulable phosphor-based imaging plate, replacing the standard x-ray film and intensifying screen combination. Conventional radiographic exposure equipment is used with no modification required to the existing system. CR can transform an analog x-ray department into a digital one and eliminates the need for chemicals, water, darkrooms and film processor headaches. (author)
Suspicions that the world might be some sort of a machine or algorithm existing 'in the mind' of some symbolic number cruncher have lingered from antiquity. Although popular at times, the most radical forms of this idea never reached mainstream. Modern developments in physics and computer science have lent support to the thesis, but empirical evidence is needed before it can begin to replace our contemporary world view
Microcephaly is associated with reduced cortical surface area and ventricular dilations. Many genetic and environmental factors precipitate this malformation, including prenatal alcohol exposure and maternal Zika infection. This complexity motivates the engineering of computation...
failure probability. Multiobjective Evolutionary Computation algorithms (MOEAs) are well-suited for Multiobjective task scheduling on heterogeneous environment. The two Multi-Objective Evolutionary Algorithms such as Multiobjective Genetic. Algorithm (MOGA) and Multiobjective Evolutionary Programming (MOEP) with.
Masys, Daniel R.
The evolution from classical genetics to biotechnology, an area of research involving key macromolecules in living cells, is chronicled and the current state of biotechnology is described, noting related advances in computing and clinical medicine. (MSE)
Annalise B. Paaby
Full Text Available Evolutionary developmental genetics has traditionally been conducted by two groups: Molecular evolutionists who emphasize divergence between species or higher taxa, and quantitative geneticists who study variation within species. Neither approach really comes to grips with the complexities of evolutionary transitions, particularly in light of the realization from genome-wide association studies that most complex traits fit an infinitesimal architecture, being influenced by thousands of loci. This paper discusses robustness, plasticity and lability, phenomena that we argue potentiate major evolutionary changes and provide a bridge between the conceptual treatments of macro- and micro-evolution. We offer cryptic genetic variation and conditional neutrality as mechanisms by which standing genetic variation can lead to developmental system drift and, sheltered within canalized processes, may facilitate developmental transitions and the evolution of novelty. Synthesis of the two dominant perspectives will require recognition that adaptation, divergence, drift and stability all depend on similar underlying quantitative genetic processes—processes that cannot be fully observed in continuously varying visible traits.
Anholt, Robert R H; Mackay, Trudy F C
Aggression mediates competition for food, mating partners, and habitats and, among social animals, establishes stable dominance hierarchies. In humans, abnormal aggression is a hallmark of neuropsychiatric disorders and can be elicited by environmental factors acting on an underlying genetic susceptibility. Identifying the genetic architecture that predisposes to aggressive behavior in people is challenging because of difficulties in quantifying the phenotype, genetic heterogeneity, and uncontrolled environmental conditions. Studies on mice have identified single-gene mutations that result in hyperaggression, contingent on genetic background. These studies can be complemented by systems genetics approaches in Drosophila melanogaster, in which mutational analyses together with genome-wide transcript analyses, artificial selection studies, and genome-wide analysis of epistasis have revealed that a large segment of the genome contributes to the manifestation of aggressive behavior with widespread epistatic interactions. Comparative genomic analyses based on the principle of evolutionary conservation are needed to enable a complete dissection of the neurogenetic underpinnings of this universal fitness trait.
Jaramillo Vasquez, J.G.
Some genetic bases of the improvement of vegetables are given. The objectives of the genetic improvement and the fundamental stages of this process are done. The sources of genetic variation are indicated and they are related the reproduction systems of the main horticultural species. It is analyzed the concept of genetic inheritance like base to determine the procedures more appropriate of improvement. The approaches are discussed, has more than enough phenotypic value, genetic action and genotypic variance; Equally the heredability concepts and value of improvement. The conventional methods of improvement are described, like they are: the introduction of species or varieties, the selection, the pure line, the pedigree method, the selection for families, the recurrent selection, the selection for unique seed, the haploids method, the selection for heterosis and the synthetic varieties
Since the introduction in the mid-1980s of analyses of minisatellites for DNA analyses, a revolution has taken place in forensic genetics. The subsequent invention of the PCR made it possible to develop forensic genetics tools that allow both very informative routine investigations and still more...... and more advanced, special investigations in cases concerning crime, paternity, relationship, disaster victim identification etc. The present review gives an update on the use of DNA investigations in forensic genetics.......Since the introduction in the mid-1980s of analyses of minisatellites for DNA analyses, a revolution has taken place in forensic genetics. The subsequent invention of the PCR made it possible to develop forensic genetics tools that allow both very informative routine investigations and still more...
By Nancy Parrish, Staff Writer One morning in early January, Amar Klar sat down at his computer and found an e-mail with a curious message from a colleague. While reading a bestselling novel, The Marriage Plot by Jeffrey Eugenides, his colleague, a professor at Princeton University, found a description of research on yeast genetics that was surprisingly similar to Klar’s early
... can be inherited? More about Inheriting Genetic Conditions Diagnosis & Management Resources Genetic Testing (4 links) Genetic Testing Registry: Ateleiotic dwarfism Genetic Testing Registry: Autosomal dominant isolated somatotropin deficiency ...
Roberts, D.F.; De Stefano, G.F.
This book contains several papers divided among three sections. The section titles are: Genetic Diversity--Its Dimensions; Genetic Diversity--Its Origin and Maintenance; and Genetic Diversity--Applications and Problems of Complex Characters
... features. Type 1 is the most common, or classical, form of this condition and is associated with ... be inherited? More about Inheriting Genetic Conditions Diagnosis & Management Resources Genetic Testing (1 link) Genetic Testing Registry: ...
Full Text Available Genome-wide association study (GWAS aims to discover genetic factors underlying phenotypic traits. The large number of genetic factors poses both computational and statistical challenges. Various computational approaches have been developed for large scale GWAS. In this chapter, we will discuss several widely used computational approaches in GWAS. The following topics will be covered: (1 An introduction to the background of GWAS. (2 The existing computational approaches that are widely used in GWAS. This will cover single-locus, epistasis detection, and machine learning methods that have been recently developed in biology, statistic, and computer science communities. This part will be the main focus of this chapter. (3 The limitations of current approaches and future directions.
Genetic heterogeneity is a phenomenon in which a genetic disease can be transmitted by several modes of inheritance. The understanding of genetic heterogeneity is important in giving genetic counselling.The presence of genetic heterogeneity can be explained by the existence of:1.different mutant alleles at a single locus, and2.mutant alleles at different loci affecting the same enzyme or protein, or affecting different enzymes or proteins.To have an overall understanding of genetic heterogene...
Chen, Yu-Ting; Gill, Michael; Reinman, Glenn; Xiao, Bingjun
Since the end of Dennard scaling in the early 2000s, improving the energy efficiency of computation has been the main concern of the research community and industry. The large energy efficiency gap between general-purpose processors and application-specific integrated circuits (ASICs) motivates the exploration of customizable architectures, where one can adapt the architecture to the workload. In this Synthesis lecture, we present an overview and introduction of the recent developments on energy-efficient customizable architectures, including customizable cores and accelerators, on-chip memory
Virvou, Maria; Jain, Lakhmi
This book at hand explores emerging scientific and technological areas in which Intelligent Computing Systems provide efficient solutions and, thus, may play a role in the years to come. It demonstrates how Intelligent Computing Systems make use of computational methodologies that mimic nature-inspired processes to address real world problems of high complexity for which exact mathematical solutions, based on physical and statistical modelling, are intractable. Common intelligent computational methodologies are presented including artificial neural networks, evolutionary computation, genetic algorithms, artificial immune systems, fuzzy logic, swarm intelligence, artificial life, virtual worlds and hybrid methodologies based on combinations of the previous. The book will be useful to researchers, practitioners and graduate students dealing with mathematically-intractable problems. It is intended for both the expert/researcher in the field of Intelligent Computing Systems, as well as for the general reader in t...
The Soft Computing techniques, which are based on the information processing of biological systems are now massively used in the area of pattern recognition, making prediction & planning, as well as acting on the environment. Ideally speaking, soft computing is not a subject of homogeneous concepts and techniques; rather, it is an amalgamation of distinct methods that confirms to its guiding principle. At present, the main aim of soft computing is to exploit the tolerance for imprecision and uncertainty to achieve tractability, robustness and low solutions cost. The principal constituents of soft computing techniques are probabilistic reasoning, fuzzy logic, neuro-computing, genetic algorithms, belief networks, chaotic systems, as well as learning theory. This book covers contributions from various authors to demonstrate the use of soft computing techniques in various applications of engineering.
Many of the most important findings concerning the genetic effects of radiation have been obtained in the Biology Division of Oak Ridge National Laboratory. The paper focuses on some of the major discoveries made in the Biology Division and on a new method of research that assesses damage to the skeletons of mice whose fathers were irradiated. The results discussed have considerable influence upon estimates of genetic risk in humans from radiation, and an attempt is made to put the estimated amount of genetic damage caused by projected nuclear power development into its proper perspective
Zhou, Ruanbao (Inventor); Gibbons, William (Inventor)
The disclosed embodiments provide cyanobacteria spp. that have been genetically engineered to have increased production of carbon-based products of interest. These genetically engineered hosts efficiently convert carbon dioxide and light into carbon-based products of interest such as long chained hydrocarbons. Several constructs containing polynucleotides encoding enzymes active in the metabolic pathways of cyanobacteria are disclosed. In many instances, the cyanobacteria strains have been further genetically modified to optimize production of the carbon-based products of interest. The optimization includes both up-regulation and down-regulation of particular genes.
Charles, Abigail Sheena
This study investigated the knowledge and skills that biology students may need to help them understand statistics/mathematics as it applies to genetics. The data are based on analyses of current representative genetics texts, practicing genetics professors' perspectives, and more directly, students' perceptions of, and performance in, doing statistically-based genetics problems. This issue is at the emerging edge of modern college-level genetics instruction, and this study attempts to identify key theoretical components for creating a specialized biological statistics curriculum. The goal of this curriculum will be to prepare biology students with the skills for assimilating quantitatively-based genetic processes, increasingly at the forefront of modern genetics. To fulfill this, two college level classes at two universities were surveyed. One university was located in the northeastern US and the other in the West Indies. There was a sample size of 42 students and a supplementary interview was administered to a select 9 students. Interviews were also administered to professors in the field in order to gain insight into the teaching of statistics in genetics. Key findings indicated that students had very little to no background in statistics (55%). Although students did perform well on exams with 60% of the population receiving an A or B grade, 77% of them did not offer good explanations on a probability question associated with the normal distribution provided in the survey. The scope and presentation of the applicable statistics/mathematics in some of the most used textbooks in genetics teaching, as well as genetics syllabi used by instructors do not help the issue. It was found that the text books, often times, either did not give effective explanations for students, or completely left out certain topics. The omission of certain statistical/mathematical oriented topics was seen to be also true with the genetics syllabi reviewed for this study. Nonetheless
Hyo Seon Park
Full Text Available Since genetic algorithm-based optimization methods are computationally expensive for practical use in the field of structural optimization, a resizing technique-based hybrid genetic algorithm for the drift design of multistory steel frame buildings is proposed to increase the convergence speed of genetic algorithms. To reduce the number of structural analyses required for the convergence, a genetic algorithm is combined with a resizing technique that is an efficient optimal technique to control the drift of buildings without the repetitive structural analysis. The resizing technique-based hybrid genetic algorithm proposed in this paper is applied to the minimum weight design of three steel frame buildings. To evaluate the performance of the algorithm, optimum weights, computational times, and generation numbers from the proposed algorithm are compared with those from a genetic algorithm. Based on the comparisons, it is concluded that the hybrid genetic algorithm shows clear improvements in convergence properties.
Wells, P.; Davis, J.; Morgan, M.
X-ray or gamma-ray transmission computed tomography (CT) is a powerful non-destructive evaluation (NDE) technique that produces two-dimensional cross-sectional images of an object without the need to physically section it. CT is also known by the acronym CAT, for computerised axial tomography. This review article presents a brief historical perspective on CT, its current status and the underlying physics. The mathematical fundamentals of computed tomography are developed for the simplest transmission CT modality. A description of CT scanner instrumentation is provided with an emphasis on radiation sources and systems. Examples of CT images are shown indicating the range of materials that can be scanned and the spatial and contrast resolutions that may be achieved. Attention is also given to the occurrence, interpretation and minimisation of various image artefacts that may arise. A final brief section is devoted to the principles and potential of a range of more recently developed tomographic modalities including diffraction CT, positron emission CT and seismic tomography. 57 refs., 2 tabs., 14 figs
Full Text Available The isolation with migration (IM model is important for studies in population genetics and phylogeography. IM program applies the IM model to genetic data drawn from a pair of closely related populations or species based on Markov chain Monte Carlo (MCMC simulations of gene genealogies. But computational burden of IM program has placed limits on its application.With strong computational power, Graphics Processing Unit (GPU has been widely used in many fields. In this article, we present an effective implementation of IM program on one GPU based on Compute Unified Device Architecture (CUDA, which we call gPGA.Compared with IM program, gPGA can achieve up to 52.30X speedup on one GPU. The evaluation results demonstrate that it allows datasets to be analyzed effectively and rapidly for research on divergence population genetics. The software is freely available with source code at https://github.com/chunbaozhou/gPGA.
Pedersen, L D; Sørensen, A C; Henryon, M
ADAM is a computer program that models selective breeding schemes for animals using stochastic simulation. The program simulates a population of animals and traces the genetic changes in the population under different selective breeding scenarios. It caters to different population structures......, genetic models, selection strategies, and mating designs. ADAM can be used to evaluate breeding schemes and generate genetic data to test statistical tools...
Classical and Quantum computing provides a self-contained, systematic and comprehensive introduction to all the subjects and techniques important in scientific computing. The style and presentation are readily accessible to undergraduates and graduates. A large number of examples, accompanied by complete C++ and Java code wherever possible, cover every topic. Features and benefits: - Comprehensive coverage of the theory with many examples - Topics in classical computing include boolean algebra, gates, circuits, latches, error detection and correction, neural networks, Turing machines, cryptography, genetic algorithms - For the first time, genetic expression programming is presented in a textbook - Topics in quantum computing include mathematical foundations, quantum algorithms, quantum information theory, hardware used in quantum computing This book serves as a textbook for courses in scientific computing and is also very suitable for self-study. Students, professionals and practitioners in computer...
Jonathan H. Young
Full Text Available Characterizing genetic interactions is crucial to understanding cellular and organismal response to gene-level perturbations. Such knowledge can inform the selection of candidate disease therapy targets, yet experimentally determining whether genes interact is technically nontrivial and time-consuming. High-fidelity prediction of different classes of genetic interactions in multiple organisms would substantially alleviate this experimental burden. Under the hypothesis that functionally related genes tend to share common genetic interaction partners, we evaluate a computational approach to predict genetic interactions in Homo sapiens, Drosophila melanogaster, and Saccharomyces cerevisiae. By leveraging knowledge of functional relationships between genes, we cross-validate predictions on known genetic interactions and observe high predictive power of multiple classes of genetic interactions in all three organisms. Additionally, our method suggests high-confidence candidate interaction pairs that can be directly experimentally tested. A web application is provided for users to query genes for predicted novel genetic interaction partners. Finally, by subsampling the known yeast genetic interaction network, we found that novel genetic interactions are predictable even when knowledge of currently known interactions is minimal.
Bay, Bjorn; Ingerslev, Hans Jakob; Lemmen, Josephine Gabriela
OBJECTIVE: To study whether women conceiving after preimplantation genetic diagnosis (PGD) and their children have greater risks of adverse pregnancy and birth outcomes compared with children conceived spontaneously or after IVF with or without intracytoplasmic sperm injection (ICSI). DESIGN...
... Testing that reveals a young child’s genet- ic destiny may affect relationships within the family or may ... linked inheritance don’t apply at all. An embryo receives its mitochondria from the mother’s egg cell, ...
Dishotsky, Norman I.; And Others
Reviews studies of the effects of lysergic acid diethylamide (LSD) on man and other organisms. Concludes that pure LSD injected in moderate doses does not cause chromosome or detectable genetic damage and is not a teratogen or carcinogen. (JM)
... be a feature of other conditions, such as Waardenburg syndrome ; these conditions have other genetic causes and additional ... 140S. Review. Citation on PubMed Spritz RA. Piebaldism, Waardenburg syndrome, and related disorders of melanocyte development. Semin Cutan ...
... inheritance of sialuria, an inborn error of feedback inhibition. Am J Hum Genet. 2001 Jun;68(6): ... Links Data Files & API Site Map Subscribe Customer Support USA.gov Copyright Privacy Accessibility FOIA Viewers & Players ...
Mellerup, Erling; Møller, Gert Lykke; Koefoed, Pernille
A complex disease with an inheritable component is polygenic, meaning that several different changes in DNA are the genetic basis for the disease. Such a disease may also be genetically heterogeneous, meaning that independent changes in DNA, i.e. various genotypes, can be the genetic basis...... for the disease. Each of these genotypes may be characterized by specific combinations of key genetic changes. It is suggested that even if all key changes are found in genes related to the biology of a certain disease, the number of combinations may be so large that the number of different genotypes may be close...... to the number of patients suffering from the disease. This hypothesis is based on a study of bipolar disorder....
... A A A Listen En Español Genetics of Diabetes You've probably wondered how you developed diabetes. ... to develop diabetes than others. What Leads to Diabetes? Type 1 and type 2 diabetes have different ...
Ibañez Cuadrado, Angela
The addictions are common chronic psychiatric diseases which represent a serious worldwide public-health problem. They have a high prevalence and negative effects at individual, family and societal level, with a high sanitary cost. Epidemiological genetic research has revealed that addictions are moderately to highly heritable. Also the investigation has evidenced that environmental and genetic factors contribute to individual differences in vulnerability to addictions. Advances in the neurobiology of addiction joined to the development of new molecular genetic technologies, have led to the identification of a variety of underlying genes and pathways in addiction process, leading to the description of common molecular mechanisms in substance and behaviour dependencies. Identifying gene-environment interactions is a crucial issue in future research. Other major goal in genetic research is the identification of new therapeutic targets for treatment and prevention.
Karthikeyan A Sadagopan
Full Text Available The eye has played a major role in human genomics including gene therapy. It is the fourth most common organ system after integument (skin, hair and nails, nervous system, and musculoskeletal system to be involved in genetic disorders. The eye is involved in single gene disorders and those caused by multifactorial etiology. Retinoblastoma was the first human cancer gene to be cloned. Leber hereditary optic neuropathy was the first mitochondrial disorder described. X-Linked red-green color deficiency was the first X-linked disorder described. The eye, unlike any other body organ, allows directly visualization of genetic phenomena such as skewed X-inactivation in the fundus of a female carrier of ocular albinism. Basic concepts of genetics and their application to clinical ophthalmological practice are important not only in making a precise diagnosis and appropriate referral, but also in management and genetic counseling.
... also helps regulate cholesterol levels in a similar fashion; normally about 50 percent of cholesterol in the ... 10 All Bulletins Features What is direct-to-consumer genetic testing? What are genome editing and CRISPR- ...
In recent years, practitioners of medicine have become increasingly aware of the importance of genetics in the understanding of physical and mental health and in the management of disease. The last decades have witnessed unprecedented developments in genetics that have increased our understanding of the basic processes of heredity enormously. New techniques and understanding have provided insights directly applicable to medicine. The fundamental fact of heredity may be considered the ability of living organisms to produce offspring that resemble their parents more than others. One of the basic characteristics of the human condition is the uniqueness and diversity of all individuals. This results from their genetic individuality (with the exception of identical twins) and the interaction of the genetic constitution (the genome) with the environment, which is generally unique to the individual as well. In short, the interaction of genes with the environment is what confers biologic uniqueness to all humans
National Oceanic and Atmospheric Administration, Department of Commerce — This database archives genetic tissue samples from marine mammals collected primarily from the U.S. east coast. The collection includes samples from field programs,...
... Care Genomic Medicine Working Group New Horizons and Research Patient Management Policy and Ethics Issues Quick Links for Patient Care Education All About the Human Genome Project Fact Sheets Genetic Education Resources for ...
Many different types of genetic mutations are found in cancer cells. This infographic outlines certain types of alterations that are present in cancer, such as missense, nonsense, frameshift, and chromosome rearrangements.
National Oceanic and Atmospheric Administration, Department of Commerce — This database archives genetic tissue samples from marine mammals collected in the North-Central Gulf of Mexico from 2010-2015. The collection includes samples from...
... for Genomics Research Intellectual Property Issues in Genetics Archive Online Bioethics Resources Privacy in Genomics Regulation of ... are not regulated, meaning that they go to market without any independent analysis to verify the claims ...
Rodriguez-Fontenla, Cristina; Gonzalez, Antonio
Osteoarthritis (OA) is a complex disease caused by the interaction of multiple genetic and environmental factors. This review focuses on the studies that have contributed to the discovery of genetic susceptibility factors in OA. The most relevant associations discovered until now are discussed in detail: GDF-5, 7q22 locus, MCF2L, DOT1L, NCOA3 and also some important findings from the arcOGEN study. Moreover, the different approaches that can be used to minimize the specific problems of the study of OA genetics are discussed. These include the study of microsatellites, phenotype standardization and other methods such as meta-analysis of GWAS and gene-based analysis. It is expected that these new approaches contribute to finding new susceptibility genetic factors for OA. Copyright © 2014 Elsevier España, S.L.U. All rights reserved.
... into human evolution and origins and serving as a springboard for important medical research. It also addresses issues of confidentiality and individual privacy for participants in genetic diversity research studies.
... Encyclopedia: Familial hypercholesterolemia Encyclopedia: High blood cholesterol and triglycerides Encyclopedia: Xanthoma Health Topic: Cholesterol Health Topic: High Cholesterol in Children and Teens Health Topic: Lipid Metabolism Disorders Genetic and Rare Diseases Information Center (1 ...
fighters’ ability to execute the mission.” Computing Services 4 We run IT Systems that: provide medical care pay the warfighters manage maintenance...users • 1,400 applications • 18 facilities • 180 software vendors • 18,000+ copies of executive software products • Virtually every type of mainframe and... chocs electriques, de branchez les deux cordons d’al imentation avant de faire le depannage P R IM A R Y SD A S B 1 2 PowerHub 7000 RST U L 00- 00
Full Text Available Berit Kerner Semel Institute for Neuroscience and Human Behavior, University of California, Los Angeles, Los Angeles, CA, USA Abstract: Bipolar disorder is a common, complex genetic disorder, but the mode of transmission remains to be discovered. Many researchers assume that common genomic variants carry some risk for manifesting the disease. The research community has celebrated the first genome-wide significant associations between common single nucleotide polymorphisms (SNPs and bipolar disorder. Currently, attempts are under way to translate these findings into clinical practice, genetic counseling, and predictive testing. However, some experts remain cautious. After all, common variants explain only a very small percentage of the genetic risk, and functional consequences of the discovered SNPs are inconclusive. Furthermore, the associated SNPs are not disease specific, and the majority of individuals with a “risk” allele are healthy. On the other hand, population-based genome-wide studies in psychiatric disorders have rediscovered rare structural variants and mutations in genes, which were previously known to cause genetic syndromes and monogenic Mendelian disorders. In many Mendelian syndromes, psychiatric symptoms are prevalent. Although these conditions do not fit the classic description of any specific psychiatric disorder, they often show nonspecific psychiatric symptoms that cross diagnostic boundaries, including intellectual disability, behavioral abnormalities, mood disorders, anxiety disorders, attention deficit, impulse control deficit, and psychosis. Although testing for chromosomal disorders and monogenic Mendelian disorders is well established, testing for common variants is still controversial. The standard concept of genetic testing includes at least three broad criteria that need to be fulfilled before new genetic tests should be introduced: analytical validity, clinical validity, and clinical utility. These criteria are
Gaunholt, Hans; Toma, Laura
In this report a study of genetic programming (GP) has been performed with respect to a number of applications such as Symbolic function regression, Solving Symbolic Differential Equations, Image encoding, the ant problem etc.......In this report a study of genetic programming (GP) has been performed with respect to a number of applications such as Symbolic function regression, Solving Symbolic Differential Equations, Image encoding, the ant problem etc....
Progress is reported on research activities in the fields of mutagenesis in Haemophilus influenzae and Escherichia coli; radioinduced chromosomal aberrations in mammalian germ cells; effects of uv radiation on xeroderma pigmentosum skin cells; mutations in Chinese hamster ovary cells; radioinduced hemoglobin variants in the mouse; analysis of mutants in yeast; Drosophila genetics; biochemical genetics of Neurospora; DNA polymerase activity in Xenopus laevis oocytes; uv-induced damage in Bacillus subtilis; and others
Smith, David H
Health is an intrinsic value that Christians should respect, but it is not the highest value. Christians should be willing to jeopardize their own health for the health of others, and should repudiate any idea that genetic problems are the result of sin. Rather, sin leads us to make genetic problems harder to live with than they should be. (c) 2009 Wiley-Liss, Inc.
Broerse, J.J.; Barendsen, G.W.; Kal, H.B.; Kogel, A.J. van der
This book contains the extended abstracts of the contributions of the poster workshop sessions on somatic and genetic effects of the 7th international congress of radiation research. They cover the following main topics: haematopoietic and immune systems, mechanisms of late effects in various tissues, endogenous and exogenous factors in radiation carcinogenesis, teratogenic effects, genetic effects, in vitro transformation, tumour induction in different tissues, carcinogenesis in incorporated tissues, cancer epidemology and risk assessment. refs.; figs.; tabs
Salk, Rachel H.; Hyde, Janet S.
Over the past century, much of genetics was deterministic, and feminist researchers framed justified criticisms of genetics research. However, over the past two decades, genetics research has evolved remarkably and has moved far from earlier deterministic approaches. Our article provides a brief primer on modern genetics, emphasizing contemporary…
Use of radioisotopes and ionizing radiations in genetics is presented. Several aspects related to men, animals,plants and microorganisms are reported highlighting biological radiation effects, evolution, mutagenesis and genetic engineering. Genetic mapping, gene mutations, genetic diversity, DNA damages, plant cultivation and plant grow are studied as well
This paper provides an overview of the ethical issues pertaining to the use of genetic insights and techniques in sport. Initially, it considers a range of scientific findings that have stimulated debate about the ethical issues associated with genetics applied to sport. It also outlines some of the early policy responses to these discoveries from world leading sports organizations, along with knowledge about actual use of gene technologies in sport. Subsequently, it considers the challenges with distinguishing between therapeutic use and human enhancement within genetic science, which is a particularly important issue for the world of sport. Next, particular attention is given to the use of genetic information, which raises questions about the legitimacy and reliability of genetic tests, along with the potential public value of having DNA databanks to economize in health care. Finally, the ethics of gene transfer are considered, inviting questions into the values of sport and humanity. It argues that, while gene modification may seem conceptually similar to other forms of doping, the requirements upon athletes are such that new forms of enhancement become increasingly necessary to discover. Insofar as genetic science is able to create safer, more effective techniques of human modification, then it may be an appealing route through which to modify athletes to safeguard the future of elite sports as enterprises of human excellence.
Cancer may result from a multistage process occurring over a long period of time. Presumably, initial and progressive stages of carcinogenesis may be modified by both genetic and environmental factors. Theoretically, genetic factors may alter susceptibility to the carcinogenic effects of an environmental agent at the initial exposure due to variation in metabolism of the carcinogen or variation in specific target cell response to the active carcinogen, or during the latent phase due to numerous factors that might increase the probability of tumor expression, including growth-promoting factors or immunodeficiency states. Observed genetic and environmental interactions in carcinogenesis include an association between genetically determined inducibility of aryl hydrocarbon hydroxylase and smoking-related cancers, familial susceptibility to certain environmental carcinogens, an association between hereditary disorders of mutagenesis and carcinogenesis, and enhancement of tissue-specific, dominantly inherited tumor predisposition by radiation. Multiple primary tumors occur frequently in genetically predisposed individuals. Specific markers for susceptibility must be sought in order that high-risk individuals be identified and appropriate measures taken for early cancer detection or prevention. Study of the nature of the genetically determined susceptibility and interactions with environmental agents may be revealing in the understanding of carcinogenesis in general
Blackwell, Kim L
Progress in Molecular Biology and Translational Science provides a forum for discussion of new discoveries, approaches, and ideas in molecular biology. It contains contributions from leaders in their fields and abundant references. This volume brings together different aspects of, and approaches to, molecular and multi-scale modeling, with applications to a diverse range of neurological diseases. Mathematical and computational modeling offers a powerful approach for examining the interaction between molecular pathways and ionic channels in producing neuron electrical activity. It is well accepted that non-linear interactions among diverse ionic channels can produce unexpected neuron behavior and hinder a deep understanding of how ion channel mutations bring about abnormal behavior and disease. Interactions with the diverse signaling pathways activated by G protein coupled receptors or calcium influx adds an additional level of complexity. Modeling is an approach to integrate myriad data sources into a cohesiv...
The past decade has witnessed a momentous transformation in the way people interact with each other. Content is now co-produced, shared, classified, and rated by millions of people, while attention has become the ephemeral and valuable resource that everyone seeks to acquire. This talk will describe how social attention determines the production and consumption of content within both the scientific community and social media, how its dynamics can be used to predict the future and the role that social media plays in setting the public agenda. About the speaker Bernardo Huberman is a Senior HP Fellow and Director of the Social Computing Lab at Hewlett Packard Laboratories. He received his Ph.D. in Physics from the University of Pennsylvania, and is currently a Consulting Professor in the Department of Applied Physics at Stanford University. He originally worked in condensed matter physics, ranging from superionic conductors to two-dimensional superfluids, and made contributions to the theory of critical p...
N. U. Ahmed
Full Text Available In this paper, we construct a new dynamic model for the Token Bucket (TB algorithm used in computer networks and use systems approach for its analysis. This model is then augmented by adding a dynamic model for a multiplexor at an access node where the TB exercises a policing function. In the model, traffic policing, multiplexing and network utilization are formally defined. Based on the model, we study such issues as (quality of service QoS, traffic sizing and network dimensioning. Also we propose an algorithm using feedback control to improve QoS and network utilization. Applying MPEG video traces as the input traffic to the model, we verify the usefulness and effectiveness of our model.
Full Text Available It is crucial that gifted and talented students should be supported by different educational methods for their interests and skills. The science and arts centres (gifted centres provide the Supportive Education Program for these students with an interdisciplinary perspective. In line with the program, an ICT lesson entitled “Computer Tree” serves for identifying learner readiness levels, and defining the basic conceptual framework. A language teacher also contributes to the process, since it caters for the creative function of the basic linguistic skills. The teaching technique is applied for 9-11 aged student level. The lesson introduces an evaluation process including basic information, skills, and interests of the target group. Furthermore, it includes an observation process by way of peer assessment. The lesson is considered to be a good sample of planning for any subject, for the unpredicted convergence of visual and technical abilities with linguistic abilities.
This paper reports on computed tomographic (CT) scanning which has improved computer-assisted imaging modalities for radiologic diagnosis. The advantage of this modality is its ability to image thin cross-sectional planes of the body, thus uncovering density information in three dimensions without tissue superposition problems. Because this enables vastly superior imaging of soft tissues in the brain and body, CT scanning was immediately successful and continues to grow in importance as improvements are made in speed, resolution, and cost efficiency. CT scanners are used for general purposes, and the more advanced machines are generally preferred in large hospitals, where volume and variety of usage justifies the cost. For imaging in the abdomen, a scanner with a rapid speed is preferred because peristalsis, involuntary motion of the diaphram, and even cardiac motion are present and can significantly degrade image quality. When contrast media is used in imaging to demonstrate scanner, immediate review of images, and multiformat hardcopy production. A second console is reserved for the radiologist to read images and perform the several types of image analysis that are available. Since CT images contain quantitative information in terms of density values and contours of organs, quantitation of volumes, areas, and masses is possible. This is accomplished with region-of- interest methods, which involve the electronic outlining of the selected region of the television display monitor with a trackball-controlled cursor. In addition, various image- processing options, such as edge enhancement (for viewing fine details of edges) or smoothing filters (for enhancing the detectability of low-contrast lesions) are useful tools
Hernandez Torres, Reynier; Irizar Mesa, Mirtha; Tavares Camara, Leoncio Diogenes
In chromatography, complex inverse problems related to the parameters estimation and process optimization are presented. Metaheuristics methods are known as general purpose approximated algorithms which seek and hopefully find good solutions at a reasonable computational cost. These methods are iterative process to perform a robust search of a solution space. Genetic algorithms are optimization techniques based on the principles of genetics and natural selection. They have demonstrated very good performance as global optimizers in many types of applications, including inverse problems. In this work, the effectiveness of genetic algorithms is investigated to estimate parameters in liquid chromatography
Cloud computing has recently emerged as a new paradigm for hosting and delivering services over the Internet. Cloud computing is attractive to business owners as it eliminates the requirement for users to plan ahead for provisioning, and allows enterprises to start from the small and increase resources only when there is a rise in service demand. The basic principles of cloud computing is to make the computing be assigned in a great number of distributed computers, rather then local computer ...
Salem, A.M.; Mohamed, A.H.
Case adaptation is the core of case based reasoning (CBR) approach that can modify the past solutions to solve new problems. It generally relies on the knowledge base and heuristics in order to achieve the required changes. It has always been a difficult process to designers within (CBR) cycle. Its difficulties can be referred to the large effort, and computational analysis needed for acquiring the knowledge's domain. To solve these problems, this research explores a new method that applying a genetic algorithm (GA) to CBR adaptation. However, it can decrease the computational complexity of determining the required changes of the problems especially those having a great amount of domain knowledge. besides, it can decrease the required time by dividing the design task into sub tasks those can be solved at the same time. Therefore, the proposed system can he practically applied for solving the complex problems. It can be used to perform a variety of design tasks on a broad set of application domains. However, it has been implemented for the tablet formulation as a domain of application. Proposed system has improved the accuracy performance of the CBR design systems
Full Text Available The paper concentrates on capability of genetic algorithms for parameter estimation of non-linear economic models. In the paper we test the ability of genetic algorithms to estimate of parameters of demand function for durable goods and simultaneously search for parameters of genetic algorithm that lead to maximum effectiveness of the computation algorithm. The genetic algorithms connect deterministic iterative computation methods with stochastic methods. In the genteic aůgorithm approach each possible solution is represented by one individual, those life and lifes of all generations of individuals run under a few parameter of genetic algorithm. Our simulations resulted in optimal mutation rate of 15% of all bits in chromosomes, optimal elitism rate 20%. We can not set the optimal extend of generation, because it proves positive correlation with effectiveness of genetic algorithm in all range under research, but its impact is degreasing. The used genetic algorithm was sensitive to mutation rate at most, than to extend of generation. The sensitivity to elitism rate is not so strong.
Full Text Available This work was aimed for developing computational intelligence for scheduling a manufacturing cell's tasks, based manily on genetic algorithms. The manufacturing cell was modelled as beign a production-line; the makespan was calculated by using heuristics adapted from several libraries for genetic algorithms computed in C++ builder. Several problems dealing with small, medium and large list of jobs and machinery were resolved. The results were compared with other heuristics. The approach developed here would seem to be promising for future research concerning scheduling manufacturing cell tasks involving mixed batches.
Algoritmos genéticos e computação paralela para problemas de roteirização de veículos com janelas de tempo e entregas fracionadas Genetic algorithms and parallel computing for a vehicle routing problem with time windows and split deliveries
Guilherme Guidolin de Campos
parallel genetic algorithms supported by a cluster of computers. The results indicate that the basic constructive heuristic provides satisfactory results for the problem, but that it can be improved through the use of more sophisticated techniques. The use of the parallel genetic algorithm with multiple populations and an initial solution, which presented the best results, reduced the total operational costs by about 10% compared with the constructive heuristic, and by 13% when compared with the company's original solutions.
Claire M. A. Haworth
Full Text Available Twin and family studies have shown that most traits are at least moderately heritable. But what are the implications of finding genetic influence for the design of intervention and prevention programs? For complex traits, heritability does not mean immutability, and research has shown that genetic influences can change with age, context and in response to behavioural and drug interventions. The most significant implications for intervention will come when we move from observational genetics to investigating dynamic genetics, including genetically sensitive interventions. Future interventions should be designed to overcome genetic risk and draw upon genetic strengths by changing the environment.
Williams, Michael A.; Rigamonti, Daniele
Human hydrocephalus is a common medical condition that is characterized by abnormalities in the flow or resorption of cerebrospinal fluid (CSF), resulting in ventricular dilatation. Human hydrocephalus can be classified into two clinical forms, congenital and acquired. Hydrocephalus is one of the complex and multifactorial neurological disorders. A growing body of evidence indicates that genetic factors play a major role in the pathogenesis of hydrocephalus. An understanding of the genetic components and mechanism of this complex disorder may offer us significant insights into the molecular etiology of impaired brain development and an accumulation of the cerebrospinal fluid in cerebral compartments during the pathogenesis of hydrocephalus. Genetic studies in animal models have started to open the way for understanding the underlying pathology of hydrocephalus. At least 43 mutants/loci linked to hereditary hydrocephalus have been identified in animal models and humans. Up to date, 9 genes associated with hydrocephalus have been identified in animal models. In contrast, only one such gene has been identified in humans. Most of known hydrocephalus gene products are the important cytokines, growth factors or related molecules in the cellular signal pathways during early brain development. The current molecular genetic evidence from animal models indicate that in the early development stage, impaired and abnormal brain development caused by abnormal cellular signaling and functioning, all these cellular and developmental events would eventually lead to the congenital hydrocephalus. Owing to our very primitive knowledge of the genetics and molecular pathogenesis of human hydrocephalus, it is difficult to evaluate whether data gained from animal models can be extrapolated to humans. Initiation of a large population genetics study in humans will certainly provide invaluable information about the molecular and cellular etiology and the developmental mechanisms of human
Rotbain, Yosi; Marbach-Ad, Gili; Stavy, Ruth
We present an active way to use a computer animation in secondary molecular genetics class. For this purpose we developed an activity booklet that helps students to work interactively with a computer animation which deals with abstract concepts and processes in molecular biology. The achievements of the experimental group were compared with those…
... mtDNA Resources Help Me Understand Genetics Share: Email Facebook Twitter Home Help Me Understand Genetics Genetic Consultation How are genetic conditions treated or managed? How are genetic conditions treated or managed? Many ...
Soupios, Pantelis; Akca, Irfan; Mpogiatzis, Petros; Basokur, Ahmet; Papazachos, Constantinos
In the earth sciences several inverse problems that require data fitting and parameter estimation are nonlinear and can involve a large number of unknown parameters. Consequently, the application of analytical inversion or optimization techniques may be quite restrictive. In practice, most analytical methods are local in nature and rely on a linearized form of the problem in question, adopting an iterative procedure using partial derivatives to improve an initial model. This approach can lead to a dependence of the final model solution on the starting model and is prone to entrapment in local misfit minima. Moreover, the calculation of derivatives can be computationally inefficient and create instabilities when numerical approximations are used. In contrast to these local minimization methods, global techniques that do not rely on partial derivatives, are independent of the form of the data misfit criterion, and are computationally robust. Such methods often use random processes to sample a selected wider span of the model space. In this situation, randomly generated models are assessed in terms of their data-fitting quality and the process may be stopped after a certain number of acceptable models is identified or continued until a satisfactory data fit is achieved. A new class of methods known as genetic algorithms achieves the aforementioned approximation through novel model representation and manipulations. Genetic algorithms (GAs) were originally developed in the field of artificial intelligence by John Holland more than 20 years ago, but even in this field it is less than a decade that the methodology has been more generally applied and only recently did the methodology attract the attention of the earth sciences community. Applications have been generally concentrated in geophysics and in particular seismology. As awareness of genetic algorithms grows there surely will be many more and varied applications to earth science problems. In the present work, the
Ichiyen, Norman; Chan, Dominic; Thompson, Paul
The major activity for the 18-month refurbishment outage at the Point Lepreau Generating Station is the replacement of all 380 fuel channel and calandria tube assemblies and the lower portion of connecting feeder pipes. New Brunswick Power would also take advantage of this outage to conduct a number of repairs, replacements, inspections and upgrades (such as rewinding or replacing the generator, replacement of shutdown system trip computers, replacement of certain valves and expansion joints, inspection of systems not normally accessible, etc.). This would allow for an additional 25 to 30 years. Among the systems to be replaced are the PDC's for both shutdown systems. Assessments have been completed for both the SDS1 and SDS2 PDC's, and it has been decided to replace the SDS2 PDCs with the same hardware and software approach that has been used successfully for the Wolsong 2, 3, and 4 and the Qinshan 1 and 2 SDS2 PDCs. For SDS1, it has been decided to use the same software development methodology that was used successfully for the Wolsong and Qinshan called the I A and to use a new hardware platform in order to ensure successful operation for the 25-30 year station operating life. The selected supplier is Triconex, which uses a triple modular redundant architecture that will enhance the robustness/fault tolerance of the design with respect to equipment failures
Ye, Fengming; Mabu, Shingo; Wang, Lutao; Eto, Shinji; Hirasawa, Kotaro
A lot of research on evolutionary computation has been done and some significant classical methods such as Genetic Algorithm (GA), Genetic Programming (GP), Evolutionary Programming (EP), and Evolution Strategies (ES) have been studied. Recently, a new approach named Genetic Network Programming (GNP) has been proposed. GNP can evolve itself and find the optimal solution. It is based on the idea of Genetic Algorithm and uses the data structure of directed graphs. Many papers have demonstrated that GNP can deal with complex problems in the dynamic environments very efficiently and effectively. As a result, recently, GNP is getting more and more attentions and is used in many different areas such as data mining, extracting trading rules of stock markets, elevator supervised control systems, etc., and GNP has obtained some outstanding results. Focusing on the GNP's distinguished expression ability of the graph structure, this paper proposes a method named Genetic Network Programming with Reconstructed Individuals (GNP-RI). The aim of GNP-RI is to balance the exploitation and exploration of GNP, that is, to strengthen the exploitation ability by using the exploited information extensively during the evolution process of GNP and finally obtain better performances than that of GNP. In the proposed method, the worse individuals are reconstructed and enhanced by the elite information before undergoing genetic operations (mutation and crossover). The enhancement of worse individuals mimics the maturing phenomenon in nature, where bad individuals can become smarter after receiving a good education. In this paper, GNP-RI is applied to the tile-world problem which is an excellent bench mark for evaluating the proposed architecture. The performance of GNP-RI is compared with that of the conventional GNP. The simulation results show some advantages of GNP-RI demonstrating its superiority over the conventional GNPs.
Lagally, Eric T; Mathies, Richard A
With the completion of the Human Genome Project and the ongoing DNA sequencing of the genomes of other animals, bacteria, plants and others, a wealth of new information about the genetic composition of organisms has become available. However, as the demand for sequence information grows, so does the workload required both to generate this sequence and to use it for targeted genetic analysis. Microfabricated genetic analysis systems are well poised to assist in the collection and use of these data through increased analysis speed, lower analysis cost and higher parallelism leading to increased assay throughput. In addition, such integrated microsystems may point the way to targeted genetic experiments on single cells and in other areas that are otherwise very difficult. Concomitant with these advantages, such systems, when fully integrated, should be capable of forming portable systems for high-speed in situ analyses, enabling a new standard in disciplines such as clinical chemistry, forensics, biowarfare detection and epidemiology. This review will discuss the various technologies available for genetic analysis on the microscale, and efforts to integrate them to form fully functional robust analysis devices. (topical review)
Full Text Available Gallstone disease is a complex disorder where both environmental and genetic factors contribute towards susceptibility to the disease. Epidemiological and family studies suggest a strong genetic component in the causation of this disease. Several genetically derived phenotypes in the population are responsible for variations in lipoprotein types, which in turn affect the amount of cholesterol available in the gall bladder. The genetic polymorphisms in various genes for apo E, apo B, apo A1, LDL receptor, cholesteryl ester transfer and LDL receptor-associated protein have been implicated in gallstone formation. However, presently available information on genetic differences is not able to account for a large number of gallstone patients. The molecular studies in the animal models have not only confirmed the present paradigm of gallstone formation but also helped in identification of novel genes in humans, which might play an important role in pathogenesis of the disease. Precise understanding of such genes and their molecular mechanisms may provide the basis of new targets for rational drug designs and dietary interventions.
Full Text Available ... Physician Resources Professions Site Index A-Z Computed Tomography (CT) - Sinuses Computed tomography (CT) of the sinuses ... CT of the Sinuses? What is CT (Computed Tomography) of the Sinuses? Computed tomography, more commonly known ...
This book provides the following information: basic aspects of computed tomography; atlas of computed tomography of the normal adult; clinical application of computed tomography; and radiotherapy planning and computed tomography
Hallinan, Dara; de Hert, Paul; Taylor, L.; Floridi, L.; van der Sloot, B.
Each person shares genetic code with others. Thus, one individual’s genome can reveal information about other individuals. When multiple individuals share aspects of genetic architecture, they form a ‘genetic group’. From a social and legal perspective, two types of genetic group exist: Those which
Hyndman, D E
Analog and Hybrid Computing focuses on the operations of analog and hybrid computers. The book first outlines the history of computing devices that influenced the creation of analog and digital computers. The types of problems to be solved on computers, computing systems, and digital computers are discussed. The text looks at the theory and operation of electronic analog computers, including linear and non-linear computing units and use of analog computers as operational amplifiers. The monograph examines the preparation of problems to be deciphered on computers. Flow diagrams, methods of ampl
In the introductory chapter we define the concept of cloud computing and cloud services, and we introduce layers and types of cloud computing. We discuss the differences between cloud computing and cloud services. New technologies that enabled cloud computing are presented next. We also discuss cloud computing features, standards, and security issues. We introduce the key cloud computing platforms, their vendors, and their offerings. We discuss cloud computing challenges and the future of cloud computing.
Kottyan, L C; Rothenberg, M E
Eosinophilic esophagitis (EoE) is a chronic, allergic disease associated with marked mucosal eosinophil accumulation. EoE disease risk is multifactorial and includes environmental and genetic factors. This review will focus on the contribution of genetic variation to EoE risk, as well as the experimental tools and statistical methodology used to identify EoE risk loci. Specific disease-risk loci that are shared between EoE and other allergic diseases (TSLP, LRRC32) or unique to EoE (CAPN14), as well as Mendellian Disorders associated with EoE, will be reviewed in the context of the insight that they provide into the molecular pathoetiology of EoE. We will also discuss the clinical opportunities that genetic analyses provide in the form of decision support tools, molecular diagnostics, and novel therapeutic approaches.
Wang, Haina; Peng, Nan; Shah, Shiraz Ali
SUMMARY: Research on archaeal extrachromosomal genetic elements (ECEs) has progressed rapidly in the past decade. To date, over 60 archaeal viruses and 60 plasmids have been isolated. These archaeal viruses exhibit an exceptional diversity in morphology, with a wide array of shapes, such as spind......SUMMARY: Research on archaeal extrachromosomal genetic elements (ECEs) has progressed rapidly in the past decade. To date, over 60 archaeal viruses and 60 plasmids have been isolated. These archaeal viruses exhibit an exceptional diversity in morphology, with a wide array of shapes...... on archaeal ECEs has just started to unravel the molecular biology of these genetic entities and their interactions with archaeal hosts, it is expected to accelerate in the next decade....
Kermani, Bahram G
Crystal Genetics, Inc. is an early-stage genetic test company, focused on achieving the highest possible clinical-grade accuracy and comprehensiveness for detecting germline (e.g., in hereditary cancer) and somatic (e.g., in early cancer detection) mutations. Crystal's mission is to significantly improve the health status of the population, by providing high accuracy, comprehensive, flexible and affordable genetic tests, primarily in cancer. Crystal's philosophy is that when it comes to detecting mutations that are strongly correlated with life-threatening diseases, the detection accuracy of every single mutation counts: a single false-positive error could cause severe anxiety for the patient. And, more importantly, a single false-negative error could potentially cost the patient's life. Crystal's objective is to eliminate both of these error types.
This paper provides part of an analysis of the use of the Maori term whakapapa in a study designed to test the compatibility and commensurability of views of members of the indigenous culture of New Zealand with other views of genetic technologies extant in the country. It is concerned with the narrow sense of whakapapa as denoting biological ancestry, leaving the wider sense of whakapapa as denoting cultural identity for discussion elsewhere. The phenomenon of genetic curiosity is employed to facilitate this comparison. Four levels of curiosity are identified, in the Maori data, which penetrate more or less deeply into the psyche of individuals, affecting their health and wellbeing. These phenomena are compared with non-Maori experiences and considerable commonalities are discovered together with a point of marked difference. The results raise important questions for the ethical application of genetic technologies. © 2010 Blackwell Publishing Ltd.
Axelrod, Felicia B
Genetic disorders affecting the autonomic nervous system can result in abnormal development of the nervous system or they can be caused by neurotransmitter imbalance, an ion-channel disturbance or by storage of deleterious material. The symptoms indicating autonomic dysfunction, however, will depend upon whether the genetic lesion has disrupted peripheral or central autonomic centers or both. Because the autonomic nervous system is pervasive and affects every organ system in the body, autonomic dysfunction will result in impaired homeostasis and symptoms will vary. The possibility of genetic confirmation by molecular testing for specific diagnosis is increasing but treatments tend to remain only supportive and directed toward particular symptoms. Copyright © 2013 Elsevier Inc. All rights reserved.
Andréa Poyastro Pinheiro
Full Text Available OBJECTIVE: To review the recent literature relevant to genetic research in eating disorders and to discuss unique issues which are crucial for the development of a genetic research project in eating disorders in Brazil. METHOD: A computer literature review was conducted in the Medline database between 1984 and may 2005 with the search terms "eating disorders", "anorexia nervosa", "bulimia nervosa", "binge eating disorder", "family", "twin" and "molecular genetic" studies. RESULTS: Current research findings suggest a substantial influence of genetic factors on the liability to anorexia nervosa and bulimia nervosa. Genetic research with admixed populations should take into consideration sample size, density of genotyping and population stratification. Through admixture mapping it is possible to study the genetic structure of admixed human populations to localize genes that underlie ethnic variation in diseases or traits of interest. CONCLUSIONS: The development of a major collaborative genetics initiative of eating disorders in Brazil and South America would represent a realistic possibility of studying the genetics of eating disorders in the context of inter ethnic groups, and also integrate a new perspective on the biological etiology of eating disorders.
Wei, Changshuai; Elston, Robert C; Lu, Qing
Converging evidence suggests that common complex diseases with the same or similar clinical manifestations could have different underlying genetic etiologies. While current research interests have shifted toward uncovering rare variants and structural variations predisposing to human diseases, the impact of heterogeneity in genetic studies of complex diseases has been largely overlooked. Most of the existing statistical methods assume the disease under investigation has a homogeneous genetic effect and could, therefore, have low power if the disease undergoes heterogeneous pathophysiological and etiological processes. In this paper, we propose a heterogeneity-weighted U (HWU) method for association analyses considering genetic heterogeneity. HWU can be applied to various types of phenotypes (e.g., binary and continuous) and is computationally efficient for high-dimensional genetic data. Through simulations, we showed the advantage of HWU when the underlying genetic etiology of a disease was heterogeneous, as well as the robustness of HWU against different model assumptions (e.g., phenotype distributions). Using HWU, we conducted a genome-wide analysis of nicotine dependence from the Study of Addiction: Genetics and Environments dataset. The genome-wide analysis of nearly one million genetic markers took 7h, identifying heterogeneous effects of two new genes (i.e., CYP3A5 and IKBKB) on nicotine dependence. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
This paper investigates a variety of unconventional quantum computation devices, including fermionic quantum computers and computers that exploit nonlinear quantum mechanics. It is shown that unconventional quantum computing devices can in principle compute some quantities more rapidly than `conventional' quantum computers.
Firoozi, Z.; Ismail, N.; Ariafar, S. H.; Tang, S. H.; Ariffin, M. K. M. A.
Network design is by nature costly and optimization models play significant role in reducing the unnecessary cost components of a distribution network. This study proposes a genetic algorithm to solve a distribution network design model. The structure of the chromosome in the proposed algorithm is defined in a novel way that in addition to producing feasible solutions, it also reduces the computational complexity of the algorithm. Computational results are presented to show the algorithm performance.
Gonzalez, Teofilo; Tucker, Allen
Overview of Computer Science Structure and Organization of Computing Peter J. DenningComputational Thinking Valerie BarrAlgorithms and Complexity Data Structures Mark WeissBasic Techniques for Design and Analysis of Algorithms Edward ReingoldGraph and Network Algorithms Samir Khuller and Balaji RaghavachariComputational Geometry Marc van KreveldComplexity Theory Eric Allender, Michael Loui, and Kenneth ReganFormal Models and Computability Tao Jiang, Ming Li, and Bala
Alexandre Rezende Vieira
Full Text Available Caries remains the most prevalent non-contagious infectious disease in humans. It is clear that the current approaches to decrease the prevalence of caries in human populations, including water fluoridation and school-based programs, are not enough to protect everyone. The scientific community has suggested the need for innovative work in a number of areas in cariology, encompassing disease etiology, epidemiology, definition, prevention, and treatment. We have pioneered the work on genetic studies to identify genes and genetic markers of diagnostic, prognostic, and therapeutic value. This paper summarizes a presentation that elaborated on these initial findings.
Garcia, Erin C
Burkholderia thailandensis is a Gram-negative bacterium endemic to Southeast Asian and northern Australian soils. It is non-pathogenic; therefore, it is commonly used as a model organism for the related human pathogens Burkholderia mallei and Burkholderia pseudomallei. B. thailandensis is relatively easily genetically manipulated and a variety of robust genetic tools can be used in this organism. This unit describes protocols for conjugation, natural transformation, mini-Tn7 insertion, and allelic exchange in B. thailandensis. © 2017 by John Wiley & Sons, Inc. Copyright © 2017 John Wiley & Sons, Inc.
Full Text Available The article discusses some paradigms of artificial intelligence in the context of their applications in computer financial systems. The proposed approach has a significant po-tential to increase the competitiveness of enterprises, including financial institutions. However, it requires the effective use of supercomputers, grids and cloud computing. A reference is made to the computing environment for Bitcoin. In addition, we characterized genetic programming and artificial neural networks to prepare investment strategies on the stock exchange market.
The article discusses some paradigms of artificial intelligence in the context of their applications in computer financial systems. The proposed approach has a significant po-tential to increase the competitiveness of enterprises, including financial institutions. However, it requires the effective use of supercomputers, grids and cloud computing. A reference is made to the computing environment for Bitcoin. In addition, we characterized genetic programming and artificial neural networks to p...
Mok, P. Y.; Wang, X. X.; Xu, J.; Kwok, Y. L.
Computer aided design is vitally important for the modern industry, particularly for the creative industry. Fashion industry faced intensive challenges to shorten the product development process. In this paper, a methodology is proposed for sketch design based on interactive genetic algorithms. The sketch design system consists of a sketch design model, a database and a multi-stage sketch design engine. First, a sketch design model is developed based on the knowledge of fashion design to describe fashion product characteristics by using parameters. Second, a database is built based on the proposed sketch design model to define general style elements. Third, a multi-stage sketch design engine is used to construct the design. Moreover, an interactive genetic algorithm (IGA) is used to accelerate the sketch design process. The experimental results have demonstrated that the proposed method is effective in helping laypersons achieve satisfied fashion design sketches.
Bates, Maxwell; Lachoff, Joe; Meech, Duncan; Zulkower, Valentin; Moisy, Anaïs; Luo, Yisha; Tekotte, Hille; Franziska Scheitz, Cornelia Johanna; Khilari, Rupal; Mazzoldi, Florencio; Chandran, Deepak; Groban, Eli
Genetic Constructor is a cloud Computer Aided Design (CAD) application developed to support synthetic biologists from design intent through DNA fabrication and experiment iteration. The platform allows users to design, manage, and navigate complex DNA constructs and libraries, using a new visual language that focuses on functional parts abstracted from sequence. Features like combinatorial libraries and automated primer design allow the user to separate design from construction by focusing on functional intent, and design constraints aid iterative refinement of designs. A plugin architecture enables contributions from scientists and coders to leverage existing powerful software and connect to DNA foundries. The software is easily accessible and platform agnostic, free for academics, and available in an open-source community edition. Genetic Constructor seeks to democratize DNA design, manufacture, and access to tools and services from the synthetic biology community.
Mahdevar, Ghasem; Zahiri, Javad; Sadeghi, Mehdi; Nowzari-Dalini, Abbas; Ahrabian, Hayedeh
Single Nucleotide Polymorphisms (SNPs) provide valuable information on human evolutionary history and may lead us to identify genetic variants responsible for human complex diseases. Unfortunately, molecular haplotyping methods are costly, laborious, and time consuming; therefore, algorithms for constructing full haplotype patterns from small available data through computational methods, Tag SNP selection problem, are convenient and attractive. This problem is proved to be an NP-hard problem, so heuristic methods may be useful. In this paper we present a heuristic method based on genetic algorithm to find reasonable solution within acceptable time. The algorithm was tested on a variety of simulated and experimental data. In comparison with the exact algorithm, based on brute force approach, results show that our method can obtain optimal solutions in almost all cases and runs much faster than exact algorithm when the number of SNP sites is large. Our software is available upon request to the corresponding author.
Goñi-Moreno, Angel; Amos, Martyn
Engineering genetic Boolean logic circuits is a major research theme of synthetic biology. By altering or introducing connections between genetic components, novel regulatory networks are built in order to mimic the behaviour of electronic devices such as logic gates. While electronics is a highly standardized science, genetic logic is still in its infancy, with few agreed standards. In this paper we focus on the interpretation of logical values in terms of molecular concentrations. We describe the results of computational investigations of a novel circuit that is able to trigger specific differential responses depending on the input standard used. The circuit can therefore be dynamically reconfigured (without modification) to serve as both a NAND/NOR logic gate. This multi-functional behaviour is achieved by a) varying the meanings of inputs, and b) using branch predictions (as in computer science) to display a constrained output. A thorough computational study is performed, which provides valuable insights for the future laboratory validation. The simulations focus on both single-cell and population behaviours. The latter give particular insights into the spatial behaviour of our engineered cells on a surface with a non-homogeneous distribution of inputs. We present a dynamically-reconfigurable NAND/NOR genetic logic circuit that can be switched between modes of operation via a simple shift in input signal concentration. The circuit addresses important issues in genetic logic that will have significance for more complex synthetic biology applications.
Voet, van der H.; Perry, J.N.; Amzal, B.; Paoletti, C.
Background - Safety assessment of genetically modified organisms is currently often performed by comparative evaluation. However, natural variation of plant characteristics between commercial varieties is usually not considered explicitly in the statistical computations underlying the assessment.
Stevenson, D. K.
In recent years, computational fluid dynamics has made significant progress in modelling aerodynamic phenomena. Currently, one of the major barriers to future development lies in the compute-intensive nature of the numerical formulations and the relative high cost of performing these computations on commercially available general purpose computers, a cost high with respect to dollar expenditure and/or elapsed time. Today's computing technology will support a program designed to create specialized computing facilities to be dedicated to the important problems of computational aerodynamics. One of the still unresolved questions is the organization of the computing components in such a facility. The characteristics of fluid dynamic problems which will have significant impact on the choice of computer architecture for a specialized facility are reviewed.
Hajri, S; Liouane, N; Hammadi, S; Borne, P
Most scheduling problems are highly complex combinatorial problems. However, stochastic methods such as genetic algorithm yield good solutions. In this paper, we present a controlled genetic algorithm (CGA) based on fuzzy logic and belief functions to solve job-shop scheduling problems. For better performance, we propose an efficient representational scheme, heuristic rules for creating the initial population, and a new methodology for mixing and computing genetic operator probabilities.
Mohammad Saad Zaghloul Salem
Nov 15, 2013 ... maps of gene loci based on information gathered, formerly, ... represented as figure or text interface data. Relevant ... The Egyptian Journal of Medical Human Genetics ... prophylactic management and genetic counseling. 17.
Acar, Sezer; Demir, Korcan; Shi, Yufei
Rickets is a metabolic bone disease that develops as a result of inadequate mineralization of growing bone due to disruption of calcium, phosphorus and/or vitamin D metabolism. Nutritional rickets remains a significant child health problem in developing countries. In addition, several rare genetic causes of rickets have also been described, which can be divided into two groups. The first group consists of genetic disorders of vitamin D biosynthesis and action, such as vitamin D-dependent rickets type 1A (VDDR1A), vitamin D-dependent rickets type 1B (VDDR1B), vitamin D-dependent rickets type 2A (VDDR2A), and vitamin D-dependent rickets type 2B (VDDR2B). The second group involves genetic disorders of excessive renal phosphate loss (hereditary hypophosphatemic rickets) due to impairment in renal tubular phosphate reabsorption as a result of FGF23-related or FGF23-independent causes. In this review, we focus on clinical, laboratory and genetic characteristics of various types of hereditary rickets as well as differential diagnosis and treatment approaches. PMID:29280738
... belongs to a class of genetic diseases called urea cycle disorders. Learn more about the genes associated with citrullinemia ... GeneReview: Citrin Deficiency GeneReview: Citrullinemia Type I GeneReview: Urea Cycle Disorders Overview MedlinePlus Encyclopedia: Hereditary Urea Cycle Abnormality National ...
Two widely-recognized committees, UNSCEAR and BEIR, have reevaluated their estimates of genetic risks from radiation. Their estimates for gene mutations are based on two different approaches, one being the doubling-dose approach and the other being a new direct approach based on an empirical determination of the amount of dominant induced damage in the skeletons of mice in the first generation following irradiation. The estimates made by these committees are in reasonably good agreement and suggest that the genetic risks from present exposures resultng from nuclear power production are small. There is room for much improvement in the reliability of the risk estimates. The relatively new approach of measuring the amount of induced damage to the mouse skeleton shows great promise of improving knowledge about how changes in the mutation frequency affect the incidence of genetic disorders. Such findings may have considerable influence on genetic risk estimates for radiation and on the development of risk estimates for other less-well-understood environmental mutagens. (author)
Full Text Available In a global society as the present, the nomenclature and terminology of diseases must be universally accepted among the specialists. This sentence is particularly true in some fields of medicine, as genetics, in which the progress of knowledge has been particularly rapid in last years.Many genetic disorders were termed using the names of the doctor (or the doctors who discovered and described them.The name of doctors and specialist were also frequently used to term sign and symptoms of diseases, including genetic syndromes.More rarely, a new disease received the name of the first patients described.In some cases the authors clearly proposed acronyms, that rapidly diffused as a good method to term genetic diseases and syndromes.Acronyms can be originated from the initial of main signs and symptoms; in some instances the acronym reproduces a word with other kind of semantic suggestions; some acronyms in their list of initials show also numbers, while others show also the initial of the words related to the physiopathology of disease.In more recent years acronyms were proposed to mark multicentric studies. Proceedings of the 10th International Workshop on Neonatology · Cagliari (Italy · October 22nd-25th, 2014 · The last ten years, the next ten years in Neonatology Guest Editors: Vassilios Fanos, Michele Mussap, Gavino Faa, Apostolos Papageorgiou
The extensive resource on ataxia has led to the development of a clinico-genetic ... Keywords: Cerebellar ataxias, SCAs, ARCAs, NGS, Gene network, iPSCs, .... Besides, mutations in different regions of the same gene result in different ..... integration with population data can also allow focussed testing/screening in specific.
John R. Jones; Norbert V. DeByle
The broad genotypic variability in quaking aspen (Populus tremuloides Michx.), that results in equally broad phenotypic variability among clones is important to the ecology and management of this species. This chapter considers principles of aspen genetics and variation, variation in aspen over its range, and local variation among clones. For a more...
Zhao, H; Lovett, B; Fang, W
Entomopathogenic fungi have been developed as environmentally friendly alternatives to chemical insecticides in biocontrol programs for agricultural pests and vectors of disease. However, mycoinsecticides currently have a small market share due to low virulence and inconsistencies in their performance. Genetic engineering has made it possible to significantly improve the virulence of fungi and their tolerance to adverse conditions. Virulence enhancement has been achieved by engineering fungi to express insect proteins and insecticidal proteins/peptides from insect predators and other insect pathogens, or by overexpressing the pathogen's own genes. Importantly, protein engineering can be used to mix and match functional domains from diverse genes sourced from entomopathogenic fungi and other organisms, producing insecticidal proteins with novel characteristics. Fungal tolerance to abiotic stresses, especially UV radiation, has been greatly improved by introducing into entomopathogens a photoreactivation system from an archaean and pigment synthesis pathways from nonentomopathogenic fungi. Conversely, gene knockout strategies have produced strains with reduced ecological fitness as recipients for genetic engineering to improve virulence; the resulting strains are hypervirulent, but will not persist in the environment. Coupled with their natural insect specificity, safety concerns can also be mitigated by using safe effector proteins with selection marker genes removed after transformation. With the increasing public concern over the continued use of synthetic chemical insecticides and growing public acceptance of genetically modified organisms, new types of biological insecticides produced by genetic engineering offer a range of environmentally friendly options for cost-effective control of insect pests. Copyright © 2016 Elsevier Inc. All rights reserved.
Ricano-Ponce, Isis; Wijmenga, Cisca; Gutierrez-Achury, Javier
New insights into the underlying molecular pathophysiology of celiac disease (CeD) over the last few years have been guided by major advances in the fields of genetics and genomics. The development and use of the Immunochip genotyping platform paved the way for the discovery of 39 non-HLA loci
MacClintic, Scott D.; Nelson, Genevieve M.
Bacterial transformation is a commonly used technique in genetic engineering that involves transferring a gene of interest into a bacterial host so that the bacteria can be used to produce large quantities of the gene product. Although several kits are available for performing bacterial transformation in the classroom, students do not always…
All definitions of the metabolic syndrome include some form of obesity as one of the possible features. Body mass index (BMI) has a known genetic component, currently estimated to account for about 70% of the population variance in weight status for non-syndromal obesity. Much research effort has be...
Genetic studies have helped us gain basic knowledge of the Tamarix invasion. We now have a better understanding of the species identities involved in the invasion, their evolutionary relationships, and the contribution of hybridization to the invasion. This information can be used to enhance the eff...
Sternberg, Robert J.; Grigorenko, Elena L.; Kidd, Kenneth K.
In this article, the authors argue that the overwhelming portion of the literature on intelligence, race, and genetics is based on folk taxonomies rather than scientific analysis. They suggest that because theorists of intelligence disagree as to what it is, any consideration of its relationships to other constructs must be tentative at best. They…
Conclusions: Over the last two decades, PGD has been shown to be a reliable and safe genetic test for couples who are at risk of a specific inher - ited disorder. For PGS, the results from several ongoing randomized controlled trials performed at different cell biopsy stage, using array-CGH and SNP array will provide the data needed to evaluate the clinical efficacy.
25, 191–194; reprinted in this issue as a J. Genet. classic, pages 3–7) ... 1932, starts with Punnett explaining that he started his work on ducks as he was asked ... text of Punnett's later comments on genes and human mental disease. Punnett's ...
The difficulties of quantifying genetic radiation effects are discussed, with reference to studies of atomic bomb survivors, and mouse germ-cells. Doubling dose methods of extrapolation and the problems of quantifying risks of diseases of irregular inheritance are also considered. (U.K.)
Ardekani, Ali M
In the past decade, the human genome has been completely sequenced and the knowledge from it has begun to influence the fields of biological and social sciences in fundamental ways. Identification of about 25000 genes in the human genome is expected to create great benefits in diagnosis and treatment of diseases in the coming years. However, Genetic technologies have also created many interesting and difficult ethical issues which can affect the human societies now and in the future. Application of genetic technologies in the areas of stem cells, cloning, gene therapy, genetic manipulation, gene selection, sex selection and preimplantation diagnosis has created a great potential for the human race to influence and change human life on earth as we know it today. Therefore, it is important for leaders of societies in the modern world to pay attention to the advances in genetic technologies and prepare themselves and those institutions under their command to face the challenges which these new technologies induce in the areas of ethics, law and social policies.
Full Text Available In the first paper on the syndrome of autism, Kanner described it as innate and inborn. He drew attention to the abnormalities in infancy without evidence of prior normal development and the intellectual, non emotional qualities shown by many of the parents and grandparents. Subsequently, the supposed lack of parental warmth led many clinicians to abandon the notions of constitutional deficit in the child and instead to postulate a psychogenic origin etiology was likely, genetic factors probably did not play a major role. Attention was draw to the low rate of autism in siblings, the lack of chromosome anomalies, and the similarities with syndromes associated with known brain trauma. Although the rate of autism in siblings was indeed low, it was much higher than in the general population rate providing a strong pointer to the genetic factors. The recognition that this was so, associated with the parallel finding of apparently high familiar loading for language delay, stimulated the first, systematic, twin study of autism, which suggested a strong genetic component. Subsequent research has produced findings in the same direction, although many questions remain unanswered. In this paper the evidence that has accumulated on genetic influences on autism is summarized and the remained dilemmas on this field are discussed.
Rosellini, D; Veronesi, F [Dipartimento di Biologia Vegetale e Biotecnologie Agroambientali e Zootecniche, Universita degli Studi di Perugia, Borgo XX giugno 74, 06121 Perugia (Italy)
The application of genetic engineering to plants has provided genetically modified plants (GMPs, or transgenic plants) that are cultivated worldwide on increasing areas. The most widespread GMPs are herbicide-resistant soybean and canola and insect-resistant corn and cotton. New GMPs that produce vaccines, pharmaceutical or industrial proteins, and fortified food are approaching the market. The techniques employed to introduce foreign genes into plants allow a quite good degree of predictability of the results, and their genome is minimally modified. However, some aspects of GMPs have raised concern: (a) control of the insertion site of the introduced DNA sequences into the plant genome and of its mutagenic effect; (b) presence of selectable marker genes conferring resistance to an antibiotic or an herbicide, linked to the useful gene; (c) insertion of undesired bacterial plasmid sequences; and (d) gene flow from transgenic plants to non-transgenic crops or wild plants. In response to public concerns, genetic engineering techniques are continuously being improved. Techniques to direct foreign gene integration into chosen genomic sites, to avoid the use of selectable genes or to remove them from the cultivated plants, to reduce the transfer of undesired bacterial sequences, and make use of alternative, safer selectable genes, are all fields of active research. In our laboratory, some of these new techniques are applied to alfalfa, an important forage plant. These emerging methods for plant genetic engineering are briefly reviewed in this work.
Rosellini, D; Veronesi, F
The application of genetic engineering to plants has provided genetically modified plants (GMPs, or transgenic plants) that are cultivated worldwide on increasing areas. The most widespread GMPs are herbicide-resistant soybean and canola and insect-resistant corn and cotton. New GMPs that produce vaccines, pharmaceutical or industrial proteins, and fortified food are approaching the market. The techniques employed to introduce foreign genes into plants allow a quite good degree of predictability of the results, and their genome is minimally modified. However, some aspects of GMPs have raised concern: (a) control of the insertion site of the introduced DNA sequences into the plant genome and of its mutagenic effect; (b) presence of selectable marker genes conferring resistance to an antibiotic or an herbicide, linked to the useful gene; (c) insertion of undesired bacterial plasmid sequences; and (d) gene flow from transgenic plants to non-transgenic crops or wild plants. In response to public concerns, genetic engineering techniques are continuously being improved. Techniques to direct foreign gene integration into chosen genomic sites, to avoid the use of selectable genes or to remove them from the cultivated plants, to reduce the transfer of undesired bacterial sequences, and make use of alternative, safer selectable genes, are all fields of active research. In our laboratory, some of these new techniques are applied to alfalfa, an important forage plant. These emerging methods for plant genetic engineering are briefly reviewed in this work
As a result of many years of domestication and selection for desirable fruit quality, watermelon cultivars (Citrullus lanatus) share a narrow genetic base. Africa is the center of origin and diversity of watermelon and is considered to be the central continent for collecting and conserving useful ge...
Seager, Robert D.
In learning genetics, many students misunderstand and misinterpret what "dominance" means. Understanding is easier if students realize that dominance is not a mechanism, but rather a consequence of underlying cellular processes. For example, metabolic pathways are often little affected by changes in enzyme concentration. This means that…
... Some studies suggest that additional genetic changes can influence the development of retinoblastoma ; these changes may help explain variations ... usually occurs in childhood, typically leading to the development of ... and there is no family history of the disease. Affected individuals are born ...
Laine, M.L.; Crielaard, W.; Loos, B.G.
In this systematic review, we explore and summarize the peer-reviewed literature on putative genetic risk factors for susceptibility to aggressive and chronic periodontitis. A comprehensive literature search on the PubMed database was performed using the keywords ‘periodontitis’ or ‘periodontal
Djémié, Tania; Weckhuysen, Sarah; von Spiczak, Sarah
BACKGROUND: Sanger sequencing, still the standard technique for genetic testing in most diagnostic laboratories and until recently widely used in research, is gradually being complemented by next-generation sequencing (NGS). No single mutation detection technique is however perfect in identifying...
Bratosin, C.C.; Sidorova, N.; Aalst, van der W.M.P.
Process mining aims at discovering process models from data logs in order to offer insight into the real use of information systems. Most of the existing process mining algorithms fail to discover complex constructs or have problems dealing with noise and infrequent behavior. The genetic process
pp 223-226 Commentary on J. Genet. Classic. Towards a genetic architecture of cryptic genetic variation and genetic assimilation: the contribution of K. G. Bateman · Ian Dworkin · More Details Fulltext PDF. pp 227-257 J. Genet. Classic. The Genetic Assimilation of Four Venation Phenocopies (Published on 1959 J. Genet.
Mar 15, 2012 ... Key words: short tandem repeat, repeat motif, genetic polymorphism, Han population, forensic genetics. INTRODUCTION. Short tandem repeat (STR) is widely .... Data analysis. The exact test of Hardy-Weinberg equilibrium was conducted with. Arlequin version 3.5 software (Computational and Molecular.
Snášel, V.; Platoš, J.; Krömer, P.; Húsek, Dušan; Frolov, A.
Roč. 17, č. 6 (2007), s. 675-688 ISSN 1210-0552 Institutional research plan: CEZ:AV0Z10300504 Keywords : data mining * genetic algorithms * Boolean factorization * binary data * machine learning * feature extraction Subject RIV: IN - Informatics, Computer Science Impact factor: 0.280, year: 2007
Sensitivity analysis of the summary odds ratio coefficients on the association between TNF-α-308G/A polymorphism and AILD risk using a random effects model. (A allele vs G allele). Results were computed by omitting each study in turn. Error bars are 95% confidence interval. Journal of Genetics, Vol. 92, No. 3, December ...
This paper presents a Genetic Algorithm software (which is a computational, search technique) for finding the zeros (roots) of any given polynomial function, and optimizing and solving N-dimensional systems of equations. The software is particularly useful since most of the classic schemes are not all embracing.
Cao, Ming; Lu, Ming; Zhang, Jian-Ping
This work presents a new approach for concrete plant operations optimization by combining a ready mixed concrete (RMC) production simulation tool (called HKCONSIM) with a genetic algorithm (GA) based optimization procedure. A revamped HKCONSIM computer system can be used to automate the simulation
Mouw, Evert; van't Noordende, Guido; van Kampen, Antoine H. C.; Louter, Baas; Santcroos, Mark; Olabarriaga, Silvia D.
European laws on privacy and data security are not explicit about the storage and processing of genetic data. Especially whole-genome data is identifying and contains a lot of personal information. Is processing of such data allowed in computing grids? To find out, we looked at legal precedents in
Apr 19, 2010 ... process computationally to describe the structure of the sys- tem and ... to the different mathematical formalisms used to model net-. Keywords. gene ..... All the algorithms were implemented in MATLAB 7.0 and run on all the 'gold ..... De Jong H. 2002 Medeling and simulation of genetic regulatory systems: a ...
The research of the scientific history and development status reflect the science and technology level of a nation. The genetic history is one of the branches of the life science and the 21st century is life science century. The genetics history in the teaching of genetics not only can help students get familiar with the birth and development of genetics, but also enhance their thinking ability and scientific qualities. The roles and approaches of teaching are discussed in this paper.
Blanco, R; Chakraborty, R
From dental casts of 94 parent-offspring and 127 full-sib pairs, sampled from two Chilean populations, shovelling indices are computed to measure the degree of shovelling of maxillary central incisors quantitatively. Genetic correlations are computed to determine the role of genetic factors in explaining the variation in this trait. Assuming only hereditary factors to be responsible for the transmission of shovel shape, 68% of total variability is ascribed to the additive effect of genes.
Full Text Available Pathogenesis of diabetes is still a mystery for medicine, the real challenge currently being the identification of genetic factors and specific mutations that cause the disease. Heterogeneity of diabetes hampers research, only a few loci inside the human genome being correlated with predisposition for disease till now. Insulin-dependent diabetes - IDDM (T1DM develops through autoimmune destruction of pancreatic beta cells. HLA complex on the short arm of chromosome 6 (6p21, where very important genes responsible for immunological condition of the person are located, plays a very important role in genetic predisposition for T1DM. Beside this region, there are also other loci in the human genome (on chromosomes 1, 2 and 11 where a correlation with T1DM has been shown. Correlation between HLA systems and T1DM was first described for class I alleles, but recently attention has been drawn to class II loci which seem to be the cause of primary predisposition for T1DM. In the case of non-insulin-dependent diabetes - NIDDM (T2DM, the situation proved to be even more complex. Only a few genetic loci on chromosomes 11, 13 and 20 and MODY variant on chromosomes 7 and 12 have been identified by now. There are two theories about genetic basis of T2DM: the first stipulates that the genetic predisposition is determined through numerous loci, each individually responsible for a small part of predisposition; the second claims that there are a limited number of "major" genes probably functioning on a polygenic basis. Further research in this area is definitely needed to enable an accurate calculation of the risks of the disease and possible consequences during a lifetime of a person.
Sloss, Brian L.; Klumb, Robert A.; Heist, Edward J.
The conservation of genetic diversity of our natural resources is overwhelmingly one of the central foci of 21st century management practices. Three recommendations related to the conservation of paddlefish Polyodon spathula genetic diversity are to (1) identify genetic diversity at both nuclear and mitochondrial DNA loci using a suggested list of 20 sampling locations, (2) use genetic diversity estimates to develop genetic management units, and (3) identify broodstock sources to minimize effects of supplemental stocking on the genetic integrity of native paddlefish populations. We review previous genetic work on paddlefish and described key principles and concepts associated with maintaining genetic diversity within and among paddlefish populations and also present a genetic case study of current paddlefish propagation at the U.S. Fish and Wildlife Service Gavins Point National Fish Hatchery. This study confirmed that three potential sources of broodfish were genetically indistinguishable at the loci examined, allowing the management agencies cooperating on this program flexibility in sampling gametes. This study also showed significant bias in the hatchery occurred in terms of male reproductive contribution, which resulted in a shift in the genetic diversity of progeny compared to the broodfish. This shift was shown to result from differential male contributions, partially attributed to the mode of egg fertilization. Genetic insights enable implementation of a paddlefish propagation program within an adaptive management strategy that conserves inherent genetic diversity while achieving demographic goals.
Discusses the claims for a brave new world of genetic manipulation" and concludes that if we could agree upon applying genetic (or any other effective) remedies to global problems we probably would need no rescourse to them. Suggests that effective methods of preventing genetic disease are prevention of mutations and detection and…
This thesis investigates the use of problem-specific knowledge to enhance a genetic algorithm approach to multiple-choice optimisation problems.It shows that such information can significantly enhance performance, but that the choice of information and the way it is included are important factors for success.Two multiple-choice problems are considered.The first is constructing a feasible nurse roster that considers as many requests as possible.In the second problem, shops are allocated to locations in a mall subject to constraints and maximising the overall income.Genetic algorithms are chosen for their well-known robustness and ability to solve large and complex discrete optimisation problems.However, a survey of the literature reveals room for further research into generic ways to include constraints into a genetic algorithm framework.Hence, the main theme of this work is to balance feasibility and cost of solutions.In particular, co-operative co-evolution with hierarchical sub-populations, problem structure exploiting repair schemes and indirect genetic algorithms with self-adjusting decoder functions are identified as promising approaches.The research starts by applying standard genetic algorithms to the problems and explaining the failure of such approaches due to epistasis.To overcome this, problem-specific information is added in a variety of ways, some of which are designed to increase the number of feasible solutions found whilst others are intended to improve the quality of such solutions.As well as a theoretical discussion as to the underlying reasons for using each operator,extensive computational experiments are carried out on a variety of data.These show that the indirect approach relies less on problem structure and hence is easier to implement and superior in solution quality.
Turcanu, C.; Alecu, L.; Craciunescu, T.; Niculae, C.
Computerized tomography being a non-destructive and non-evasive technique is frequently used in medical application to generate three dimensional images of objects. Genetic algorithms are efficient, domain independent for a large variety of problems. The proposed method produces good quality reconstructions even in case of very small number of projection angles. It requests no a priori knowledge about the solution and takes into account the statistical uncertainties. The main drawback of the method is the amount of computer memory and time needed. (author)
Full Text Available The splitting approach is developed for the numerical simulation of genetic regulatory networks with a stable steady-state structure. The numerical results of the simulation of a one-gene network, a two-gene network, and a p53-mdm2 network show that the new splitting methods constructed in this paper are remarkably more effective and more suitable for long-term computation with large steps than the traditional general-purpose Runge-Kutta methods. The new methods have no restriction on the choice of stepsize due to their infinitely large stability regions.
Martínez, Héctor P.; Yannakakis, Georgios N.
Automatic feature selection is a critical step towards the generation of successful computational models of affect. This paper presents a genetic search-based feature selection method which is developed as a global-search algorithm for improving the accuracy of the affective models built....... The method is tested and compared against sequential forward feature selection and random search in a dataset derived from a game survey experiment which contains bimodal input features (physiological and gameplay) and expressed pairwise preferences of affect. Results suggest that the proposed method...
A dictionary of more than 150 genetics-related terms written for healthcare professionals. This resource was developed to support the comprehensive, evidence-based, peer-reviewed PDQ cancer genetics information summaries.
... 5 links) Diagnostic Tests Drug Therapy Genetic Counseling Palliative Care Surgery and Rehabilitation Related Information How are genetic ... G, Tzioumi D, Sillence DO, Mowat D. Three patients with terminal deletions within the subtelomeric region of chromosome 9q. ...
... my area? Other Names for This Condition Diastrophic dwarfism DTD Related Information How are genetic conditions and ... 2 links) Health Topic: Bone Diseases Health Topic: Dwarfism Genetic and Rare Diseases Information Center (1 link) ...
Olufunmilayo F. Olopade MD, FACP, Professor of Medicine and Human Genetics and Director of the Cancer Risk Clinic Department of Medicine, BSD Section of Hematology/Oncology University of Chicago, presented "Clinical Cancer Genetics and Prevention".
... Thomas MG, Swallow DM. Lactose digestion and the evolutionary genetics of lactase persistence. Hum Genet. 2009 Jan; ... genome editing and CRISPR-Cas9? What is precision medicine? What is newborn screening? New Pages Alopecia areata ...
Wijk, van H.J.
This thesis describes the genetics of carcass composition and pork quality traits. A large population of commercial finishers was extensively phenotyped for growth, carcass composition and meat quality traits. Genetic parameters were estimated based on those measurements. The population was
Kottow, Miguel H
Genetics research has shown enormous developments in recent decades, although as yet with only limited clinical application. Bioethical analysis has been unable to deal with the vast problems of genetics because emphasis has been put on the principlism applied to both clinical and research bioethics. Genetics nevertheless poses its most complex moral dilemmas at the public level, where a social brand of ethics ought to supersede the essentially interpersonal perspective of principlism. A more social understanding of ethics in genetics is required to unravel issues such as research and clinical explorations, ownership and patents, genetic manipulation, and allocation of resources. All these issues require reflection based on the requirements of citizenry, consideration of common assets, and definition of public policies in regulating genetic endeavors and protecting the society as a whole Bioethics has privileged the approach to individual ethical issues derived from genetic intervention, thereby neglecting the more salient aspects of genetics and social ethics.
Mertens, Thomas R.; Robinson, Sandra K.
Describes different sources of readings for understanding issues and concepts of genetic engineering. Broad categories of reading materials are: concerns about genetic engineering; its background; procedures; and social, ethical and legal issues. References are listed. (PS)
... of CLPB is associated with congenital microcephaly, severe encephalopathy and 3-methylglutaconic aciduria. J Med Genet. 2015 ... genetic testing? What is precision medicine? What is newborn screening? New Pages LMNA-related congenital muscular dystrophy ...
... TJ, Vreeburg M, Rouhl RPW, Stevens SJC, Stegmann APA, Schieving J, Pfundt R, van Dijk K, Smeets ... article on PubMed Central More from Genetics Home Reference Bulletins Genetics Home Reference Celebrates Its 15th Anniversary ...
... may include a rounded upper back that also curves to the side ( kyphoscoliosis ), severely flattened bones of ... Information What information about a genetic condition can statistics provide? Why are some genetic conditions more common ...
... deformed hips, a rounded upper back that also curves to the side ( kyphoscoliosis ), and knees that are ... Information What information about a genetic condition can statistics provide? Why are some genetic conditions more common ...
... such as a rounded upper back that also curves to the side ( kyphoscoliosis ). Some people with Czech ... Information What information about a genetic condition can statistics provide? Why are some genetic conditions more common ...
... bones ( osteoporosis ) throughout the skeleton. These abnormalities make bones brittle and more prone to fracture. The bone abnormalities ... information about a genetic condition can statistics provide? Why are some genetic conditions more common in particular ...
Rapid technological advances in genetic analysis have revealed the genetic background of various diseases. Elucidation of the genes responsible for a disease enables better clinical management of the disease and helps to develop targeted drugs. Also, early diagnosis and management of at-risk family members can be made by identification of a genetic disease in the proband. On the other hand, genetic issues often cause psychological distress to the family. To perform genetic testing appropriately and to protect patients and family members from any harm, guidelines for genetic testing were released from the alliance of Japanese genetics-related academic societies in 2003. As genetic testing is becoming incorporated into clinical practice more broadly, the guideline was revised and released by the Japanese Society of Medical Sciences in 2011. All medical professionals in Japan are expected to follow this guideline.
Madsen, Kaj; NA NA NA Olesen, Dorte
Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)......Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)...
Hansen, Karl Damkjær
unmanned aircraft are used for aerial surveying of the crops. The farmer takes the role of the analyst above, who does not necessarily have any specific interest in remote controlled aircraft but needs the outcome of the survey. The recurring method in the study is the genetic algorithm; a flexible...... contributions are made in the area of the genetic algorithms. One is a method to decide on the right time to stop the computation of the plan, when the right balance is stricken between using the time planning and using the time flying. The other contribution is a characterization of the evolutionary operators...... used in the genetic algorithm. The result is a measure based on entropy to evaluate and control the diversity of the population of the genetic algorithm, which is an important factor its effectiveness....
Fry, T F
Further Computer Appreciation is a comprehensive cover of the principles and aspects in computer appreciation. The book starts by describing the development of computers from the first to the third computer generations, to the development of processors and storage systems, up to the present position of computers and future trends. The text tackles the basic elements, concepts and functions of digital computers, computer arithmetic, input media and devices, and computer output. The basic central processor functions, data storage and the organization of data by classification of computer files,
Bouckaert, Stefan; Vanhie-Van Gerwen, Jono; Moerman, Ingrid; Phillips, Stephen; Wilander, Jerker
The benchmarking concept is not new in the field of computing or computer networking. With “benchmarking tools”, one usually refers to a program or set of programs, used to evaluate the performance of a solution under certain reference conditions, relative to the performance of another solution. Since the 1970s, benchmarking techniques have been used to measure the performance of computers and computer networks. Benchmarking of applications and virtual machines in an Infrastructure-as-a-Servi...
Margolis, Jane; Goode, Joanna; Ryoo, Jean J.
Computer science programs are too often identified with a narrow stratum of the student population, often white or Asian boys who have access to computers at home. But because computers play such a huge role in our world today, all students can benefit from the study of computer science and the opportunity to build skills related to computing. The…
Feigenbaum, Edward A.; Nielsen, Norman R.
This article provides a current status report on the computing and computer science activities at Stanford University, focusing on the Computer Science Department, the Stanford Computation Center, the recently established regional computing network, and the Institute for Mathematical Studies in the Social Sciences. Also considered are such topics…
Foundations of Genetic Algorithms 1991 (FOGA 1) discusses the theoretical foundations of genetic algorithms (GA) and classifier systems.This book compiles research papers on selection and convergence, coding and representation, problem hardness, deception, classifier system design, variation and recombination, parallelization, and population divergence. Other topics include the non-uniform Walsh-schema transform; spurious correlations and premature convergence in genetic algorithms; and variable default hierarchy separation in a classifier system. The grammar-based genetic algorithm; condition
Herrera, Blanca M.; Keildson, Sarah; Lindgren, Cecilia M.
Obesity results from interactions between environmental and genetic factors. Despite a relatively high heritability of common, non-syndromic obesity (40?70%), the search for genetic variants contributing to susceptibility has been a challenging task. Genome wide association (GWA) studies have dramatically changed the pace of detection of common genetic susceptibility variants. To date, more than 40 genetic variants have been associated with obesity and fat distribution. However, since these v...
Computer and Information 2012
The series "Studies in Computational Intelligence" (SCI) publishes new developments and advances in the various areas of computational intelligence – quickly and with a high quality. The intent is to cover the theory, applications, and design methods of computational intelligence, as embedded in the fields of engineering, computer science, physics and life science, as well as the methodologies behind them. The series contains monographs, lecture notes and edited volumes in computational intelligence spanning the areas of neural networks, connectionist systems, genetic algorithms, evolutionary computation, artificial intelligence, cellular automata, self-organizing systems, soft computing, fuzzy systems, and hybrid intelligent systems. Critical to both contributors and readers are the short publication time and world-wide distribution - this permits a rapid and broad dissemination of research results. The purpose of the 11th IEEE/ACIS International Conference on Computer and Information Science (ICIS 2012...
... are some genetic conditions more common in particular ethnic groups? Genetic Changes A variety of genetic and environmental factors are likely involved in the development of ulcerative colitis . Recent studies have identified variations in dozens of genes that may be linked ...
Constance I. Millar; Diane L. Delany; Lawrence A. Riggs
In forestry the importance of genetic variation for successful reproduction, survival and growth has been widely documented for commercial conifers; until recently, little genetic work has been done on the California oaks. Even before the nature of genetic variation was scientifically investigated, its importance was suspected in operational forestry. Many failures of...
Stewart, J. Bird
Claims that most instruction dealing with genetics is limited to sex education and personal hygiene. Suggests that the biology curriculum should begin to deal with other issues related to genetics, including genetic normality, prenatal diagnoses, race, and intelligence. Predicts these topics will begin to appear in British examination programs.…
Boone, C. Keith
Discusses the main ethical issues generated by the new genetics and suggests ways to think about them. Concerns include "playing God," violation of the natural order of the universe, and abuse of genetic technology. Critical distinctions for making difficult decisions about genetic engineering issues are noted. (DH)
The task of the genetic counselor who identifies genetic causes of mental retardation and assists families to understand risk of recurrence is described. Considered are chromosomal genetic disorders such as Down's syndrome, inherited disorders such as Tay-Sachs disease, identification by testing the amniotic fluid cells (amniocentresis) in time…
Prins, J.C.P.; Smant, G.; Jansen, R.C.
Genetical genomics combines acquired high-throughput genomic data with genetic analysis. In this chapter, we discuss the application of genetical genomics for evolutionary studies, where new high-throughput molecular technologies are combined with mapping quantitative trait loci (QTL) on the genome
The mutagenic effects of ionising radiation on germ cells with resulting genetic abnormalities in subsequent generations, are considered. Having examined a simple model to explain the interaction of ionising radiation with genetic material and discussed its limitations, the methods whereby mutations are transmitted are discussed. Methods of estimating genetic risks and the results of such studies are examined. (U.K.)
In this review, the recent progress on genetic transformation of forest trees were discussed. Its described also, different applications of genetic engineering for improving forest trees or understanding the mechanisms governing genes expression in woody plants. Key words: Genetic transformation, transgenic forest trees, ...
Reconfigurable computing is emerging as an important area of research in computer architectures and software systems. Many algorithms can be greatly accelerated by placing the computationally intense portions of an algorithm into reconfigurable hardware. Reconfigurable computing combines many benefits of both software and ASIC implementations. Like software, the mapped circuit is flexible, and can be changed over the lifetime of the system. Similar to an ASIC, reconfigurable systems provide a method to map circuits into hardware. Reconfigurable systems therefore have the potential to achieve far greater performance than software as a result of bypassing the fetch-decode-execute operations of traditional processors, and possibly exploiting a greater level of parallelism. Such a field, where there is many different algorithms which can be accelerated, is an artificial intelligence. This paper presents example hardware implementations of Artificial Neural Networks, Genetic Algorithms and Expert Systems.
Houfek, Julia Fisco; Soltis-Vaughan, Brigette S; Atwood, Jan R; Reiser, Gwendolyn M; Schaefer, G Bradley
This study described the perceptions of genetic counseling and testing of adults (N = 116) attending a genetic education program. Understanding perceptions of genetic counseling, including the importance of counseling topics, will contribute to patient-focused care as clinical genetic applications for common, complex disorders evolve. Participants completed a survey addressing: the importance of genetic counseling topics, benefits and negative effects of genetic testing, and sharing test results. Topics addressing practical information about genetic conditions were rated most important; topics involving conceptual genetic/genomic principles were rated least important. The most frequently identified benefit and negative effect of testing were prevention/early detection/treatment and psychological distress. Participants perceived that they were more likely to share test results with first-degree than other relatives. Findings suggest providing patients with practical information about genetic testing and genetic contributions to disease, while also determining whether their self-care abilities would be enhanced by teaching genetic/genomic principles. Copyright © 2014 Elsevier Inc. All rights reserved.
Libbrecht, Maxwell W; Noble, William Stafford
The field of machine learning, which aims to develop computer algorithms that improve with experience, holds promise to enable computers to assist humans in the analysis of large, complex data sets. Here, we provide an overview of machine learning applications for the analysis of genome sequencing data sets, including the annotation of sequence elements and epigenetic, proteomic or metabolomic data. We present considerations and recurrent challenges in the application of supervised, semi-supervised and unsupervised machine learning methods, as well as of generative and discriminative modelling approaches. We provide general guidelines to assist in the selection of these machine learning methods and their practical application for the analysis of genetic and genomic data sets.
Rothstein, M.A. [ed.
Few developments are likely to affect human beings more profoundly in the long run than the discoveries resulting from advances in modern genetics. Although the developments in genetic technology promise to provide many additional benefits, their application to genetic screening poses ethical, social, and legal questions, many of which are rooted in issues of privacy and confidentiality. The ethical, practical, and legal ramifications of these and related questions are explored in depth. The broad range of topics includes: the privacy and confidentiality of genetic information; the challenges to privacy and confidentiality that may be projected to result from the emerging genetic technologies; the role of informed consent in protecting the confidentiality of genetic information in the clinical setting; the potential uses of genetic information by third parties; the implications of changes in the health care delivery system for privacy and confidentiality; relevant national and international developments in public policies, professional standards, and laws; recommendations; and the identification of research needs.
Holtzman, N A
Pressures to lower health-care costs remain an important stimulus to eugenic approaches. Prenatal diagnosis followed by abortion of affected fetuses has replaced sterilization as the major eugenic technique. Voluntary acceptance has replaced coercion, but subtle pressures undermine personal autonomy. The failure of the old eugenics to accurately predict who will have affected offspring virtually disappears when prenatal diagnosis is used to predict Mendelian disorders. However, when prenatal diagnosis is used to detect inherited susceptibilities to adult-onset, common, complex disorders, considerable uncertainty is inherent in the prediction. Intolerance and the resurgence of genetic determinism are current pressures for a eugenic approach. The increasing use of carrier screening (to identify those at risk of having affected offspring) and of prenatal diagnosis could itself generate intolerance for those who refuse the procedures. Genetic determinism deflects society from social action that would reduce the burden of disease far more than even the maximum use of eugenics.
Harris, H.; Hirschhorn, K. (eds.)
This book has five chapters covering peroxisomal diseases, X-linked immunodeficiencies, genetic mutations affecting human lipoproteins and their receptors and enzymes, genetic aspects of cancer, and Gaucher disease. The chapter on peroxisomes covers their discovery, structure, functions, disorders, etc. The chapter on X-linked immunodeficiencies discusses such diseases as agammaglobulinemia, severe combined immunodeficiency, Wiskott-Aldrich syndrome, animal models, linkage analysis, etc. Apolipoprotein formation, synthesis, gene regulation, proteins, etc. are the main focus of chapter 3. The chapter on cancer covers such topics as oncogene mapping and the molecular characterization of some recessive oncogenes. Gaucher disease is covered from its diagnosis, classification, and prevention, to its organ system involvement and molecular biology.
Zhu, Ena C; Soundy, Timothy J; Hu, Yueshan
Consuming excessive amounts of alcohol has the potential to modify an individual's brain and lead to alcohol dependence. Alcohol use leads to 88,000 deaths every year in the U.S. alone and can lead to other health issues including cancers, such as colorectal cancer, and mental health problems. While drinking behavior varies due to environmental factors, genetic factors also contribute to the risk of alcoholism. Certain genes affecting alcohol metabolism and neurotransmitters have been found to contribute to or inhibit the risk. Geneenvironment interactions may also play a role in the susceptibility of alcoholism. With a better understanding of the different components that can contribute to alcoholism, more personalized treatment could cater to the individual. This review discusses the major genetic factors and some small variants in other genes that contribute to alcoholism, as well as considers the gene-environmental interactions. Copyright© South Dakota State Medical Association.
Full Text Available Fast development of medical genetics and it’s subdisciplines is noticed in last thirty years. Modern diagnostic methods made possible to establish human genome and its impairment. In human genetics, ethic is main principle in working. Ethic is science about biggest goodness for human or society, and its aim protecting human health.Today's conditions for leaving and science development open a wide way for ethical approaches, but also for non-ethical manipulations with human even before his conception. We must keep to attitude that without law, with our behavior will must conduct our conscience. It is best to have neutral eugenetic attitude, which allows free ethical choice of each individual, in any case, for the well being of man.
Bedate, C.A.; Morales, J.C.; Lopez, E.H.
The objective of this book is to encourage the use of genetic engineering for economic development. The report covers: (1) Precedents of genetic engineering; (2) a brief description of the technology, including the transfer of DNA in bacteria (vectors, E. coli and B. subtilis hosts, stages, and technical problems), practical examples of techniques used and their products (interferon; growth hormone; insulin; treatment of blood cells, Talasemia, and Lesch-Nyhan syndrome; and more nutritious soya), transfer to higher organisms, and cellular fusion; (3) biological risks and precautions; (4) possible applications (production of hydrogen, hydrocarbons, alcohol, chemicals, enzymes, peptides, viral antigens, monoclonal antibodies, genes, proteins, and insecticides; metal extraction; nitrogen fixation; biodegradation; and new varieties of plants and animals; and (5) international activities.
Otlowski, M; Taylor, S; Bombard, Y
Genetic discrimination (GD) is a complex, multifaceted ethical, psychosocial, and legal phenomenon. It is defined as the differential treatment of asymptomatic individuals or their relatives on the basis of their real or assumed genetic characteristics. This article presents an overview of GD within the contemporary international context. It describes the concept of GD and its contextual features, reviews research evidence regarding people's experiences of GD and the impact of GD within a range of domains, and provides an overview of legal and policy responses to GD that have emerged globally. We argue that GD is a significant and internationally established phenomenon that requires multilevel responses to ensure social justice and equitable outcomes for all citizens. Future research should monitor GD and its impacts within the community as well as institutions and should evaluate the effectiveness of legislative, policy, community education, and systemic responses.