WorldWideScience

Sample records for site prediction methods

  1. Polyadenylation site prediction using PolyA-iEP method.

    Science.gov (United States)

    Kavakiotis, Ioannis; Tzanis, George; Vlahavas, Ioannis

    2014-01-01

    This chapter presents a method called PolyA-iEP that has been developed for the prediction of polyadenylation sites. More precisely, PolyA-iEP is a method that recognizes mRNA 3'ends which contain polyadenylation sites. It is a modular system which consists of two main components. The first exploits the advantages of emerging patterns and the second is a distance-based scoring method. The outputs of the two components are finally combined by a classifier. The final results reach very high scores of sensitivity and specificity.

  2. Method of predicting Splice Sites based on signal interactions

    Directory of Open Access Journals (Sweden)

    Deogun Jitender S

    2006-04-01

    Full Text Available Abstract Background Predicting and proper ranking of canonical splice sites (SSs is a challenging problem in bioinformatics and machine learning communities. Any progress in SSs recognition will lead to better understanding of splicing mechanism. We introduce several new approaches of combining a priori knowledge for improved SS detection. First, we design our new Bayesian SS sensor based on oligonucleotide counting. To further enhance prediction quality, we applied our new de novo motif detection tool MHMMotif to intronic ends and exons. We combine elements found with sensor information using Naive Bayesian Network, as implemented in our new tool SpliceScan. Results According to our tests, the Bayesian sensor outperforms the contemporary Maximum Entropy sensor for 5' SS detection. We report a number of putative Exonic (ESE and Intronic (ISE Splicing Enhancers found by MHMMotif tool. T-test statistics on mouse/rat intronic alignments indicates, that detected elements are on average more conserved as compared to other oligos, which supports our assumption of their functional importance. The tool has been shown to outperform the SpliceView, GeneSplicer, NNSplice, Genio and NetUTR tools for the test set of human genes. SpliceScan outperforms all contemporary ab initio gene structural prediction tools on the set of 5' UTR gene fragments. Conclusion Designed methods have many attractive properties, compared to existing approaches. Bayesian sensor, MHMMotif program and SpliceScan tools are freely available on our web site. Reviewers This article was reviewed by Manyuan Long, Arcady Mushegian and Mikhail Gelfand.

  3. Predicting proteasomal cleavage sites: a comparison of available methods

    DEFF Research Database (Denmark)

    Saxova, P.; Buus, S.; Brunak, Søren

    2003-01-01

    -terminal, in particular, of CTL epitopes is cleaved precisely by the proteasome, whereas the N-terminal is produced with an extension, and later trimmed by peptidases in the cytoplasm and in the endoplasmic reticulum. Recently, three publicly available methods have been developed for prediction of the specificity...

  4. Ensemble approach combining multiple methods improves human transcription start site prediction.

    LENUS (Irish Health Repository)

    Dineen, David G

    2010-01-01

    The computational prediction of transcription start sites is an important unsolved problem. Some recent progress has been made, but many promoters, particularly those not associated with CpG islands, are still difficult to locate using current methods. These methods use different features and training sets, along with a variety of machine learning techniques and result in different prediction sets.

  5. Prediction of allosteric sites on protein surfaces with an elastic-network-model-based thermodynamic method.

    Science.gov (United States)

    Su, Ji Guo; Qi, Li Sheng; Li, Chun Hua; Zhu, Yan Ying; Du, Hui Jing; Hou, Yan Xue; Hao, Rui; Wang, Ji Hua

    2014-08-01

    Allostery is a rapid and efficient way in many biological processes to regulate protein functions, where binding of an effector at the allosteric site alters the activity and function at a distant active site. Allosteric regulation of protein biological functions provides a promising strategy for novel drug design. However, how to effectively identify the allosteric sites remains one of the major challenges for allosteric drug design. In the present work, a thermodynamic method based on the elastic network model was proposed to predict the allosteric sites on the protein surface. In our method, the thermodynamic coupling between the allosteric and active sites was considered, and then the allosteric sites were identified as those where the binding of an effector molecule induces a large change in the binding free energy of the protein with its ligand. Using the proposed method, two proteins, i.e., the 70 kD heat shock protein (Hsp70) and GluA2 alpha-amino-3-hydroxy-5-methyl-4-isoxazole propionic acid (AMPA) receptor, were studied and the allosteric sites on the protein surface were successfully identified. The predicted results are consistent with the available experimental data, which indicates that our method is a simple yet effective approach for the identification of allosteric sites on proteins.

  6. Ensemble approach combining multiple methods improves human transcription start site prediction

    LENUS (Irish Health Repository)

    Dineen, David G

    2010-11-30

    Abstract Background The computational prediction of transcription start sites is an important unsolved problem. Some recent progress has been made, but many promoters, particularly those not associated with CpG islands, are still difficult to locate using current methods. These methods use different features and training sets, along with a variety of machine learning techniques and result in different prediction sets. Results We demonstrate the heterogeneity of current prediction sets, and take advantage of this heterogeneity to construct a two-level classifier (\\'Profisi Ensemble\\') using predictions from 7 programs, along with 2 other data sources. Support vector machines using \\'full\\' and \\'reduced\\' data sets are combined in an either\\/or approach. We achieve a 14% increase in performance over the current state-of-the-art, as benchmarked by a third-party tool. Conclusions Supervised learning methods are a useful way to combine predictions from diverse sources.

  7. Predicting Flavin and Nicotinamide Adenine Dinucleotide-Binding Sites in Proteins Using the Fragment Transformation Method

    Directory of Open Access Journals (Sweden)

    Chih-Hao Lu

    2015-01-01

    Full Text Available We developed a computational method to identify NAD- and FAD-binding sites in proteins. First, we extracted from the Protein Data Bank structures of proteins that bind to at least one of these ligands. NAD-/FAD-binding residue templates were then constructed by identifying binding residues through the ligand-binding database BioLiP. The fragment transformation method was used to identify structures within query proteins that resembled the ligand-binding templates. By comparing residue types and their relative spatial positions, potential binding sites were identified and a ligand-binding potential for each residue was calculated. Setting the false positive rate at 5%, our method predicted NAD- and FAD-binding sites at true positive rates of 67.1% and 68.4%, respectively. Our method provides excellent results for identifying FAD- and NAD-binding sites in proteins, and the most important is that the requirement of conservation of residue types and local structures in the FAD- and NAD-binding sites can be verified.

  8. A novel method for improved accuracy of transcription factor binding site prediction

    KAUST Repository

    Khamis, Abdullah M.; Motwalli, Olaa Amin; Oliva, Romina; Jankovic, Boris R.; Medvedeva, Yulia; Ashoor, Haitham; Essack, Magbubah; Gao, Xin; Bajic, Vladimir B.

    2018-01-01

    Identifying transcription factor (TF) binding sites (TFBSs) is important in the computational inference of gene regulation. Widely used computational methods of TFBS prediction based on position weight matrices (PWMs) usually have high false positive rates. Moreover, computational studies of transcription regulation in eukaryotes frequently require numerous PWM models of TFBSs due to a large number of TFs involved. To overcome these problems we developed DRAF, a novel method for TFBS prediction that requires only 14 prediction models for 232 human TFs, while at the same time significantly improves prediction accuracy. DRAF models use more features than PWM models, as they combine information from TFBS sequences and physicochemical properties of TF DNA-binding domains into machine learning models. Evaluation of DRAF on 98 human ChIP-seq datasets shows on average 1.54-, 1.96- and 5.19-fold reduction of false positives at the same sensitivities compared to models from HOCOMOCO, TRANSFAC and DeepBind, respectively. This observation suggests that one can efficiently replace the PWM models for TFBS prediction by a small number of DRAF models that significantly improve prediction accuracy. The DRAF method is implemented in a web tool and in a stand-alone software freely available at http://cbrc.kaust.edu.sa/DRAF.

  9. A novel method for improved accuracy of transcription factor binding site prediction

    KAUST Repository

    Khamis, Abdullah M.

    2018-03-20

    Identifying transcription factor (TF) binding sites (TFBSs) is important in the computational inference of gene regulation. Widely used computational methods of TFBS prediction based on position weight matrices (PWMs) usually have high false positive rates. Moreover, computational studies of transcription regulation in eukaryotes frequently require numerous PWM models of TFBSs due to a large number of TFs involved. To overcome these problems we developed DRAF, a novel method for TFBS prediction that requires only 14 prediction models for 232 human TFs, while at the same time significantly improves prediction accuracy. DRAF models use more features than PWM models, as they combine information from TFBS sequences and physicochemical properties of TF DNA-binding domains into machine learning models. Evaluation of DRAF on 98 human ChIP-seq datasets shows on average 1.54-, 1.96- and 5.19-fold reduction of false positives at the same sensitivities compared to models from HOCOMOCO, TRANSFAC and DeepBind, respectively. This observation suggests that one can efficiently replace the PWM models for TFBS prediction by a small number of DRAF models that significantly improve prediction accuracy. The DRAF method is implemented in a web tool and in a stand-alone software freely available at http://cbrc.kaust.edu.sa/DRAF.

  10. Feature selection for splice site prediction: A new method using EDA-based feature ranking

    Directory of Open Access Journals (Sweden)

    Rouzé Pierre

    2004-05-01

    Full Text Available Abstract Background The identification of relevant biological features in large and complex datasets is an important step towards gaining insight in the processes underlying the data. Other advantages of feature selection include the ability of the classification system to attain good or even better solutions using a restricted subset of features, and a faster classification. Thus, robust methods for fast feature selection are of key importance in extracting knowledge from complex biological data. Results In this paper we present a novel method for feature subset selection applied to splice site prediction, based on estimation of distribution algorithms, a more general framework of genetic algorithms. From the estimated distribution of the algorithm, a feature ranking is derived. Afterwards this ranking is used to iteratively discard features. We apply this technique to the problem of splice site prediction, and show how it can be used to gain insight into the underlying biological process of splicing. Conclusion We show that this technique proves to be more robust than the traditional use of estimation of distribution algorithms for feature selection: instead of returning a single best subset of features (as they normally do this method provides a dynamical view of the feature selection process, like the traditional sequential wrapper methods. However, the method is faster than the traditional techniques, and scales better to datasets described by a large number of features.

  11. A neural network method for identification of prokaryotic and eukaryotic signal peptides and prediction of their cleavage sites

    DEFF Research Database (Denmark)

    Nielsen, Henrik; Engelbrecht, Jacob; Brunak, Søren

    1997-01-01

    We have developed a new method for the identication of signal peptides and their cleavage sites based on neural networks trained on separate sets of prokaryotic and eukaryotic sequences. The method performs signicantly better than previous prediction schemes, and can easily be applied to genome...

  12. Predicting ecological flow regime at ungaged sites: A comparison of methods

    Science.gov (United States)

    Murphy, Jennifer C.; Knight, Rodney R.; Wolfe, William J.; Gain, W. Scott

    2012-01-01

    Nineteen ecologically relevant streamflow characteristics were estimated using published rainfall–runoff and regional regression models for six sites with observed daily streamflow records in Kentucky. The regional regression model produced median estimates closer to the observed median for all but two characteristics. The variability of predictions from both models was generally less than the observed variability. The variability of the predictions from the rainfall–runoff model was greater than that from the regional regression model for all but three characteristics. Eight characteristics predicted by the rainfall–runoff model display positive or negative bias across all six sites; biases are not as pronounced for the regional regression model. Results suggest that a rainfall–runoff model calibrated on a single characteristic is less likely to perform well as a predictor of a range of other characteristics (flow regime) when compared with a regional regression model calibrated individually on multiple characteristics used to represent the flow regime. Poor model performance may misrepresent hydrologic conditions, potentially distorting the perceived risk of ecological degradation. Without prior selection of streamflow characteristics, targeted calibration, and error quantification, the widespread application of general hydrologic models to ecological flow studies is problematic. Published 2012. This article is a U.S. Government work and is in the public domain in the USA.

  13. Prediction of protein binding sites using physical and chemical descriptors and the support vector machine regression method

    International Nuclear Information System (INIS)

    Sun Zhong-Hua; Jiang Fan

    2010-01-01

    In this paper a new continuous variable called core-ratio is defined to describe the probability for a residue to be in a binding site, thereby replacing the previous binary description of the interface residue using 0 and 1. So we can use the support vector machine regression method to fit the core-ratio value and predict the protein binding sites. We also design a new group of physical and chemical descriptors to characterize the binding sites. The new descriptors are more effective, with an averaging procedure used. Our test shows that much better prediction results can be obtained by the support vector regression (SVR) method than by the support vector classification method. (rapid communication)

  14. A Genetic Programming Method for the Identification of Signal Peptides and Prediction of Their Cleavage Sites

    Directory of Open Access Journals (Sweden)

    David Lennartsson

    2004-01-01

    Full Text Available A novel approach to signal peptide identification is presented. We use an evolutionary algorithm for automatic evolution of classification programs, so-called programmatic motifs. The variant of evolutionary algorithm used is called genetic programming where a population of solution candidates in the form of full computer programs is evolved, based on training examples consisting of signal peptide sequences. The method is compared with a previous work using artificial neural network (ANN approaches. Some advantages compared to ANNs are noted. The programmatic motif can perform computational tasks beyond that of feed-forward neural networks and has also other advantages such as readability. The best motif evolved was analyzed and shown to detect the h-region of the signal peptide. A powerful parallel computer cluster was used for the experiment.

  15. pDHS-SVM: A prediction method for plant DNase I hypersensitive sites based on support vector machine.

    Science.gov (United States)

    Zhang, Shanxin; Zhou, Zhiping; Chen, Xinmeng; Hu, Yong; Yang, Lindong

    2017-08-07

    DNase I hypersensitive sites (DHSs) are accessible chromatin regions hypersensitive to cleavages by DNase I endonucleases. DHSs are indicative of cis-regulatory DNA elements (CREs), all of which play important roles in global gene expression regulation. It is helpful for discovering CREs by recognition of DHSs in genome. To accelerate the investigation, it is an important complement to develop cost-effective computational methods to identify DHSs. However, there is a lack of tools used for identifying DHSs in plant genome. Here we presented pDHS-SVM, a computational predictor to identify plant DHSs. To integrate the global sequence-order information and local DNA properties, reverse complement kmer and dinucleotide-based auto covariance of DNA sequences were applied to construct the feature space. In this work, fifteen physical-chemical properties of dinucleotides were used and Support Vector Machine (SVM) was employed. To further improve the performance of the predictor and extract an optimized subset of nucleotide physical-chemical properties positive for the DHSs, a heuristic nucleotide physical-chemical property selection algorithm was introduced. With the optimized subset of properties, experimental results of Arabidopsis thaliana and rice (Oryza sativa) showed that pDHS-SVM could achieve accuracies up to 87.00%, and 85.79%, respectively. The results indicated the effectiveness of proposed method for predicting DHSs. Furthermore, pDHS-SVM could provide a helpful complement for predicting CREs in plant genome. Our implementation of the novel proposed method pDHS-SVM is freely available as source code, at https://github.com/shanxinzhang/pDHS-SVM. Copyright © 2017 Elsevier Ltd. All rights reserved.

  16. A consistency-based feature selection method allied with linear SVMs for HIV-1 protease cleavage site prediction.

    Directory of Open Access Journals (Sweden)

    Orkun Oztürk

    Full Text Available BACKGROUND: Predicting type-1 Human Immunodeficiency Virus (HIV-1 protease cleavage site in protein molecules and determining its specificity is an important task which has attracted considerable attention in the research community. Achievements in this area are expected to result in effective drug design (especially for HIV-1 protease inhibitors against this life-threatening virus. However, some drawbacks (like the shortage of the available training data and the high dimensionality of the feature space turn this task into a difficult classification problem. Thus, various machine learning techniques, and specifically several classification methods have been proposed in order to increase the accuracy of the classification model. In addition, for several classification problems, which are characterized by having few samples and many features, selecting the most relevant features is a major factor for increasing classification accuracy. RESULTS: We propose for HIV-1 data a consistency-based feature selection approach in conjunction with recursive feature elimination of support vector machines (SVMs. We used various classifiers for evaluating the results obtained from the feature selection process. We further demonstrated the effectiveness of our proposed method by comparing it with a state-of-the-art feature selection method applied on HIV-1 data, and we evaluated the reported results based on attributes which have been selected from different combinations. CONCLUSION: Applying feature selection on training data before realizing the classification task seems to be a reasonable data-mining process when working with types of data similar to HIV-1. On HIV-1 data, some feature selection or extraction operations in conjunction with different classifiers have been tested and noteworthy outcomes have been reported. These facts motivate for the work presented in this paper. SOFTWARE AVAILABILITY: The software is available at http

  17. SitesIdentify: a protein functional site prediction tool

    Directory of Open Access Journals (Sweden)

    Doig Andrew J

    2009-11-01

    Full Text Available Abstract Background The rate of protein structures being deposited in the Protein Data Bank surpasses the capacity to experimentally characterise them and therefore computational methods to analyse these structures have become increasingly important. Identifying the region of the protein most likely to be involved in function is useful in order to gain information about its potential role. There are many available approaches to predict functional site, but many are not made available via a publicly-accessible application. Results Here we present a functional site prediction tool (SitesIdentify, based on combining sequence conservation information with geometry-based cleft identification, that is freely available via a web-server. We have shown that SitesIdentify compares favourably to other functional site prediction tools in a comparison of seven methods on a non-redundant set of 237 enzymes with annotated active sites. Conclusion SitesIdentify is able to produce comparable accuracy in predicting functional sites to its closest available counterpart, but in addition achieves improved accuracy for proteins with few characterised homologues. SitesIdentify is available via a webserver at http://www.manchester.ac.uk/bioinformatics/sitesidentify/

  18. Climate Prediction Center - Site Index

    Science.gov (United States)

    Weather Service NWS logo - Click to go to the NWS home page Climate Prediction Center Home Site Map News Means Bulletins Annual Winter Stratospheric Ozone Climate Diagnostics Bulletin (Most Recent) Climate (Hazards Outlook) Climate Assessment: Dec. 1999-Feb. 2000 (Seasonal) Climate Assessment: Mar-May 2000

  19. ChloroP, a neural network-based method for predicting chloroplast transitpeptides and their cleavage sites

    DEFF Research Database (Denmark)

    Emanuelsson, O.; Nielsen, Henrik; von Heijne, Gunnar

    1999-01-01

    the cleavage sites given in SWISS-PROT. An analysis of 715 Arabidopsis thaliana sequences from SWISS-PROT suggests that the ChloroP method should be useful for the identification of putative transit peptides in genome-wide sequence data. The ChloroP predictor is available as a web-server at http......We present a neural network based method (ChloroP) for identifying chloroplast transit peptides and their cleavage sites. Using cross-validation, 88% of the sequences in our homology reduced training set were correctly classified as transit peptides or nontransit peptides. This performance level...

  20. Text mining improves prediction of protein functional sites.

    Directory of Open Access Journals (Sweden)

    Karin M Verspoor

    Full Text Available We present an approach that integrates protein structure analysis and text mining for protein functional site prediction, called LEAP-FS (Literature Enhanced Automated Prediction of Functional Sites. The structure analysis was carried out using Dynamics Perturbation Analysis (DPA, which predicts functional sites at control points where interactions greatly perturb protein vibrations. The text mining extracts mentions of residues in the literature, and predicts that residues mentioned are functionally important. We assessed the significance of each of these methods by analyzing their performance in finding known functional sites (specifically, small-molecule binding sites and catalytic sites in about 100,000 publicly available protein structures. The DPA predictions recapitulated many of the functional site annotations and preferentially recovered binding sites annotated as biologically relevant vs. those annotated as potentially spurious. The text-based predictions were also substantially supported by the functional site annotations: compared to other residues, residues mentioned in text were roughly six times more likely to be found in a functional site. The overlap of predictions with annotations improved when the text-based and structure-based methods agreed. Our analysis also yielded new high-quality predictions of many functional site residues that were not catalogued in the curated data sources we inspected. We conclude that both DPA and text mining independently provide valuable high-throughput protein functional site predictions, and that integrating the two methods using LEAP-FS further improves the quality of these predictions.

  1. Text Mining Improves Prediction of Protein Functional Sites

    Science.gov (United States)

    Cohn, Judith D.; Ravikumar, Komandur E.

    2012-01-01

    We present an approach that integrates protein structure analysis and text mining for protein functional site prediction, called LEAP-FS (Literature Enhanced Automated Prediction of Functional Sites). The structure analysis was carried out using Dynamics Perturbation Analysis (DPA), which predicts functional sites at control points where interactions greatly perturb protein vibrations. The text mining extracts mentions of residues in the literature, and predicts that residues mentioned are functionally important. We assessed the significance of each of these methods by analyzing their performance in finding known functional sites (specifically, small-molecule binding sites and catalytic sites) in about 100,000 publicly available protein structures. The DPA predictions recapitulated many of the functional site annotations and preferentially recovered binding sites annotated as biologically relevant vs. those annotated as potentially spurious. The text-based predictions were also substantially supported by the functional site annotations: compared to other residues, residues mentioned in text were roughly six times more likely to be found in a functional site. The overlap of predictions with annotations improved when the text-based and structure-based methods agreed. Our analysis also yielded new high-quality predictions of many functional site residues that were not catalogued in the curated data sources we inspected. We conclude that both DPA and text mining independently provide valuable high-throughput protein functional site predictions, and that integrating the two methods using LEAP-FS further improves the quality of these predictions. PMID:22393388

  2. Prediction of site-specific interactions in antibody-antigen complexes: the proABC method and server.

    KAUST Repository

    Olimpieri, Pier Paolo

    2013-06-26

    MOTIVATION: Antibodies or immunoglobulins are proteins of paramount importance in the immune system. They are extremely relevant as diagnostic, biotechnological and therapeutic tools. Their modular structure makes it easy to re-engineer them for specific purposes. Short of undergoing a trial and error process, these experiments, as well as others, need to rely on an understanding of the specific determinants of the antibody binding mode. RESULTS: In this article, we present a method to identify, on the basis of the antibody sequence alone, which residues of an antibody directly interact with its cognate antigen. The method, based on the random forest automatic learning techniques, reaches a recall and specificity as high as 80% and is implemented as a free and easy-to-use server, named prediction of Antibody Contacts. We believe that it can be of great help in re-design experiments as well as a guide for molecular docking experiments. The results that we obtained also allowed us to dissect which features of the antibody sequence contribute most to the involvement of specific residues in binding to the antigen. AVAILABILITY: http://www.biocomputing.it/proABC. CONTACT: anna.tramontano@uniroma1.it or paolo.marcatili@gmail.com SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.

  3. A method for predicting individual residue contributions to enzyme specificity and binding-site energies, and its application to MTH1.

    Science.gov (United States)

    Stewart, James J P

    2016-11-01

    A new method for predicting the energy contributions to substrate binding and to specificity has been developed. Conventional global optimization methods do not permit the subtle effects responsible for these properties to be modeled with sufficient precision to allow confidence to be placed in the results, but by making simple alterations to the model, the precisions of the various energies involved can be improved from about ±2 kcal mol -1 to ±0.1 kcal mol -1 . This technique was applied to the oxidized nucleotide pyrophosphohydrolase enzyme MTH1. MTH1 is unusual in that the binding and reaction sites are well separated-an advantage from a computational chemistry perspective, as it allows the energetics involved in docking to be modeled without the need to consider any issues relating to reaction mechanisms. In this study, two types of energy terms were investigated: the noncovalent interactions between the binding site and the substrate, and those responsible for discriminating between the oxidized nucleotide 8-oxo-dGTP and the normal dGTP. Both of these were investigated using the semiempirical method PM7 in the program MOPAC. The contributions of the individual residues to both the binding energy and the specificity of MTH1 were calculated by simulating the effect of mutations. Where comparisons were possible, all calculated results were in agreement with experimental observations. This technique provides fresh insight into the binding mechanism that enzymes use for discriminating between possible substrates.

  4. Positive-Unlabeled Learning for Pupylation Sites Prediction

    Directory of Open Access Journals (Sweden)

    Ming Jiang

    2016-01-01

    Full Text Available Pupylation plays a key role in regulating various protein functions as a crucial posttranslational modification of prokaryotes. In order to understand the molecular mechanism of pupylation, it is important to identify pupylation substrates and sites accurately. Several computational methods have been developed to identify pupylation sites because the traditional experimental methods are time-consuming and labor-sensitive. With the existing computational methods, the experimentally annotated pupylation sites are used as the positive training set and the remaining nonannotated lysine residues as the negative training set to build classifiers to predict new pupylation sites from the unknown proteins. However, the remaining nonannotated lysine residues may contain pupylation sites which have not been experimentally validated yet. Unlike previous methods, in this study, the experimentally annotated pupylation sites were used as the positive training set whereas the remaining nonannotated lysine residues were used as the unlabeled training set. A novel method named PUL-PUP was proposed to predict pupylation sites by using positive-unlabeled learning technique. Our experimental results indicated that PUL-PUP outperforms the other methods significantly for the prediction of pupylation sites. As an application, PUL-PUP was also used to predict the most likely pupylation sites in nonannotated lysine sites.

  5. Prediction of glycosylation sites using random forests

    Directory of Open Access Journals (Sweden)

    Hirst Jonathan D

    2008-11-01

    Full Text Available Abstract Background Post translational modifications (PTMs occur in the vast majority of proteins and are essential for function. Prediction of the sequence location of PTMs enhances the functional characterisation of proteins. Glycosylation is one type of PTM, and is implicated in protein folding, transport and function. Results We use the random forest algorithm and pairwise patterns to predict glycosylation sites. We identify pairwise patterns surrounding glycosylation sites and use an odds ratio to weight their propensity of association with modified residues. Our prediction program, GPP (glycosylation prediction program, predicts glycosylation sites with an accuracy of 90.8% for Ser sites, 92.0% for Thr sites and 92.8% for Asn sites. This is significantly better than current glycosylation predictors. We use the trepan algorithm to extract a set of comprehensible rules from GPP, which provide biological insight into all three major glycosylation types. Conclusion We have created an accurate predictor of glycosylation sites and used this to extract comprehensible rules about the glycosylation process. GPP is available online at http://comp.chem.nottingham.ac.uk/glyco/.

  6. A grammar inference approach for predicting kinase specific phosphorylation sites.

    Science.gov (United States)

    Datta, Sutapa; Mukhopadhyay, Subhasis

    2015-01-01

    Kinase mediated phosphorylation site detection is the key mechanism of post translational mechanism that plays an important role in regulating various cellular processes and phenotypes. Many diseases, like cancer are related with the signaling defects which are associated with protein phosphorylation. Characterizing the protein kinases and their substrates enhances our ability to understand the mechanism of protein phosphorylation and extends our knowledge of signaling network; thereby helping us to treat such diseases. Experimental methods for predicting phosphorylation sites are labour intensive and expensive. Also, manifold increase of protein sequences in the databanks over the years necessitates the improvement of high speed and accurate computational methods for predicting phosphorylation sites in protein sequences. Till date, a number of computational methods have been proposed by various researchers in predicting phosphorylation sites, but there remains much scope of improvement. In this communication, we present a simple and novel method based on Grammatical Inference (GI) approach to automate the prediction of kinase specific phosphorylation sites. In this regard, we have used a popular GI algorithm Alergia to infer Deterministic Stochastic Finite State Automata (DSFA) which equally represents the regular grammar corresponding to the phosphorylation sites. Extensive experiments on several datasets generated by us reveal that, our inferred grammar successfully predicts phosphorylation sites in a kinase specific manner. It performs significantly better when compared with the other existing phosphorylation site prediction methods. We have also compared our inferred DSFA with two other GI inference algorithms. The DSFA generated by our method performs superior which indicates that our method is robust and has a potential for predicting the phosphorylation sites in a kinase specific manner.

  7. A Grammar Inference Approach for Predicting Kinase Specific Phosphorylation Sites

    Science.gov (United States)

    Datta, Sutapa; Mukhopadhyay, Subhasis

    2015-01-01

    Kinase mediated phosphorylation site detection is the key mechanism of post translational mechanism that plays an important role in regulating various cellular processes and phenotypes. Many diseases, like cancer are related with the signaling defects which are associated with protein phosphorylation. Characterizing the protein kinases and their substrates enhances our ability to understand the mechanism of protein phosphorylation and extends our knowledge of signaling network; thereby helping us to treat such diseases. Experimental methods for predicting phosphorylation sites are labour intensive and expensive. Also, manifold increase of protein sequences in the databanks over the years necessitates the improvement of high speed and accurate computational methods for predicting phosphorylation sites in protein sequences. Till date, a number of computational methods have been proposed by various researchers in predicting phosphorylation sites, but there remains much scope of improvement. In this communication, we present a simple and novel method based on Grammatical Inference (GI) approach to automate the prediction of kinase specific phosphorylation sites. In this regard, we have used a popular GI algorithm Alergia to infer Deterministic Stochastic Finite State Automata (DSFA) which equally represents the regular grammar corresponding to the phosphorylation sites. Extensive experiments on several datasets generated by us reveal that, our inferred grammar successfully predicts phosphorylation sites in a kinase specific manner. It performs significantly better when compared with the other existing phosphorylation site prediction methods. We have also compared our inferred DSFA with two other GI inference algorithms. The DSFA generated by our method performs superior which indicates that our method is robust and has a potential for predicting the phosphorylation sites in a kinase specific manner. PMID:25886273

  8. Epitope prediction methods

    DEFF Research Database (Denmark)

    Karosiene, Edita

    Analysis. The chapter provides detailed explanations on how to use different methods for T cell epitope discovery research, explaining how input should be given as well as how to interpret the output. In the last chapter, I present the results of a bioinformatics analysis of epitopes from the yellow fever...... peptide-MHC interactions. Furthermore, using yellow fever virus epitopes, we demonstrated the power of the %Rank score when compared with the binding affinity score of MHC prediction methods, suggesting that this score should be considered to be used for selecting potential T cell epitopes. In summary...... immune responses. Therefore, it is of great importance to be able to identify peptides that bind to MHC molecules, in order to understand the nature of immune responses and discover T cell epitopes useful for designing new vaccines and immunotherapies. MHC molecules in humans, referred to as human...

  9. Motor degradation prediction methods

    Energy Technology Data Exchange (ETDEWEB)

    Arnold, J.R.; Kelly, J.F.; Delzingaro, M.J.

    1996-12-01

    Motor Operated Valve (MOV) squirrel cage AC motor rotors are susceptible to degradation under certain conditions. Premature failure can result due to high humidity/temperature environments, high running load conditions, extended periods at locked rotor conditions (i.e. > 15 seconds) or exceeding the motor`s duty cycle by frequent starts or multiple valve stroking. Exposure to high heat and moisture due to packing leaks, pressure seal ring leakage or other causes can significantly accelerate the degradation. ComEd and Liberty Technologies have worked together to provide and validate a non-intrusive method using motor power diagnostics to evaluate MOV rotor condition and predict failure. These techniques have provided a quick, low radiation dose method to evaluate inaccessible motors, identify degradation and allow scheduled replacement of motors prior to catastrophic failures.

  10. Motor degradation prediction methods

    International Nuclear Information System (INIS)

    Arnold, J.R.; Kelly, J.F.; Delzingaro, M.J.

    1996-01-01

    Motor Operated Valve (MOV) squirrel cage AC motor rotors are susceptible to degradation under certain conditions. Premature failure can result due to high humidity/temperature environments, high running load conditions, extended periods at locked rotor conditions (i.e. > 15 seconds) or exceeding the motor's duty cycle by frequent starts or multiple valve stroking. Exposure to high heat and moisture due to packing leaks, pressure seal ring leakage or other causes can significantly accelerate the degradation. ComEd and Liberty Technologies have worked together to provide and validate a non-intrusive method using motor power diagnostics to evaluate MOV rotor condition and predict failure. These techniques have provided a quick, low radiation dose method to evaluate inaccessible motors, identify degradation and allow scheduled replacement of motors prior to catastrophic failures

  11. Variable context Markov chains for HIV protease cleavage site prediction.

    Science.gov (United States)

    Oğul, Hasan

    2009-06-01

    Deciphering the knowledge of HIV protease specificity and developing computational tools for detecting its cleavage sites in protein polypeptide chain are very desirable for designing efficient and specific chemical inhibitors to prevent acquired immunodeficiency syndrome. In this study, we developed a generative model based on a generalization of variable order Markov chains (VOMC) for peptide sequences and adapted the model for prediction of their cleavability by certain proteases. The new method, called variable context Markov chains (VCMC), attempts to identify the context equivalence based on the evolutionary similarities between individual amino acids. It was applied for HIV-1 protease cleavage site prediction problem and shown to outperform existing methods in terms of prediction accuracy on a common dataset. In general, the method is a promising tool for prediction of cleavage sites of all proteases and encouraged to be used for any kind of peptide classification problem as well.

  12. A systems biology approach to transcription factor binding site prediction.

    Directory of Open Access Journals (Sweden)

    Xiang Zhou

    2010-03-01

    Full Text Available The elucidation of mammalian transcriptional regulatory networks holds great promise for both basic and translational research and remains one the greatest challenges to systems biology. Recent reverse engineering methods deduce regulatory interactions from large-scale mRNA expression profiles and cross-species conserved regulatory regions in DNA. Technical challenges faced by these methods include distinguishing between direct and indirect interactions, associating transcription regulators with predicted transcription factor binding sites (TFBSs, identifying non-linearly conserved binding sites across species, and providing realistic accuracy estimates.We address these challenges by closely integrating proven methods for regulatory network reverse engineering from mRNA expression data, linearly and non-linearly conserved regulatory region discovery, and TFBS evaluation and discovery. Using an extensive test set of high-likelihood interactions, which we collected in order to provide realistic prediction-accuracy estimates, we show that a careful integration of these methods leads to significant improvements in prediction accuracy. To verify our methods, we biochemically validated TFBS predictions made for both transcription factors (TFs and co-factors; we validated binding site predictions made using a known E2F1 DNA-binding motif on E2F1 predicted promoter targets, known E2F1 and JUND motifs on JUND predicted promoter targets, and a de novo discovered motif for BCL6 on BCL6 predicted promoter targets. Finally, to demonstrate accuracy of prediction using an external dataset, we showed that sites matching predicted motifs for ZNF263 are significantly enriched in recent ZNF263 ChIP-seq data.Using an integrative framework, we were able to address technical challenges faced by state of the art network reverse engineering methods, leading to significant improvement in direct-interaction detection and TFBS-discovery accuracy. We estimated the accuracy

  13. Empirical Flutter Prediction Method.

    Science.gov (United States)

    1988-03-05

    been used in this way to discover species or subspecies of animals, and to discover different types of voter or comsumer requiring different persuasions...respect to behavior or performance or response variables. Once this were done, corresponding clusters might be sought among descriptive or predictive or...jump in a response. The first sort of usage does not apply to the flutter prediction problem. Here the types of behavior are the different kinds of

  14. Human Splice-Site Prediction with Deep Neural Networks.

    Science.gov (United States)

    Naito, Tatsuhiko

    2018-04-18

    Accurate splice-site prediction is essential to delineate gene structures from sequence data. Several computational techniques have been applied to create a system to predict canonical splice sites. For classification tasks, deep neural networks (DNNs) have achieved record-breaking results and often outperformed other supervised learning techniques. In this study, a new method of splice-site prediction using DNNs was proposed. The proposed system receives an input sequence data and returns an answer as to whether it is splice site. The length of input is 140 nucleotides, with the consensus sequence (i.e., "GT" and "AG" for the donor and acceptor sites, respectively) in the middle. Each input sequence model is applied to the pretrained DNN model that determines the probability that an input is a splice site. The model consists of convolutional layers and bidirectional long short-term memory network layers. The pretraining and validation were conducted using the data set tested in previously reported methods. The performance evaluation results showed that the proposed method can outperform the previous methods. In addition, the pattern learned by the DNNs was visualized as position frequency matrices (PFMs). Some of PFMs were very similar to the consensus sequence. The trained DNN model and the brief source code for the prediction system are uploaded. Further improvement will be achieved following the further development of DNNs.

  15. Finding protein sites using machine learning methods

    Directory of Open Access Journals (Sweden)

    Jaime Leonardo Bobadilla Molina

    2003-07-01

    Full Text Available The increasing amount of protein three-dimensional (3D structures determined by x-ray and NMR technologies as well as structures predicted by computational methods results in the need for automated methods to provide inital annotations. We have developed a new method for recognizing sites in three-dimensional protein structures. Our method is based on a previosly reported algorithm for creating descriptions of protein microenviroments using physical and chemical properties at multiple levels of detail. The recognition method takes three inputs: 1. A set of control nonsites that share some structural or functional role. 2. A set of control nonsites that lack this role. 3. A single query site. A support vector machine classifier is built using feature vectors where each component represents a property in a given volume. Validation against an independent test set shows that this recognition approach has high sensitivity and specificity. We also describe the results of scanning four calcium binding proteins (with the calcium removed using a three dimensional grid of probe points at 1.25 angstrom spacing. The system finds the sites in the proteins giving points at or near the blinding sites. Our results show that property based descriptions along with support vector machines can be used for recognizing protein sites in unannotated structures.

  16. Prediction method abstracts

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1994-12-31

    This conference was held December 4--8, 1994 in Asilomar, California. The purpose of this meeting was to provide a forum for exchange of state-of-the-art information concerning the prediction of protein structure. Attention if focused on the following: comparative modeling; sequence to fold assignment; and ab initio folding.

  17. Earthquake prediction by Kina Method

    International Nuclear Information System (INIS)

    Kianoosh, H.; Keypour, H.; Naderzadeh, A.; Motlagh, H.F.

    2005-01-01

    Earthquake prediction has been one of the earliest desires of the man. Scientists have worked hard to predict earthquakes for a long time. The results of these efforts can generally be divided into two methods of prediction: 1) Statistical Method, and 2) Empirical Method. In the first method, earthquakes are predicted using statistics and probabilities, while the second method utilizes variety of precursors for earthquake prediction. The latter method is time consuming and more costly. However, the result of neither method has fully satisfied the man up to now. In this paper a new method entitled 'Kiana Method' is introduced for earthquake prediction. This method offers more accurate results yet lower cost comparing to other conventional methods. In Kiana method the electrical and magnetic precursors are measured in an area. Then, the time and the magnitude of an earthquake in the future is calculated using electrical, and in particular, electrical capacitors formulas. In this method, by daily measurement of electrical resistance in an area we make clear that the area is capable of earthquake occurrence in the future or not. If the result shows a positive sign, then the occurrence time and the magnitude can be estimated by the measured quantities. This paper explains the procedure and details of this prediction method. (authors)

  18. Predictive Methods of Pople

    Indian Academy of Sciences (India)

    Chemistry for their pioneering contri butions to the development of computational methods in quantum chemistry and density functional theory .... program of Pop Ie for ab-initio electronic structure calculation of molecules. This ab-initio MO ...

  19. XenoSite: accurately predicting CYP-mediated sites of metabolism with neural networks.

    Science.gov (United States)

    Zaretzki, Jed; Matlock, Matthew; Swamidass, S Joshua

    2013-12-23

    Understanding how xenobiotic molecules are metabolized is important because it influences the safety, efficacy, and dose of medicines and how they can be modified to improve these properties. The cytochrome P450s (CYPs) are proteins responsible for metabolizing 90% of drugs on the market, and many computational methods can predict which atomic sites of a molecule--sites of metabolism (SOMs)--are modified during CYP-mediated metabolism. This study improves on prior methods of predicting CYP-mediated SOMs by using new descriptors and machine learning based on neural networks. The new method, XenoSite, is faster to train and more accurate by as much as 4% or 5% for some isozymes. Furthermore, some "incorrect" predictions made by XenoSite were subsequently validated as correct predictions by revaluation of the source literature. Moreover, XenoSite output is interpretable as a probability, which reflects both the confidence of the model that a particular atom is metabolized and the statistical likelihood that its prediction for that atom is correct.

  20. Pripper: prediction of caspase cleavage sites from whole proteomes

    Directory of Open Access Journals (Sweden)

    Salmi Jussi

    2010-06-01

    Full Text Available Abstract Background Caspases are a family of proteases that have central functions in programmed cell death (apoptosis and inflammation. Caspases mediate their effects through aspartate-specific cleavage of their target proteins, and at present almost 400 caspase substrates are known. There are several methods developed to predict caspase cleavage sites from individual proteins, but currently none of them can be used to predict caspase cleavage sites from multiple proteins or entire proteomes, or to use several classifiers in combination. The possibility to create a database from predicted caspase cleavage products for the whole genome could significantly aid in identifying novel caspase targets from tandem mass spectrometry based proteomic experiments. Results Three different pattern recognition classifiers were developed for predicting caspase cleavage sites from protein sequences. Evaluation of the classifiers with quality measures indicated that all of the three classifiers performed well in predicting caspase cleavage sites, and when combining different classifiers the accuracy increased further. A new tool, Pripper, was developed to utilize the classifiers and predict the caspase cut sites from an arbitrary number of input sequences. A database was constructed with the developed tool, and it was used to identify caspase target proteins from tandem mass spectrometry data from two different proteomic experiments. Both known caspase cleavage products as well as novel cleavage products were identified using the database demonstrating the usefulness of the tool. Pripper is not restricted to predicting only caspase cut sites, but it gives the possibility to scan protein sequences for any given motif(s and predict cut sites once a suitable cut site prediction model for any other protease has been developed. Pripper is freely available and can be downloaded from http://users.utu.fi/mijopi/Pripper. Conclusions We have developed Pripper, a tool for

  1. Predicting protein amidation sites by orchestrating amino acid sequence features

    Science.gov (United States)

    Zhao, Shuqiu; Yu, Hua; Gong, Xiujun

    2017-08-01

    Amidation is the fourth major category of post-translational modifications, which plays an important role in physiological and pathological processes. Identifying amidation sites can help us understanding the amidation and recognizing the original reason of many kinds of diseases. But the traditional experimental methods for predicting amidation sites are often time-consuming and expensive. In this study, we propose a computational method for predicting amidation sites by orchestrating amino acid sequence features. Three kinds of feature extraction methods are used to build a feature vector enabling to capture not only the physicochemical properties but also position related information of the amino acids. An extremely randomized trees algorithm is applied to choose the optimal features to remove redundancy and dependence among components of the feature vector by a supervised fashion. Finally the support vector machine classifier is used to label the amidation sites. When tested on an independent data set, it shows that the proposed method performs better than all the previous ones with the prediction accuracy of 0.962 at the Matthew's correlation coefficient of 0.89 and area under curve of 0.964.

  2. Transcription factor binding sites prediction based on modified nucleosomes.

    Directory of Open Access Journals (Sweden)

    Mohammad Talebzadeh

    Full Text Available In computational methods, position weight matrices (PWMs are commonly applied for transcription factor binding site (TFBS prediction. Although these matrices are more accurate than simple consensus sequences to predict actual binding sites, they usually produce a large number of false positive (FP predictions and so are impoverished sources of information. Several studies have employed additional sources of information such as sequence conservation or the vicinity to transcription start sites to distinguish true binding regions from random ones. Recently, the spatial distribution of modified nucleosomes has been shown to be associated with different promoter architectures. These aligned patterns can facilitate DNA accessibility for transcription factors. We hypothesize that using data from these aligned and periodic patterns can improve the performance of binding region prediction. In this study, we propose two effective features, "modified nucleosomes neighboring" and "modified nucleosomes occupancy", to decrease FP in binding site discovery. Based on these features, we designed a logistic regression classifier which estimates the probability of a region as a TFBS. Our model learned each feature based on Sp1 binding sites on Chromosome 1 and was tested on the other chromosomes in human CD4+T cells. In this work, we investigated 21 histone modifications and found that only 8 out of 21 marks are strongly correlated with transcription factor binding regions. To prove that these features are not specific to Sp1, we combined the logistic regression classifier with the PWM, and created a new model to search TFBSs on the genome. We tested the model using transcription factors MAZ, PU.1 and ELF1 and compared the results to those using only the PWM. The results show that our model can predict Transcription factor binding regions more successfully. The relative simplicity of the model and capability of integrating other features make it a superior method

  3. Rainfall prediction with backpropagation method

    Science.gov (United States)

    Wahyuni, E. G.; Fauzan, L. M. F.; Abriyani, F.; Muchlis, N. F.; Ulfa, M.

    2018-03-01

    Rainfall is an important factor in many fields, such as aviation and agriculture. Although it has been assisted by technology but the accuracy can not reach 100% and there is still the possibility of error. Though current rainfall prediction information is needed in various fields, such as agriculture and aviation fields. In the field of agriculture, to obtain abundant and quality yields, farmers are very dependent on weather conditions, especially rainfall. Rainfall is one of the factors that affect the safety of aircraft. To overcome the problems above, then it’s required a system that can accurately predict rainfall. In predicting rainfall, artificial neural network modeling is applied in this research. The method used in modeling this artificial neural network is backpropagation method. Backpropagation methods can result in better performance in repetitive exercises. This means that the weight of the ANN interconnection can approach the weight it should be. Another advantage of this method is the ability in the learning process adaptively and multilayer owned on this method there is a process of weight changes so as to minimize error (fault tolerance). Therefore, this method can guarantee good system resilience and consistently work well. The network is designed using 4 input variables, namely air temperature, air humidity, wind speed, and sunshine duration and 3 output variables ie low rainfall, medium rainfall, and high rainfall. Based on the research that has been done, the network can be used properly, as evidenced by the results of the prediction of the system precipitation is the same as the results of manual calculations.

  4. MetWAMer: eukaryotic translation initiation site prediction

    Directory of Open Access Journals (Sweden)

    Brendel Volker

    2008-09-01

    Full Text Available Abstract Background Translation initiation site (TIS identification is an important aspect of the gene annotation process, requisite for the accurate delineation of protein sequences from transcript data. We have developed the MetWAMer package for TIS prediction in eukaryotic open reading frames of non-viral origin. MetWAMer can be used as a stand-alone, third-party tool for post-processing gene structure annotations generated by external computational programs and/or pipelines, or directly integrated into gene structure prediction software implementations. Results MetWAMer currently implements five distinct methods for TIS prediction, the most accurate of which is a routine that combines weighted, signal-based translation initiation site scores and the contrast in coding potential of sequences flanking TISs using a perceptron. Also, our program implements clustering capabilities through use of the k-medoids algorithm, thereby enabling cluster-specific TIS parameter utilization. In practice, our static weight array matrix-based indexing method for parameter set lookup can be used with good results in data sets exhibiting moderate levels of 5'-complete coverage. Conclusion We demonstrate that improvements in statistically-based models for TIS prediction can be achieved by taking the class of each potential start-methionine into account pending certain testing conditions, and that our perceptron-based model is suitable for the TIS identification task. MetWAMer represents a well-documented, extensible, and freely available software system that can be readily re-trained for differing target applications and/or extended with existing and novel TIS prediction methods, to support further research efforts in this area.

  5. Ensemble method for dengue prediction.

    Science.gov (United States)

    Buczak, Anna L; Baugher, Benjamin; Moniz, Linda J; Bagley, Thomas; Babin, Steven M; Guven, Erhan

    2018-01-01

    In the 2015 NOAA Dengue Challenge, participants made three dengue target predictions for two locations (Iquitos, Peru, and San Juan, Puerto Rico) during four dengue seasons: 1) peak height (i.e., maximum weekly number of cases during a transmission season; 2) peak week (i.e., week in which the maximum weekly number of cases occurred); and 3) total number of cases reported during a transmission season. A dengue transmission season is the 12-month period commencing with the location-specific, historical week with the lowest number of cases. At the beginning of the Dengue Challenge, participants were provided with the same input data for developing the models, with the prediction testing data provided at a later date. Our approach used ensemble models created by combining three disparate types of component models: 1) two-dimensional Method of Analogues models incorporating both dengue and climate data; 2) additive seasonal Holt-Winters models with and without wavelet smoothing; and 3) simple historical models. Of the individual component models created, those with the best performance on the prior four years of data were incorporated into the ensemble models. There were separate ensembles for predicting each of the three targets at each of the two locations. Our ensemble models scored higher for peak height and total dengue case counts reported in a transmission season for Iquitos than all other models submitted to the Dengue Challenge. However, the ensemble models did not do nearly as well when predicting the peak week. The Dengue Challenge organizers scored the dengue predictions of the Challenge participant groups. Our ensemble approach was the best in predicting the total number of dengue cases reported for transmission season and peak height for Iquitos, Peru.

  6. Ensemble method for dengue prediction.

    Directory of Open Access Journals (Sweden)

    Anna L Buczak

    Full Text Available In the 2015 NOAA Dengue Challenge, participants made three dengue target predictions for two locations (Iquitos, Peru, and San Juan, Puerto Rico during four dengue seasons: 1 peak height (i.e., maximum weekly number of cases during a transmission season; 2 peak week (i.e., week in which the maximum weekly number of cases occurred; and 3 total number of cases reported during a transmission season. A dengue transmission season is the 12-month period commencing with the location-specific, historical week with the lowest number of cases. At the beginning of the Dengue Challenge, participants were provided with the same input data for developing the models, with the prediction testing data provided at a later date.Our approach used ensemble models created by combining three disparate types of component models: 1 two-dimensional Method of Analogues models incorporating both dengue and climate data; 2 additive seasonal Holt-Winters models with and without wavelet smoothing; and 3 simple historical models. Of the individual component models created, those with the best performance on the prior four years of data were incorporated into the ensemble models. There were separate ensembles for predicting each of the three targets at each of the two locations.Our ensemble models scored higher for peak height and total dengue case counts reported in a transmission season for Iquitos than all other models submitted to the Dengue Challenge. However, the ensemble models did not do nearly as well when predicting the peak week.The Dengue Challenge organizers scored the dengue predictions of the Challenge participant groups. Our ensemble approach was the best in predicting the total number of dengue cases reported for transmission season and peak height for Iquitos, Peru.

  7. Combining specificity determining and conserved residues improves functional site prediction

    Directory of Open Access Journals (Sweden)

    Gelfand Mikhail S

    2009-06-01

    Full Text Available Abstract Background Predicting the location of functionally important sites from protein sequence and/or structure is a long-standing problem in computational biology. Most current approaches make use of sequence conservation, assuming that amino acid residues conserved within a protein family are most likely to be functionally important. Most often these approaches do not consider many residues that act to define specific sub-functions within a family, or they make no distinction between residues important for function and those more relevant for maintaining structure (e.g. in the hydrophobic core. Many protein families bind and/or act on a variety of ligands, meaning that conserved residues often only bind a common ligand sub-structure or perform general catalytic activities. Results Here we present a novel method for functional site prediction based on identification of conserved positions, as well as those responsible for determining ligand specificity. We define Specificity-Determining Positions (SDPs, as those occupied by conserved residues within sub-groups of proteins in a family having a common specificity, but differ between groups, and are thus likely to account for specific recognition events. We benchmark the approach on enzyme families of known 3D structure with bound substrates, and find that in nearly all families residues predicted by SDPsite are in contact with the bound substrate, and that the addition of SDPs significantly improves functional site prediction accuracy. We apply SDPsite to various families of proteins containing known three-dimensional structures, but lacking clear functional annotations, and discusse several illustrative examples. Conclusion The results suggest a better means to predict functional details for the thousands of protein structures determined prior to a clear understanding of molecular function.

  8. Candidate Prediction Models and Methods

    DEFF Research Database (Denmark)

    Nielsen, Henrik Aalborg; Nielsen, Torben Skov; Madsen, Henrik

    2005-01-01

    This document lists candidate prediction models for Work Package 3 (WP3) of the PSO-project called ``Intelligent wind power prediction systems'' (FU4101). The main focus is on the models transforming numerical weather predictions into predictions of power production. The document also outlines...... the possibilities w.r.t. different numerical weather predictions actually available to the project....

  9. A method for evaluating the effectiveness of site characterization measurements

    International Nuclear Information System (INIS)

    Ditmars, J.D.

    1987-01-01

    A quantitative approach for evaluating the effectiveness of site characterization measurement activities is developed and illustrated with an example application to hypothetical measurement schemes at a potential geologic repository site for radioactive waste. The method is a general one and could also be applied at sites for underground disposal of hazardous chemicals. The approach presumes that measurements will be undertaken to support predictions of the performance of some aspect of a constructed facility or natural system. It requires a quantitative performance objective, such as groundwater travel time or contaminant concentration, against which to compare predictions of performance. The approach recognizes that such predictions are uncertain because the measurements upon which they are based are uncertain. The effectiveness of measurement activities is quantified by a confidence index, β, that reflects the number of standard deviations separating the best estimate of performance from the predetermined performance objective. Measurements that reduce the uncertainty in predictions lead to increased values of β. 5 refs., 4 figs

  10. Predicting sumoylation sites using support vector machines based on various sequence features, conformational flexibility and disorder.

    Science.gov (United States)

    Yavuz, Ahmet Sinan; Sezerman, Osman Ugur

    2014-01-01

    Sumoylation, which is a reversible and dynamic post-translational modification, is one of the vital processes in a cell. Before a protein matures to perform its function, sumoylation may alter its localization, interactions, and possibly structural conformation. Abberations in protein sumoylation has been linked with a variety of disorders and developmental anomalies. Experimental approaches to identification of sumoylation sites may not be effective due to the dynamic nature of sumoylation, laborsome experiments and their cost. Therefore, computational approaches may guide experimental identification of sumoylation sites and provide insights for further understanding sumoylation mechanism. In this paper, the effectiveness of using various sequence properties in predicting sumoylation sites was investigated with statistical analyses and machine learning approach employing support vector machines. These sequence properties were derived from windows of size 7 including position-specific amino acid composition, hydrophobicity, estimated sub-window volumes, predicted disorder, and conformational flexibility. 5-fold cross-validation results on experimentally identified sumoylation sites revealed that our method successfully predicts sumoylation sites with a Matthew's correlation coefficient, sensitivity, specificity, and accuracy equal to 0.66, 73%, 98%, and 97%, respectively. Additionally, we have showed that our method compares favorably to the existing prediction methods and basic regular expressions scanner. By using support vector machines, a new, robust method for sumoylation site prediction was introduced. Besides, the possible effects of predicted conformational flexibility and disorder on sumoylation site recognition were explored computationally for the first time to our knowledge as an additional parameter that could aid in sumoylation site prediction.

  11. Geospatial Analytics in Retail Site Selection and Sales Prediction.

    Science.gov (United States)

    Ting, Choo-Yee; Ho, Chiung Ching; Yee, Hui Jia; Matsah, Wan Razali

    2018-03-01

    Studies have shown that certain features from geography, demography, trade area, and environment can play a vital role in retail site selection, largely due to the impact they asserted on retail performance. Although the relevant features could be elicited by domain experts, determining the optimal feature set can be intractable and labor-intensive exercise. The challenges center around (1) how to determine features that are important to a particular retail business and (2) how to estimate retail sales performance given a new location? The challenges become apparent when the features vary across time. In this light, this study proposed a nonintervening approach by employing feature selection algorithms and subsequently sales prediction through similarity-based methods. The results of prediction were validated by domain experts. In this study, data sets from different sources were transformed and aggregated before an analytics data set that is ready for analysis purpose could be obtained. The data sets included data about feature location, population count, property type, education status, and monthly sales from 96 branches of a telecommunication company in Malaysia. The finding suggested that (1) optimal retail performance can only be achieved through fulfillment of specific location features together with the surrounding trade area characteristics and (2) similarity-based method can provide solution to retail sales prediction.

  12. Computational predictive methods for fracture and fatigue

    Science.gov (United States)

    Cordes, J.; Chang, A. T.; Nelson, N.; Kim, Y.

    1994-09-01

    The damage-tolerant design philosophy as used by aircraft industries enables aircraft components and aircraft structures to operate safely with minor damage, small cracks, and flaws. Maintenance and inspection procedures insure that damages developed during service remain below design values. When damage is found, repairs or design modifications are implemented and flight is resumed. Design and redesign guidelines, such as military specifications MIL-A-83444, have successfully reduced the incidence of damage and cracks. However, fatigue cracks continue to appear in aircraft well before the design life has expired. The F16 airplane, for instance, developed small cracks in the engine mount, wing support, bulk heads, the fuselage upper skin, the fuel shelf joints, and along the upper wings. Some cracks were found after 600 hours of the 8000 hour design service life and design modifications were required. Tests on the F16 plane showed that the design loading conditions were close to the predicted loading conditions. Improvements to analytic methods for predicting fatigue crack growth adjacent to holes, when multiple damage sites are present, and in corrosive environments would result in more cost-effective designs, fewer repairs, and fewer redesigns. The overall objective of the research described in this paper is to develop, verify, and extend the computational efficiency of analysis procedures necessary for damage tolerant design. This paper describes an elastic/plastic fracture method and an associated fatigue analysis method for damage tolerant design. Both methods are unique in that material parameters such as fracture toughness, R-curve data, and fatigue constants are not required. The methods are implemented with a general-purpose finite element package. Several proof-of-concept examples are given. With further development, the methods could be extended for analysis of multi-site damage, creep-fatigue, and corrosion fatigue problems.

  13. Site and stand analysis for growth prediction of Eucalyptus grandis ...

    African Journals Online (AJOL)

    The integration of site information with that of tree growth is of special importance in Zululand, where sustainable supply of timber is essential for local processing and export commitments. Site prediction growth models need to be based on easily attainable input variables that are suitable for operational implementation by ...

  14. Using sequence-specific chemical and structural properties of DNA to predict transcription factor binding sites.

    Directory of Open Access Journals (Sweden)

    Amy L Bauer

    2010-11-01

    Full Text Available An important step in understanding gene regulation is to identify the DNA binding sites recognized by each transcription factor (TF. Conventional approaches to prediction of TF binding sites involve the definition of consensus sequences or position-specific weight matrices and rely on statistical analysis of DNA sequences of known binding sites. Here, we present a method called SiteSleuth in which DNA structure prediction, computational chemistry, and machine learning are applied to develop models for TF binding sites. In this approach, binary classifiers are trained to discriminate between true and false binding sites based on the sequence-specific chemical and structural features of DNA. These features are determined via molecular dynamics calculations in which we consider each base in different local neighborhoods. For each of 54 TFs in Escherichia coli, for which at least five DNA binding sites are documented in RegulonDB, the TF binding sites and portions of the non-coding genome sequence are mapped to feature vectors and used in training. According to cross-validation analysis and a comparison of computational predictions against ChIP-chip data available for the TF Fis, SiteSleuth outperforms three conventional approaches: Match, MATRIX SEARCH, and the method of Berg and von Hippel. SiteSleuth also outperforms QPMEME, a method similar to SiteSleuth in that it involves a learning algorithm. The main advantage of SiteSleuth is a lower false positive rate.

  15. An Overview of the Prediction of Protein DNA-Binding Sites

    Directory of Open Access Journals (Sweden)

    Jingna Si

    2015-03-01

    Full Text Available Interactions between proteins and DNA play an important role in many essential biological processes such as DNA replication, transcription, splicing, and repair. The identification of amino acid residues involved in DNA-binding sites is critical for understanding the mechanism of these biological activities. In the last decade, numerous computational approaches have been developed to predict protein DNA-binding sites based on protein sequence and/or structural information, which play an important role in complementing experimental strategies. At this time, approaches can be divided into three categories: sequence-based DNA-binding site prediction, structure-based DNA-binding site prediction, and homology modeling and threading. In this article, we review existing research on computational methods to predict protein DNA-binding sites, which includes data sets, various residue sequence/structural features, machine learning methods for comparison and selection, evaluation methods, performance comparison of different tools, and future directions in protein DNA-binding site prediction. In particular, we detail the meta-analysis of protein DNA-binding sites. We also propose specific implications that are likely to result in novel prediction methods, increased performance, or practical applications.

  16. An overview of the prediction of protein DNA-binding sites.

    Science.gov (United States)

    Si, Jingna; Zhao, Rui; Wu, Rongling

    2015-03-06

    Interactions between proteins and DNA play an important role in many essential biological processes such as DNA replication, transcription, splicing, and repair. The identification of amino acid residues involved in DNA-binding sites is critical for understanding the mechanism of these biological activities. In the last decade, numerous computational approaches have been developed to predict protein DNA-binding sites based on protein sequence and/or structural information, which play an important role in complementing experimental strategies. At this time, approaches can be divided into three categories: sequence-based DNA-binding site prediction, structure-based DNA-binding site prediction, and homology modeling and threading. In this article, we review existing research on computational methods to predict protein DNA-binding sites, which includes data sets, various residue sequence/structural features, machine learning methods for comparison and selection, evaluation methods, performance comparison of different tools, and future directions in protein DNA-binding site prediction. In particular, we detail the meta-analysis of protein DNA-binding sites. We also propose specific implications that are likely to result in novel prediction methods, increased performance, or practical applications.

  17. Non-Invasive Seismic Methods for Earthquake Site Classification Applied to Ontario Bridge Sites

    Science.gov (United States)

    Bilson Darko, A.; Molnar, S.; Sadrekarimi, A.

    2017-12-01

    How a site responds to earthquake shaking and its corresponding damage is largely influenced by the underlying ground conditions through which it propagates. The effects of site conditions on propagating seismic waves can be predicted from measurements of the shear wave velocity (Vs) of the soil layer(s) and the impedance ratio between bedrock and soil. Currently the seismic design of new buildings and bridges (2015 Canadian building and bridge codes) requires determination of the time-averaged shear-wave velocity of the upper 30 metres (Vs30) of a given site. In this study, two in situ Vs profiling methods; Multichannel Analysis of Surface Waves (MASW) and Ambient Vibration Array (AVA) methods are used to determine Vs30 at chosen bridge sites in Ontario, Canada. Both active-source (MASW) and passive-source (AVA) surface wave methods are used at each bridge site to obtain Rayleigh-wave phase velocities over a wide frequency bandwidth. The dispersion curve is jointly inverted with each site's amplification function (microtremor horizontal-to-vertical spectral ratio) to obtain shear-wave velocity profile(s). We apply our non-invasive testing at three major infrastructure projects, e.g., five bridge sites along the Rt. Hon. Herb Gray Parkway in Windsor, Ontario. Our non-invasive testing is co-located with previous invasive testing, including Standard Penetration Test (SPT), Cone Penetration Test and downhole Vs data. Correlations between SPT blowcount and Vs are developed for the different soil types sampled at our Ontario bridge sites. A robust earthquake site classification procedure (reliable Vs30 estimates) for bridge sites across Ontario is evaluated from available combinations of invasive and non-invasive site characterization methods.

  18. NEURAL METHODS FOR THE FINANCIAL PREDICTION

    OpenAIRE

    Jerzy Balicki; Piotr Dryja; Waldemar Korłub; Piotr Przybyłek; Maciej Tyszka; Marcin Zadroga; Marcin Zakidalski

    2016-01-01

    Artificial neural networks can be used to predict share investment on the stock market, assess the reliability of credit client or predicting banking crises. Moreover, this paper discusses the principles of cooperation neural network algorithms with evolutionary method, and support vector machines. In addition, a reference is made to other methods of artificial intelligence, which are used in finance prediction.

  19. NEURAL METHODS FOR THE FINANCIAL PREDICTION

    Directory of Open Access Journals (Sweden)

    Jerzy Balicki

    2016-06-01

    Full Text Available Artificial neural networks can be used to predict share investment on the stock market, assess the reliability of credit client or predicting banking crises. Moreover, this paper discusses the principles of cooperation neural network algorithms with evolutionary method, and support vector machines. In addition, a reference is made to other methods of artificial intelligence, which are used in finance prediction.

  20. Predictive Temperature Equations for Three Sites at the Grand Canyon

    Science.gov (United States)

    McLaughlin, Katrina Marie Neitzel

    Climate data collected at a number of automated weather stations were used to create a series of predictive equations spanning from December 2009 to May 2010 in order to better predict the temperatures along hiking trails within the Grand Canyon. The central focus of this project is how atmospheric variables interact and can be combined to predict the weather in the Grand Canyon at the Indian Gardens, Phantom Ranch, and Bright Angel sites. Through the use of statistical analysis software and data regression, predictive equations were determined. The predictive equations are simple or multivariable best fits that reflect the curvilinear nature of the data. With data analysis software curves resulting from the predictive equations were plotted along with the observed data. Each equation's reduced chi2 was determined to aid the visual examination of the predictive equations' ability to reproduce the observed data. From this information an equation or pair of equations was determined to be the best of the predictive equations. Although a best predictive equation for each month and season was determined for each site, future work may refine equations to result in a more accurate predictive equation.

  1. Prediction methods and databases within chemoinformatics

    DEFF Research Database (Denmark)

    Jónsdóttir, Svava Osk; Jørgensen, Flemming Steen; Brunak, Søren

    2005-01-01

    MOTIVATION: To gather information about available databases and chemoinformatics methods for prediction of properties relevant to the drug discovery and optimization process. RESULTS: We present an overview of the most important databases with 2-dimensional and 3-dimensional structural information...... about drugs and drug candidates, and of databases with relevant properties. Access to experimental data and numerical methods for selecting and utilizing these data is crucial for developing accurate predictive in silico models. Many interesting predictive methods for classifying the suitability...

  2. Prediction of protein-protein interaction sites in sequences and 3D structures by random forests.

    Directory of Open Access Journals (Sweden)

    Mile Sikić

    2009-01-01

    Full Text Available Identifying interaction sites in proteins provides important clues to the function of a protein and is becoming increasingly relevant in topics such as systems biology and drug discovery. Although there are numerous papers on the prediction of interaction sites using information derived from structure, there are only a few case reports on the prediction of interaction residues based solely on protein sequence. Here, a sliding window approach is combined with the Random Forests method to predict protein interaction sites using (i a combination of sequence- and structure-derived parameters and (ii sequence information alone. For sequence-based prediction we achieved a precision of 84% with a 26% recall and an F-measure of 40%. When combined with structural information, the prediction performance increases to a precision of 76% and a recall of 38% with an F-measure of 51%. We also present an attempt to rationalize the sliding window size and demonstrate that a nine-residue window is the most suitable for predictor construction. Finally, we demonstrate the applicability of our prediction methods by modeling the Ras-Raf complex using predicted interaction sites as target binding interfaces. Our results suggest that it is possible to predict protein interaction sites with quite a high accuracy using only sequence information.

  3. Fast dynamics perturbation analysis for prediction of protein functional sites

    Directory of Open Access Journals (Sweden)

    Cohn Judith D

    2008-01-01

    Full Text Available Abstract Background We present a fast version of the dynamics perturbation analysis (DPA algorithm to predict functional sites in protein structures. The original DPA algorithm finds regions in proteins where interactions cause a large change in the protein conformational distribution, as measured using the relative entropy Dx. Such regions are associated with functional sites. Results The Fast DPA algorithm, which accelerates DPA calculations, is motivated by an empirical observation that Dx in a normal-modes model is highly correlated with an entropic term that only depends on the eigenvalues of the normal modes. The eigenvalues are accurately estimated using first-order perturbation theory, resulting in a N-fold reduction in the overall computational requirements of the algorithm, where N is the number of residues in the protein. The performance of the original and Fast DPA algorithms was compared using protein structures from a standard small-molecule docking test set. For nominal implementations of each algorithm, top-ranked Fast DPA predictions overlapped the true binding site 94% of the time, compared to 87% of the time for original DPA. In addition, per-protein recall statistics (fraction of binding-site residues that are among predicted residues were slightly better for Fast DPA. On the other hand, per-protein precision statistics (fraction of predicted residues that are among binding-site residues were slightly better using original DPA. Overall, the performance of Fast DPA in predicting ligand-binding-site residues was comparable to that of the original DPA algorithm. Conclusion Compared to the original DPA algorithm, the decreased run time with comparable performance makes Fast DPA well-suited for implementation on a web server and for high-throughput analysis.

  4. A GIS approach for predicting prehistoric site locations.

    Energy Technology Data Exchange (ETDEWEB)

    Kuiper, J. A.; Wescott, K. L.

    1999-08-04

    Use of geographic information system (GIS)-based predictive mapping to locate areas of high potential for prehistoric archaeological sites is becoming increasingly popular among archaeologists. Knowledge of the environmental variables influencing activities of original inhabitants is used to produce GIS layers representing the spatial distribution of those variables. The GIS layers are then analyzed to identify locations where combinations of environmental variables match patterns observed at known prehistoric sites. Presented are the results of a study to locate high-potential areas for prehistoric sites in a largely unsurveyed area of 39,000 acres in the Upper Chesapeake Bay region, including details of the analysis process. The project used environmental data from over 500 known sites in other parts of the region and the results corresponded well with known sites in the study area.

  5. Computational prediction of muon stopping sites using ab initio random structure searching (AIRSS)

    Science.gov (United States)

    Liborio, Leandro; Sturniolo, Simone; Jochym, Dominik

    2018-04-01

    The stopping site of the muon in a muon-spin relaxation experiment is in general unknown. There are some techniques that can be used to guess the muon stopping site, but they often rely on approximations and are not generally applicable to all cases. In this work, we propose a purely theoretical method to predict muon stopping sites in crystalline materials from first principles. The method is based on a combination of ab initio calculations, random structure searching, and machine learning, and it has successfully predicted the MuT and MuBC stopping sites of muonium in Si, diamond, and Ge, as well as the muonium stopping site in LiF, without any recourse to experimental results. The method makes use of Soprano, a Python library developed to aid ab initio computational crystallography, that was publicly released and contains all the software tools necessary to reproduce our analysis.

  6. The IntFOLD server: an integrated web resource for protein fold recognition, 3D model quality assessment, intrinsic disorder prediction, domain prediction and ligand binding site prediction.

    Science.gov (United States)

    Roche, Daniel B; Buenavista, Maria T; Tetchner, Stuart J; McGuffin, Liam J

    2011-07-01

    The IntFOLD server is a novel independent server that integrates several cutting edge methods for the prediction of structure and function from sequence. Our guiding principles behind the server development were as follows: (i) to provide a simple unified resource that makes our prediction software accessible to all and (ii) to produce integrated output for predictions that can be easily interpreted. The output for predictions is presented as a simple table that summarizes all results graphically via plots and annotated 3D models. The raw machine readable data files for each set of predictions are also provided for developers, which comply with the Critical Assessment of Methods for Protein Structure Prediction (CASP) data standards. The server comprises an integrated suite of five novel methods: nFOLD4, for tertiary structure prediction; ModFOLD 3.0, for model quality assessment; DISOclust 2.0, for disorder prediction; DomFOLD 2.0 for domain prediction; and FunFOLD 1.0, for ligand binding site prediction. Predictions from the IntFOLD server were found to be competitive in several categories in the recent CASP9 experiment. The IntFOLD server is available at the following web site: http://www.reading.ac.uk/bioinf/IntFOLD/.

  7. Machine learning methods for metabolic pathway prediction

    Directory of Open Access Journals (Sweden)

    Karp Peter D

    2010-01-01

    Full Text Available Abstract Background A key challenge in systems biology is the reconstruction of an organism's metabolic network from its genome sequence. One strategy for addressing this problem is to predict which metabolic pathways, from a reference database of known pathways, are present in the organism, based on the annotated genome of the organism. Results To quantitatively validate methods for pathway prediction, we developed a large "gold standard" dataset of 5,610 pathway instances known to be present or absent in curated metabolic pathway databases for six organisms. We defined a collection of 123 pathway features, whose information content we evaluated with respect to the gold standard. Feature data were used as input to an extensive collection of machine learning (ML methods, including naïve Bayes, decision trees, and logistic regression, together with feature selection and ensemble methods. We compared the ML methods to the previous PathoLogic algorithm for pathway prediction using the gold standard dataset. We found that ML-based prediction methods can match the performance of the PathoLogic algorithm. PathoLogic achieved an accuracy of 91% and an F-measure of 0.786. The ML-based prediction methods achieved accuracy as high as 91.2% and F-measure as high as 0.787. The ML-based methods output a probability for each predicted pathway, whereas PathoLogic does not, which provides more information to the user and facilitates filtering of predicted pathways. Conclusions ML methods for pathway prediction perform as well as existing methods, and have qualitative advantages in terms of extensibility, tunability, and explainability. More advanced prediction methods and/or more sophisticated input features may improve the performance of ML methods. However, pathway prediction performance appears to be limited largely by the ability to correctly match enzymes to the reactions they catalyze based on genome annotations.

  8. Machine learning methods for metabolic pathway prediction

    Science.gov (United States)

    2010-01-01

    Background A key challenge in systems biology is the reconstruction of an organism's metabolic network from its genome sequence. One strategy for addressing this problem is to predict which metabolic pathways, from a reference database of known pathways, are present in the organism, based on the annotated genome of the organism. Results To quantitatively validate methods for pathway prediction, we developed a large "gold standard" dataset of 5,610 pathway instances known to be present or absent in curated metabolic pathway databases for six organisms. We defined a collection of 123 pathway features, whose information content we evaluated with respect to the gold standard. Feature data were used as input to an extensive collection of machine learning (ML) methods, including naïve Bayes, decision trees, and logistic regression, together with feature selection and ensemble methods. We compared the ML methods to the previous PathoLogic algorithm for pathway prediction using the gold standard dataset. We found that ML-based prediction methods can match the performance of the PathoLogic algorithm. PathoLogic achieved an accuracy of 91% and an F-measure of 0.786. The ML-based prediction methods achieved accuracy as high as 91.2% and F-measure as high as 0.787. The ML-based methods output a probability for each predicted pathway, whereas PathoLogic does not, which provides more information to the user and facilitates filtering of predicted pathways. Conclusions ML methods for pathway prediction perform as well as existing methods, and have qualitative advantages in terms of extensibility, tunability, and explainability. More advanced prediction methods and/or more sophisticated input features may improve the performance of ML methods. However, pathway prediction performance appears to be limited largely by the ability to correctly match enzymes to the reactions they catalyze based on genome annotations. PMID:20064214

  9. Exploiting protein flexibility to predict the location of allosteric sites

    Directory of Open Access Journals (Sweden)

    Panjkovich Alejandro

    2012-10-01

    Full Text Available Abstract Background Allostery is one of the most powerful and common ways of regulation of protein activity. However, for most allosteric proteins identified to date the mechanistic details of allosteric modulation are not yet well understood. Uncovering common mechanistic patterns underlying allostery would allow not only a better academic understanding of the phenomena, but it would also streamline the design of novel therapeutic solutions. This relatively unexplored therapeutic potential and the putative advantages of allosteric drugs over classical active-site inhibitors fuel the attention allosteric-drug research is receiving at present. A first step to harness the regulatory potential and versatility of allosteric sites, in the context of drug-discovery and design, would be to detect or predict their presence and location. In this article, we describe a simple computational approach, based on the effect allosteric ligands exert on protein flexibility upon binding, to predict the existence and position of allosteric sites on a given protein structure. Results By querying the literature and a recently available database of allosteric sites, we gathered 213 allosteric proteins with structural information that we further filtered into a non-redundant set of 91 proteins. We performed normal-mode analysis and observed significant changes in protein flexibility upon allosteric-ligand binding in 70% of the cases. These results agree with the current view that allosteric mechanisms are in many cases governed by changes in protein dynamics caused by ligand binding. Furthermore, we implemented an approach that achieves 65% positive predictive value in identifying allosteric sites within the set of predicted cavities of a protein (stricter parameters set, 0.22 sensitivity, by combining the current analysis on dynamics with previous results on structural conservation of allosteric sites. We also analyzed four biological examples in detail, revealing

  10. CaMELS: In silico prediction of calmodulin binding proteins and their binding sites.

    Science.gov (United States)

    Abbasi, Wajid Arshad; Asif, Amina; Andleeb, Saiqa; Minhas, Fayyaz Ul Amir Afsar

    2017-09-01

    Due to Ca 2+ -dependent binding and the sequence diversity of Calmodulin (CaM) binding proteins, identifying CaM interactions and binding sites in the wet-lab is tedious and costly. Therefore, computational methods for this purpose are crucial to the design of such wet-lab experiments. We present an algorithm suite called CaMELS (CalModulin intEraction Learning System) for predicting proteins that interact with CaM as well as their binding sites using sequence information alone. CaMELS offers state of the art accuracy for both CaM interaction and binding site prediction and can aid biologists in studying CaM binding proteins. For CaM interaction prediction, CaMELS uses protein sequence features coupled with a large-margin classifier. CaMELS models the binding site prediction problem using multiple instance machine learning with a custom optimization algorithm which allows more effective learning over imprecisely annotated CaM-binding sites during training. CaMELS has been extensively benchmarked using a variety of data sets, mutagenic studies, proteome-wide Gene Ontology enrichment analyses and protein structures. Our experiments indicate that CaMELS outperforms simple motif-based search and other existing methods for interaction and binding site prediction. We have also found that the whole sequence of a protein, rather than just its binding site, is important for predicting its interaction with CaM. Using the machine learning model in CaMELS, we have identified important features of protein sequences for CaM interaction prediction as well as characteristic amino acid sub-sequences and their relative position for identifying CaM binding sites. Python code for training and evaluating CaMELS together with a webserver implementation is available at the URL: http://faculty.pieas.edu.pk/fayyaz/software.html#camels. © 2017 Wiley Periodicals, Inc.

  11. NetAcet: prediction of N-terminal acetylation sites

    DEFF Research Database (Denmark)

    Kiemer, Lars; Bendtsen, Jannick Dyrløv; Blom, Nikolaj

    2005-01-01

    Summary: We present here a neural network based method for prediction of N-terminal acetylation-by far the most abundant post-translational modification in eukaryotes. The method was developed on a yeast dataset for N-acetyltransferase A (NatA) acetylation, which is the type of N-acetylation for ......Summary: We present here a neural network based method for prediction of N-terminal acetylation-by far the most abundant post-translational modification in eukaryotes. The method was developed on a yeast dataset for N-acetyltransferase A (NatA) acetylation, which is the type of N...

  12. Seismic Hazard Assessment in Site Evaluation for Nuclear Installations: Ground Motion Prediction Equations and Site Response

    International Nuclear Information System (INIS)

    2016-07-01

    The objective of this publication is to provide the state-of-the-art practice and detailed technical elements related to ground motion evaluation by ground motion prediction equations (GMPEs) and site response in the context of seismic hazard assessments as recommended in IAEA Safety Standards Series No. SSG-9, Seismic Hazards in Site Evaluation for Nuclear Installations. The publication includes the basics of GMPEs, ground motion simulation, selection and adjustment of GMPEs, site characterization, and modelling of site response in order to improve seismic hazard assessment. The text aims at delineating the most important aspects of these topics (including current practices, criticalities and open problems) within a coherent framework. In particular, attention has been devoted to filling conceptual gaps. It is written as a reference text for trained users who are responsible for planning preparatory seismic hazard analyses for siting of all nuclear installations and/or providing constraints for anti-seismic design and retrofitting of existing structures

  13. Development of METAL-ACTIVE SITE and ZINCCLUSTER tool to predict active site pockets.

    Science.gov (United States)

    Ajitha, M; Sundar, K; Arul Mugilan, S; Arumugam, S

    2018-03-01

    The advent of whole genome sequencing leads to increasing number of proteins with known amino acid sequences. Despite many efforts, the number of proteins with resolved three dimensional structures is still low. One of the challenging tasks the structural biologists face is the prediction of the interaction of metal ion with any protein for which the structure is unknown. Based on the information available in Protein Data Bank, a site (METALACTIVE INTERACTION) has been generated which displays information for significant high preferential and low-preferential combination of endogenous ligands for 49 metal ions. User can also gain information about the residues present in the first and second coordination sphere as it plays a major role in maintaining the structure and function of metalloproteins in biological system. In this paper, a novel computational tool (ZINCCLUSTER) is developed, which can predict the zinc metal binding sites of proteins even if only the primary sequence is known. The purpose of this tool is to predict the active site cluster of an uncharacterized protein based on its primary sequence or a 3D structure. The tool can predict amino acids interacting with a metal or vice versa. This tool is based on the occurrence of significant triplets and it is tested to have higher prediction accuracy when compared to that of other available techniques. © 2017 Wiley Periodicals, Inc.

  14. Better estimation of protein-DNA interaction parameters improve prediction of functional sites

    Directory of Open Access Journals (Sweden)

    O'Flanagan Ruadhan A

    2008-12-01

    Full Text Available Abstract Background Characterizing transcription factor binding motifs is a common bioinformatics task. For transcription factors with variable binding sites, we need to get many suboptimal binding sites in our training dataset to get accurate estimates of free energy penalties for deviating from the consensus DNA sequence. One procedure to do that involves a modified SELEX (Systematic Evolution of Ligands by Exponential Enrichment method designed to produce many such sequences. Results We analyzed low stringency SELEX data for E. coli Catabolic Activator Protein (CAP, and we show here that appropriate quantitative analysis improves our ability to predict in vitro affinity. To obtain large number of sequences required for this analysis we used a SELEX SAGE protocol developed by Roulet et al. The sequences obtained from here were subjected to bioinformatic analysis. The resulting bioinformatic model characterizes the sequence specificity of the protein more accurately than those sequence specificities predicted from previous analysis just by using a few known binding sites available in the literature. The consequences of this increase in accuracy for prediction of in vivo binding sites (and especially functional ones in the E. coli genome are also discussed. We measured the dissociation constants of several putative CAP binding sites by EMSA (Electrophoretic Mobility Shift Assay and compared the affinities to the bioinformatics scores provided by methods like the weight matrix method and QPMEME (Quadratic Programming Method of Energy Matrix Estimation trained on known binding sites as well as on the new sites from SELEX SAGE data. We also checked predicted genome sites for conservation in the related species S. typhimurium. We found that bioinformatics scores based on SELEX SAGE data does better in terms of prediction of physical binding energies as well as in detecting functional sites. Conclusion We think that training binding site detection

  15. Prediction of nucleosome positioning based on transcription factor binding sites.

    Directory of Open Access Journals (Sweden)

    Xianfu Yi

    Full Text Available BACKGROUND: The DNA of all eukaryotic organisms is packaged into nucleosomes, the basic repeating units of chromatin. The nucleosome consists of a histone octamer around which a DNA core is wrapped and the linker histone H1, which is associated with linker DNA. By altering the accessibility of DNA sequences, the nucleosome has profound effects on all DNA-dependent processes. Understanding the factors that influence nucleosome positioning is of great importance for the study of genomic control mechanisms. Transcription factors (TFs have been suggested to play a role in nucleosome positioning in vivo. PRINCIPAL FINDINGS: Here, the minimum redundancy maximum relevance (mRMR feature selection algorithm, the nearest neighbor algorithm (NNA, and the incremental feature selection (IFS method were used to identify the most important TFs that either favor or inhibit nucleosome positioning by analyzing the numbers of transcription factor binding sites (TFBSs in 53,021 nucleosomal DNA sequences and 50,299 linker DNA sequences. A total of nine important families of TFs were extracted from 35 families, and the overall prediction accuracy was 87.4% as evaluated by the jackknife cross-validation test. CONCLUSIONS: Our results are consistent with the notion that TFs are more likely to bind linker DNA sequences than the sequences in the nucleosomes. In addition, our results imply that there may be some TFs that are important for nucleosome positioning but that play an insignificant role in discriminating nucleosome-forming DNA sequences from nucleosome-inhibiting DNA sequences. The hypothesis that TFs play a role in nucleosome positioning is, thus, confirmed by the results of this study.

  16. Predicting on-site environmental impacts of municipal engineering works

    International Nuclear Information System (INIS)

    Gangolells, Marta; Casals, Miquel; Forcada, Núria; Macarulla, Marcel

    2014-01-01

    The research findings fill a gap in the body of knowledge by presenting an effective way to evaluate the significance of on-site environmental impacts of municipal engineering works prior to the construction stage. First, 42 on-site environmental impacts of municipal engineering works were identified by means of a process-oriented approach. Then, 46 indicators and their corresponding significance limits were determined on the basis of a statistical analysis of 25 new-build and remodelling municipal engineering projects. In order to ensure the objectivity of the assessment process, direct and indirect indicators were always based on quantitative data from the municipal engineering project documents. Finally, two case studies were analysed and found to illustrate the practical use of the proposed model. The model highlights the significant environmental impacts of a particular municipal engineering project prior to the construction stage. Consequently, preventive actions can be planned and implemented during on-site activities. The results of the model also allow a comparison of proposed municipal engineering projects and alternatives with respect to the overall on-site environmental impact and the absolute importance of a particular environmental aspect. These findings are useful within the framework of the environmental impact assessment process, as they help to improve the identification and evaluation of on-site environmental aspects of municipal engineering works. The findings may also be of use to construction companies that are willing to implement an environmental management system or simply wish to improve on-site environmental performance in municipal engineering projects. -- Highlights: • We present a model to predict the environmental impacts of municipal engineering works. • It highlights significant on-site environmental impacts prior to the construction stage. • Findings are useful within the environmental impact assessment process. • They also

  17. Predicting on-site environmental impacts of municipal engineering works

    Energy Technology Data Exchange (ETDEWEB)

    Gangolells, Marta, E-mail: marta.gangolells@upc.edu; Casals, Miquel, E-mail: miquel.casals@upc.edu; Forcada, Núria, E-mail: nuria.forcada@upc.edu; Macarulla, Marcel, E-mail: marcel.macarulla@upc.edu

    2014-01-15

    The research findings fill a gap in the body of knowledge by presenting an effective way to evaluate the significance of on-site environmental impacts of municipal engineering works prior to the construction stage. First, 42 on-site environmental impacts of municipal engineering works were identified by means of a process-oriented approach. Then, 46 indicators and their corresponding significance limits were determined on the basis of a statistical analysis of 25 new-build and remodelling municipal engineering projects. In order to ensure the objectivity of the assessment process, direct and indirect indicators were always based on quantitative data from the municipal engineering project documents. Finally, two case studies were analysed and found to illustrate the practical use of the proposed model. The model highlights the significant environmental impacts of a particular municipal engineering project prior to the construction stage. Consequently, preventive actions can be planned and implemented during on-site activities. The results of the model also allow a comparison of proposed municipal engineering projects and alternatives with respect to the overall on-site environmental impact and the absolute importance of a particular environmental aspect. These findings are useful within the framework of the environmental impact assessment process, as they help to improve the identification and evaluation of on-site environmental aspects of municipal engineering works. The findings may also be of use to construction companies that are willing to implement an environmental management system or simply wish to improve on-site environmental performance in municipal engineering projects. -- Highlights: • We present a model to predict the environmental impacts of municipal engineering works. • It highlights significant on-site environmental impacts prior to the construction stage. • Findings are useful within the environmental impact assessment process. • They also

  18. Assessment of Mars Exploration Rover Landing Site Predictions

    Science.gov (United States)

    Golombek, M. P.

    2005-05-01

    Comprehensive analyses of remote sensing data during the 3-year effort to select the Mars Exploration Rover landing sites at Gusev crater and Meridiani Planum correctly predicted the safe and trafficable surfaces explored by the two rovers. Gusev crater was predicted to be a relatively low relief surface that was comparably dusty, but less rocky than the Viking landing sites. Available data for Meridiani Planum indicated a very flat plain composed of basaltic sand to granules and hematite that would look completely unlike any of the existing landing sites with a dark, low albedo surface, little dust and very few rocks. Orbital thermal inertia measurements of 315 J m-2 s-0.5 K-1 at Gusev suggested surfaces dominated by duricrust to cemented soil-like materials or cohesionless sand or granules, which is consistent with observed soil characteristics and measured thermal inertias from the surface. THEMIS thermal inertias along the traverse at Gusev vary from 285 at the landing site to 330 around Bonneville rim and show systematic variations that can be related to the observed increase in rock abundance (5-30%). Meridiani has an orbital bulk inertia of ~200, similar to measured surface inertias that correspond to observed surfaces dominated by 0.2 mm sand size particles. Rock abundance derived from orbital thermal differencing techniques suggested that Meridiani Planum would have very low rock abundance, consistent with the rock free plain traversed by Opportunity. Spirit landed in an 8% orbital rock abundance pixel, consistent with the measured 7% of the surface covered by rocks >0.04 m diameter at the landing site, which is representative of the plains away from craters. The orbital albedo of the Spirit traverse varies from 0.19 to 0.30, consistent with surface measurements in and out of dust devil tracks. Opportunity is the first landing in a low albedo portion of Mars as seen from orbit, which is consistent with the dark, dust-free surface and measured albedos. The

  19. Predicting lodgepole pine site index from climatic parameters in Alberta.

    Science.gov (United States)

    Robert A. Monserud; Shongming Huang; Yuqing. Yang

    2006-01-01

    We sought to evaluate the impact of climatic variables on site productivity of lodgepole pine (Pinus contorta var. latifolia Engelm.) for the province of Alberta. Climatic data were obtained from the Alberta Climate Model, which is based on 30-year normals from the provincial weather station network. Mapping methods were based...

  20. Method for Predicting Thermal Buckling in Rails

    Science.gov (United States)

    2018-01-01

    A method is proposed herein for predicting the onset of thermal buckling in rails in such a way as to provide a means of avoiding this type of potentially devastating failure. The method consists of the development of a thermomechanical model of rail...

  1. Work characteristics predict the development of multi-site musculoskeletal pain

    NARCIS (Netherlands)

    Oakman, J.; Wind, A. de; Heuvel, S.G. van den; Beek, A.J. van der

    2017-01-01

    Purpose. Musculoskeletal pain in more than one body region is common and a barrier to sustaining employment. We aimed to examine whether work characteristics predict the development of multi-site pain (MSP), and to determine differences in work-related predictors between age groups. Methods. This

  2. Application of GIS in prediction and assessment system of off-site accident consequence for NPP

    International Nuclear Information System (INIS)

    Wang Xingyu; Shi Zhongqi

    2002-01-01

    The assessment and prediction software system of off-site accident consequence for Guangdong Nuclear Power Plant (GNARD2.0) is a GIS-based software system. The spatial analysis of radioactive materials and doses with geographic information is available in this system. The structure and functions of the GNARD system and the method of applying ArcView GIS are presented

  3. Methods and systems for identifying ligand-protein binding sites

    KAUST Repository

    Gao, Xin

    2016-05-06

    The invention provides a novel integrated structure and system-based approach for drug target prediction that enables the large-scale discovery of new targets for existing drugs Novel computer-readable storage media and computer systems are also provided. Methods and systems of the invention use novel sequence order-independent structure alignment, hierarchical clustering, and probabilistic sequence similarity techniques to construct a probabilistic pocket ensemble (PPE) that captures even promiscuous structural features of different binding sites for a drug on known targets. The drug\\'s PPE is combined with an approximation of the drug delivery profile to facilitate large-scale prediction of novel drug- protein interactions with several applications to biological research and drug development.

  4. Prediction Methods for Blood Glucose Concentration

    DEFF Research Database (Denmark)

    “Recent Results on Glucose–Insulin Predictions by Means of a State Observer for Time-Delay Systems” by Pasquale Palumbo et al. introduces a prediction model which in real time predicts the insulin concentration in blood which in turn is used in a control system. The method is tested in simulation...... EEG signals to predict upcoming hypoglycemic situations in real-time by employing artificial neural networks. The results of a 30-day long clinical study with the implanted device and the developed algorithm are presented. The chapter “Meta-Learning Based Blood Glucose Predictor for Diabetic......, but the insulin amount is chosen using factors that account for this expectation. The increasing availability of more accurate continuous blood glucose measurement (CGM) systems is attracting much interest to the possibilities of explicit prediction of future BG values. Against this background, in 2014 a two...

  5. A method for predicting monthly rainfall patterns

    International Nuclear Information System (INIS)

    Njau, E.C.

    1987-11-01

    A brief survey is made of previous methods that have been used to predict rainfall trends or drought spells in different parts of the earth. The basic methodologies or theoretical strategies used in these methods are compared with contents of a recent theory of Sun-Weather/Climate links (Njau, 1985a; 1985b; 1986; 1987a; 1987b; 1987c) which point towards the possibility of practical climatic predictions. It is shown that not only is the theoretical basis of each of these methodologies or strategies fully incorporated into the above-named theory, but also this theory may be used to develop a technique by which future monthly rainfall patterns can be predicted in further and finer details. We describe the latter technique and then illustrate its workability by means of predictions made on monthly rainfall patterns in some East African meteorological stations. (author). 43 refs, 11 figs, 2 tabs

  6. Protein-protein interaction site predictions with minimum covariance determinant and Mahalanobis distance.

    Science.gov (United States)

    Qiu, Zhijun; Zhou, Bo; Yuan, Jiangfeng

    2017-11-21

    Protein-protein interaction site (PPIS) prediction must deal with the diversity of interaction sites that limits their prediction accuracy. Use of proteins with unknown or unidentified interactions can also lead to missing interfaces. Such data errors are often brought into the training dataset. In response to these two problems, we used the minimum covariance determinant (MCD) method to refine the training data to build a predictor with better performance, utilizing its ability of removing outliers. In order to predict test data in practice, a method based on Mahalanobis distance was devised to select proper test data as input for the predictor. With leave-one-validation and independent test, after the Mahalanobis distance screening, our method achieved higher performance according to Matthews correlation coefficient (MCC), although only a part of test data could be predicted. These results indicate that data refinement is an efficient approach to improve protein-protein interaction site prediction. By further optimizing our method, it is hopeful to develop predictors of better performance and wide range of application. Copyright © 2017 Elsevier Ltd. All rights reserved.

  7. Site characterization and validation. Stage 2 - Preliminary predictions

    International Nuclear Information System (INIS)

    Olsson, O.; Black, J.H.; Gale, J.E.; Holmes, D.C.

    1989-05-01

    The Site Characterization and Validation (SCV) project is designed to assess how well we can characterize a volume of rock prior to using it as a repository. The programme of work focuses on the validation of the techniques used in site characterization. The SCV project contains 5 stages of work arranged in two 'cycles' of data-gathering, prediction, and validation. The first stage of work has included drilling of 6 boreholes (N2, N3, N4, W1, W2 and V3) and measurements of geology, fracture characteristics, stess, single borehole geophysical logging, radar, seismics and hydrogeology. The rock at the SCV site is granite with small lithological variations. Based essentially on radar and seismic results 5 'fracture zones' have been identified, named GA, GB, GC, GH and GI. They all extend acroos the entire SCV site. They aer basically in in two groups (GA, GB, GC and GH, GI). The first group are aligned N40 degree E with a dip of 35 degree to the south. The second group are aligned approximately N10 degree W dipping 60 degree E. From the stochastic analysis of the joint data it was possible to identify three main fracture orientation clusters. The orientation of two of these clusters agree roughly with orientation of the main features. Cluster B has roughly the same orientation as GH and GI, while features GA, GB and GC have an orientation similar to the more loosely defined cluster C. The orientation of the third cluster (A) is northwest with a dip to northeast. It is found that 94% of all measured hydraulic transmissivity is accounted for by 4% of the tested rock, not all of this 'concentrated' transmissivity is with the major features defined by geophysics. When the hydraulic connections across the site are examied they show that there are several welldefined zones which permit rapid transmission of hydraulic signals. These are essentially from the northeast to the southwest. (66 figs., 21 tabs., 33 refs.)

  8. Investigation into Methods for Predicting Connection Temperatures

    Directory of Open Access Journals (Sweden)

    K. Anderson

    2009-01-01

    Full Text Available The mechanical response of connections in fire is largely based on material strength degradation and the interactions between the various components of the connection. In order to predict connection performance in fire, temperature profiles must initially be established in order to evaluate the material strength degradation over time. This paper examines two current methods for predicting connection temperatures: The percentage method, where connection temperatures are calculated as a percentage of the adjacent beam lower-flange, mid-span temperatures; and the lumped capacitance method, based on the lumped mass of the connection. Results from the percentage method do not correlate well with experimental results, whereas the lumped capacitance method shows much better agreement with average connection temperatures. A 3D finite element heat transfer model was also created in Abaqus, and showed good correlation with experimental results. 

  9. Predicting Ligand Binding Sites on Protein Surfaces by 3-Dimensional Probability Density Distributions of Interacting Atoms

    Science.gov (United States)

    Jian, Jhih-Wei; Elumalai, Pavadai; Pitti, Thejkiran; Wu, Chih Yuan; Tsai, Keng-Chang; Chang, Jeng-Yih; Peng, Hung-Pin; Yang, An-Suei

    2016-01-01

    Predicting ligand binding sites (LBSs) on protein structures, which are obtained either from experimental or computational methods, is a useful first step in functional annotation or structure-based drug design for the protein structures. In this work, the structure-based machine learning algorithm ISMBLab-LIG was developed to predict LBSs on protein surfaces with input attributes derived from the three-dimensional probability density maps of interacting atoms, which were reconstructed on the query protein surfaces and were relatively insensitive to local conformational variations of the tentative ligand binding sites. The prediction accuracy of the ISMBLab-LIG predictors is comparable to that of the best LBS predictors benchmarked on several well-established testing datasets. More importantly, the ISMBLab-LIG algorithm has substantial tolerance to the prediction uncertainties of computationally derived protein structure models. As such, the method is particularly useful for predicting LBSs not only on experimental protein structures without known LBS templates in the database but also on computationally predicted model protein structures with structural uncertainties in the tentative ligand binding sites. PMID:27513851

  10. Soft Computing Methods for Disulfide Connectivity Prediction.

    Science.gov (United States)

    Márquez-Chamorro, Alfonso E; Aguilar-Ruiz, Jesús S

    2015-01-01

    The problem of protein structure prediction (PSP) is one of the main challenges in structural bioinformatics. To tackle this problem, PSP can be divided into several subproblems. One of these subproblems is the prediction of disulfide bonds. The disulfide connectivity prediction problem consists in identifying which nonadjacent cysteines would be cross-linked from all possible candidates. Determining the disulfide bond connectivity between the cysteines of a protein is desirable as a previous step of the 3D PSP, as the protein conformational search space is highly reduced. The most representative soft computing approaches for the disulfide bonds connectivity prediction problem of the last decade are summarized in this paper. Certain aspects, such as the different methodologies based on soft computing approaches (artificial neural network or support vector machine) or features of the algorithms, are used for the classification of these methods.

  11. Analysis and prediction of Multiple-Site Damage (MSD) fatigue crack growth

    Science.gov (United States)

    Dawicke, D. S.; Newman, J. C., Jr.

    1992-08-01

    A technique was developed to calculate the stress intensity factor for multiple interacting cracks. The analysis was verified through comparison with accepted methods of calculating stress intensity factors. The technique was incorporated into a fatigue crack growth prediction model and used to predict the fatigue crack growth life for multiple-site damage (MSD). The analysis was verified through comparison with experiments conducted on uniaxially loaded flat panels with multiple cracks. Configuration with nearly equal and unequal crack distribution were examined. The fatigue crack growth predictions agreed within 20 percent of the experimental lives for all crack configurations considered.

  12. Analysis and prediction of Multiple-Site Damage (MSD) fatigue crack growth

    Science.gov (United States)

    Dawicke, D. S.; Newman, J. C., Jr.

    1992-01-01

    A technique was developed to calculate the stress intensity factor for multiple interacting cracks. The analysis was verified through comparison with accepted methods of calculating stress intensity factors. The technique was incorporated into a fatigue crack growth prediction model and used to predict the fatigue crack growth life for multiple-site damage (MSD). The analysis was verified through comparison with experiments conducted on uniaxially loaded flat panels with multiple cracks. Configuration with nearly equal and unequal crack distribution were examined. The fatigue crack growth predictions agreed within 20 percent of the experimental lives for all crack configurations considered.

  13. New prediction methods for collaborative filtering

    Directory of Open Access Journals (Sweden)

    Hasan BULUT

    2016-05-01

    Full Text Available Companies, in particular e-commerce companies, aims to increase customer satisfaction, hence in turn increase their profits, using recommender systems. Recommender Systems are widely used nowadays and they provide strategic advantages to the companies that use them. These systems consist of different stages. In the first stage, the similarities between the active user and other users are computed using the user-product ratings matrix. Then, the neighbors of the active user are found from these similarities. In prediction calculation stage, the similarities computed at the first stage are used to generate the weight vector of the closer neighbors. Neighbors affect the prediction value by the corresponding value of the weight vector. In this study, we developed two new methods for the prediction calculation stage which is the last stage of collaborative filtering. The performance of these methods are measured with evaluation metrics used in the literature and compared with other studies in this field.

  14. Novel hyperspectral prediction method and apparatus

    Science.gov (United States)

    Kemeny, Gabor J.; Crothers, Natalie A.; Groth, Gard A.; Speck, Kathy A.; Marbach, Ralf

    2009-05-01

    Both the power and the challenge of hyperspectral technologies is the very large amount of data produced by spectral cameras. While off-line methodologies allow the collection of gigabytes of data, extended data analysis sessions are required to convert the data into useful information. In contrast, real-time monitoring, such as on-line process control, requires that compression of spectral data and analysis occur at a sustained full camera data rate. Efficient, high-speed practical methods for calibration and prediction are therefore sought to optimize the value of hyperspectral imaging. A novel method of matched filtering known as science based multivariate calibration (SBC) was developed for hyperspectral calibration. Classical (MLR) and inverse (PLS, PCR) methods are combined by spectroscopically measuring the spectral "signal" and by statistically estimating the spectral "noise." The accuracy of the inverse model is thus combined with the easy interpretability of the classical model. The SBC method is optimized for hyperspectral data in the Hyper-CalTM software used for the present work. The prediction algorithms can then be downloaded into a dedicated FPGA based High-Speed Prediction EngineTM module. Spectral pretreatments and calibration coefficients are stored on interchangeable SD memory cards, and predicted compositions are produced on a USB interface at real-time camera output rates. Applications include minerals, pharmaceuticals, food processing and remote sensing.

  15. Lattice gas methods for predicting intrinsic permeability of porous media

    Energy Technology Data Exchange (ETDEWEB)

    Santos, L.O.E.; Philippi, P.C. [Santa Catarina Univ., Florianopolis, SC (Brazil). Dept. de Engenharia Mecanica. Lab. de Propriedades Termofisicas e Meios Porosos)]. E-mail: emerich@lmpt.ufsc.br; philippi@lmpt.ufsc.br; Damiani, M.C. [Engineering Simulation and Scientific Software (ESSS), Florianopolis, SC (Brazil). Parque Tecnologico]. E-mail: damiani@lmpt.ufsc.br

    2000-07-01

    This paper presents a method for predicting intrinsic permeability of porous media based on Lattice Gas Cellular Automata methods. Two methods are presented. The first is based on a Boolean model (LGA). The second is Boltzmann method (LB) based on Boltzmann relaxation equation. LGA is a relatively recent method developed to perform hydrodynamic calculations. The method, in its simplest form, consists of a regular lattice populated with particles that hop from site to site in discrete time steps in a process, called propagation. After propagation, the particles in each site interact with each other in a process called collision, in which the number of particles and momentum are conserved. An exclusion principle is imposed in order to achieve better computational efficiency. In despite of its simplicity, this model evolves in agreement with Navier-Stokes equation for low Mach numbers. LB methods were recently developed for the numerical integration of the Navier-Stokes equation based on discrete Boltzmann transport equation. Derived from LGA, LB is a powerful alternative to the standard methods in computational fluid dynamics. In recent years, it has received much attention and has been used in several applications like simulations of flows through porous media, turbulent flows and multiphase flows. It is important to emphasize some aspects that make Lattice Gas Cellular Automata methods very attractive for simulating flows through porous media. In fact, boundary conditions in flows through complex geometry structures are very easy to describe in simulations using these methods. In LGA methods simulations are performed with integers needing less resident memory capability and boolean arithmetic reduces running time. The two methods are used to simulate flows through several Brazilian reservoir petroleum rocks leading to intrinsic permeability prediction. Simulation is compared with experimental results. (author)

  16. Features generated for computational splice-site prediction correspond to functional elements

    Directory of Open Access Journals (Sweden)

    Wilbur W John

    2007-10-01

    Full Text Available Abstract Background Accurate selection of splice sites during the splicing of precursors to messenger RNA requires both relatively well-characterized signals at the splice sites and auxiliary signals in the adjacent exons and introns. We previously described a feature generation algorithm (FGA that is capable of achieving high classification accuracy on human 3' splice sites. In this paper, we extend the splice-site prediction to 5' splice sites and explore the generated features for biologically meaningful splicing signals. Results We present examples from the observed features that correspond to known signals, both core signals (including the branch site and pyrimidine tract and auxiliary signals (including GGG triplets and exon splicing enhancers. We present evidence that features identified by FGA include splicing signals not found by other methods. Conclusion Our generated features capture known biological signals in the expected sequence interval flanking splice sites. The method can be easily applied to other species and to similar classification problems, such as tissue-specific regulatory elements, polyadenylation sites, promoters, etc.

  17. Artificial neural network intelligent method for prediction

    Science.gov (United States)

    Trifonov, Roumen; Yoshinov, Radoslav; Pavlova, Galya; Tsochev, Georgi

    2017-09-01

    Accounting and financial classification and prediction problems are high challenge and researchers use different methods to solve them. Methods and instruments for short time prediction of financial operations using artificial neural network are considered. The methods, used for prediction of financial data as well as the developed forecasting system with neural network are described in the paper. The architecture of a neural network used four different technical indicators, which are based on the raw data and the current day of the week is presented. The network developed is used for forecasting movement of stock prices one day ahead and consists of an input layer, one hidden layer and an output layer. The training method is algorithm with back propagation of the error. The main advantage of the developed system is self-determination of the optimal topology of neural network, due to which it becomes flexible and more precise The proposed system with neural network is universal and can be applied to various financial instruments using only basic technical indicators as input data.

  18. A web server for analysis, comparison and prediction of protein ligand binding sites.

    Science.gov (United States)

    Singh, Harinder; Srivastava, Hemant Kumar; Raghava, Gajendra P S

    2016-03-25

    One of the major challenges in the field of system biology is to understand the interaction between a wide range of proteins and ligands. In the past, methods have been developed for predicting binding sites in a protein for a limited number of ligands. In order to address this problem, we developed a web server named 'LPIcom' to facilitate users in understanding protein-ligand interaction. Analysis, comparison and prediction modules are available in the "LPIcom' server to predict protein-ligand interacting residues for 824 ligands. Each ligand must have at least 30 protein binding sites in PDB. Analysis module of the server can identify residues preferred in interaction and binding motif for a given ligand; for example residues glycine, lysine and arginine are preferred in ATP binding sites. Comparison module of the server allows comparing protein-binding sites of multiple ligands to understand the similarity between ligands based on their binding site. This module indicates that ATP, ADP and GTP ligands are in the same cluster and thus their binding sites or interacting residues exhibit a high level of similarity. Propensity-based prediction module has been developed for predicting ligand-interacting residues in a protein for more than 800 ligands. In addition, a number of web-based tools have been integrated to facilitate users in creating web logo and two-sample between ligand interacting and non-interacting residues. In summary, this manuscript presents a web-server for analysis of ligand interacting residue. This server is available for public use from URL http://crdd.osdd.net/raghava/lpicom .

  19. Prediction methods environmental-effect reporting

    International Nuclear Information System (INIS)

    Jonker, R.J.; Koester, H.W.

    1987-12-01

    This report provides a survey of prediction methods which can be applied to the calculation of emissions in cuclear-reactor accidents, in the framework of environment-effect reports (dutch m.e.r.) or risk analyses. Also emissions during normal operation are important for m.e.r.. These can be derived from measured emissions of power plants being in operation. Data concerning the latter are reported. The report consists of an introduction into reactor technology, among which a description of some reactor types, the corresponding fuel cycle and dismantling scenarios - a discussion of risk-analyses for nuclear power plants and the physical processes which can play a role during accidents - a discussion of prediction methods to be employed and the expected developments in this area - some background information. (aughor). 145 refs.; 21 figs.; 20 tabs

  20. A comparison of methods for cascade prediction

    OpenAIRE

    Guo, Ruocheng; Shakarian, Paulo

    2016-01-01

    Information cascades exist in a wide variety of platforms on Internet. A very important real-world problem is to identify which information cascades can go viral. A system addressing this problem can be used in a variety of applications including public health, marketing and counter-terrorism. As a cascade can be considered as compound of the social network and the time series. However, in related literature where methods for solving the cascade prediction problem were proposed, the experimen...

  1. Methods of Identification and Evaluation of Brownfield Sites

    Directory of Open Access Journals (Sweden)

    Safet Kurtović

    2014-04-01

    Full Text Available The basic objective of this paper was to determine the importance and potential restoration of brownfield sites in terms of economic prosperity of a particular region or country. In addition, in a theoretical sense, this paper presents the methods used in the identification of brownfield sites such as Smart Growth Network model and Thomas GIS model, and methods for evaluation of brownfield sites or the indexing method, cost-benefit and multivariate analysis.

  2. Detection and characterization of 3D-signature phosphorylation site motifs and their contribution towards improved phosphorylation site prediction in proteins

    Directory of Open Access Journals (Sweden)

    Selbig Joachim

    2009-04-01

    Full Text Available Abstract Background Phosphorylation of proteins plays a crucial role in the regulation and activation of metabolic and signaling pathways and constitutes an important target for pharmaceutical intervention. Central to the phosphorylation process is the recognition of specific target sites by protein kinases followed by the covalent attachment of phosphate groups to the amino acids serine, threonine, or tyrosine. The experimental identification as well as computational prediction of phosphorylation sites (P-sites has proved to be a challenging problem. Computational methods have focused primarily on extracting predictive features from the local, one-dimensional sequence information surrounding phosphorylation sites. Results We characterized the spatial context of phosphorylation sites and assessed its usability for improved phosphorylation site predictions. We identified 750 non-redundant, experimentally verified sites with three-dimensional (3D structural information available in the protein data bank (PDB and grouped them according to their respective kinase family. We studied the spatial distribution of amino acids around phosphorserines, phosphothreonines, and phosphotyrosines to extract signature 3D-profiles. Characteristic spatial distributions of amino acid residue types around phosphorylation sites were indeed discernable, especially when kinase-family-specific target sites were analyzed. To test the added value of using spatial information for the computational prediction of phosphorylation sites, Support Vector Machines were applied using both sequence as well as structural information. When compared to sequence-only based prediction methods, a small but consistent performance improvement was obtained when the prediction was informed by 3D-context information. Conclusion While local one-dimensional amino acid sequence information was observed to harbor most of the discriminatory power, spatial context information was identified as

  3. Prediction of flow and drawdown for the site characterization and validation site in the Stripa Mine

    International Nuclear Information System (INIS)

    Long, J.C.S.; Mauldon, A.D.; Nelson, K.; Martel, S.; Fuller, P.; and Karasaki, K.

    1992-01-01

    Geophysical and hydrologic data from a location in the Stripa Mine in Sweden, called the Site Characterization and Validation (SCV) block, has been used to create a series of models for flow through the fracture network. The models can be characterized as ''equivalent discontinuum'' models. Equivalent discontinuum models are derived starting from a specified lattice or 6 ''template''. An inverse analysis called ''Simulated Annealing'' is used to make a random search through the elements of the lattice to find a configuration that can reproduce the measured responses. Evidence at Stripa points to hydrology which is dominated by fracture zones. These have been identified and located through extensive characterization efforts. Lattice templates were arranged to lie on the fracture zones identified by Black and Olsson. The fundamental goal of this project was to build a fracture flow model based an initial data set, and use this model to make predictions of the flow behavior during a new test. Then given data from the new test, predict a second test, etc. The first data set was an interference test called C1-2. Both a two-dimensional and a three-dimensional model were annealed to the C1-2 data and use this model to predict the behavior of the Simulated Drift Experiment (SDE). The SDE measured the flow into, and drawdown due to reducing the pressure in a group of 6 parallel boreholes. Then both the C1-2 and SDE data were used to predict the flow into and drawdown due to an excavation, the Validation Drift (VD), made through the boreholes. Finally, all the data was used to predict the hydrologic response to opening another hole, T1

  4. A review of influenza detection and prediction through social networking sites.

    Science.gov (United States)

    Alessa, Ali; Faezipour, Miad

    2018-02-01

    Early prediction of seasonal epidemics such as influenza may reduce their impact in daily lives. Nowadays, the web can be used for surveillance of diseases. Search engines and social networking sites can be used to track trends of different diseases seven to ten days faster than government agencies such as Center of Disease Control and Prevention (CDC). CDC uses the Illness-Like Influenza Surveillance Network (ILINet), which is a program used to monitor Influenza-Like Illness (ILI) sent by thousands of health care providers in order to detect influenza outbreaks. It is a reliable tool, however, it is slow and expensive. For that reason, many studies aim to develop methods that do real time analysis to track ILI using social networking sites. Social media data such as Twitter can be used to predict the spread of flu in the population and can help in getting early warnings. Today, social networking sites (SNS) are used widely by many people to share thoughts and even health status. Therefore, SNS provides an efficient resource for disease surveillance and a good way to communicate to prevent disease outbreaks. The goal of this study is to review existing alternative solutions that track flu outbreak in real time using social networking sites and web blogs. Many studies have shown that social networking sites can be used to conduct real time analysis for better predictions.

  5. Hybrid methods for airframe noise numerical prediction

    Energy Technology Data Exchange (ETDEWEB)

    Terracol, M.; Manoha, E.; Herrero, C.; Labourasse, E.; Redonnet, S. [ONERA, Department of CFD and Aeroacoustics, BP 72, Chatillon (France); Sagaut, P. [Laboratoire de Modelisation en Mecanique - UPMC/CNRS, Paris (France)

    2005-07-01

    This paper describes some significant steps made towards the numerical simulation of the noise radiated by the high-lift devices of a plane. Since the full numerical simulation of such configuration is still out of reach for present supercomputers, some hybrid strategies have been developed to reduce the overall cost of such simulations. The proposed strategy relies on the coupling of an unsteady nearfield CFD with an acoustic propagation solver based on the resolution of the Euler equations for midfield propagation in an inhomogeneous field, and the use of an integral solver for farfield acoustic predictions. In the first part of this paper, this CFD/CAA coupling strategy is presented. In particular, the numerical method used in the propagation solver is detailed, and two applications of this coupling method to the numerical prediction of the aerodynamic noise of an airfoil are presented. Then, a hybrid RANS/LES method is proposed in order to perform some unsteady simulations of complex noise sources. This method allows for significant reduction of the cost of such a simulation by considerably reducing the extent of the LES zone. This method is described and some results of the numerical simulation of the three-dimensional unsteady flow in the slat cove of a high-lift profile are presented. While these results remain very difficult to validate with experiments on similar configurations, they represent up to now the first 3D computations of this kind of flow. (orig.)

  6. WSDM : A user-centred design method for web sites

    NARCIS (Netherlands)

    de Troyer, O.M.F.; Leune, C.J.

    1998-01-01

    WSDM is a user-centered method for the design of kiosk Web Sites. By explicitly starting from the requirements of the users or visitors, WSDM solves Web site problems that are primarily caused by that fact that a site has no underlying design at all, or that the design is mostly data-driven.

  7. Mechatronics technology in predictive maintenance method

    Science.gov (United States)

    Majid, Nurul Afiqah A.; Muthalif, Asan G. A.

    2017-11-01

    This paper presents recent mechatronics technology that can help to implement predictive maintenance by combining intelligent and predictive maintenance instrument. Vibration Fault Simulation System (VFSS) is an example of mechatronics system. The focus of this study is the prediction on the use of critical machines to detect vibration. Vibration measurement is often used as the key indicator of the state of the machine. This paper shows the choice of the appropriate strategy in the vibration of diagnostic process of the mechanical system, especially rotating machines, in recognition of the failure during the working process. In this paper, the vibration signature analysis is implemented to detect faults in rotary machining that includes imbalance, mechanical looseness, bent shaft, misalignment, missing blade bearing fault, balancing mass and critical speed. In order to perform vibration signature analysis for rotating machinery faults, studies have been made on how mechatronics technology is used as predictive maintenance methods. Vibration Faults Simulation Rig (VFSR) is designed to simulate and understand faults signatures. These techniques are based on the processing of vibrational data in frequency-domain. The LabVIEW-based spectrum analyzer software is developed to acquire and extract frequency contents of faults signals. This system is successfully tested based on the unique vibration fault signatures that always occur in a rotating machinery.

  8. Methods for estimating on-site ambient air concentrations at disposal sites

    International Nuclear Information System (INIS)

    Hwang, S.T.

    1987-01-01

    Currently, Gaussian type dispersion modeling and point source approximation are combined to estimate the ambient air concentrations of pollutants dispersed downwind of an areawide emission source, using the approach of virtual point source approximation. This Gaussian dispersion modeling becomes less accurate as the receptor comes closer to the source, and becomes inapplicable for the estimation of on-site ambient air concentrations at disposal sites. Partial differential equations are solved with appropriate boundary conditions for use in estimating the on-site concentrations in the ambient air impacted by emissions from an area source such as land disposal sites. Two variations of solution techniques are presented, and their predictions are compared

  9. Prediction of site specific ground motion for large earthquake

    International Nuclear Information System (INIS)

    Kamae, Katsuhiro; Irikura, Kojiro; Fukuchi, Yasunaga.

    1990-01-01

    In this paper, we apply the semi-empirical synthesis method by IRIKURA (1983, 1986) to the estimation of site specific ground motion using accelerograms observed at Kumatori in Osaka prefecture. Target earthquakes used here are a comparatively distant earthquake (Δ=95 km, M=5.6) caused by the YAMASAKI fault and a near earthquake (Δ=27 km, M=5.6). The results obtained are as follows. 1) The accelerograms from the distant earthquake (M=5.6) are synthesized using the aftershock records (M=4.3) for 1983 YAMASAKI fault earthquake whose source parameters have been obtained by other authors from the hypocentral distribution of the aftershocks. The resultant synthetic motions show a good agreement with the observed ones. 2) The synthesis for a near earthquake (M=5.6, we call this target earthquake) are made using a small earthquake which occurred in the neighborhood of the target earthquake. Here, we apply two methods for giving the parameters for synthesis. One method is to use the parameters of YAMASAKI fault earthquake which has the same magnitude as the target earthquake, and the other is to use the parameters obtained from several existing empirical formulas. The resultant synthetic motion with the former parameters shows a good agreement with the observed one, but that with the latter does not. 3) We estimate the source parameters from the source spectra of several earthquakes which have been observed in this site. Consequently we find that the small earthquakes (M<4) as Green's functions should be carefully used because the stress drops are not constant. 4) We propose that we should designate not only the magnitudes but also seismic moments of the target earthquake and the small earthquake. (J.P.N.)

  10. Prediction Methods for Blood Glucose Concentration

    DEFF Research Database (Denmark)

    -day workshop on the design, use and evaluation of prediction methods for blood glucose concentration was held at the Johannes Kepler University Linz, Austria. One intention of the workshop was to bring together experts working in various fields on the same topic, in order to shed light from different angles...... discussions which allowed to receive direct feedback from the point of view of different disciplines. This book is based on the contributions of that workshop and is intended to convey an overview of the different aspects involved in the prediction. The individual chapters are based on the presentations given...... in the process of writing this book: All authors for their individual contributions, all reviewers of the book chapters, Daniela Hummer for the entire organization of the workshop, Boris Tasevski for helping with the typesetting, Florian Reiterer for his help editing the book, as well as Oliver Jackson and Karin...

  11. Position-specific prediction of methylation sites from sequence conservation based on information theory.

    Science.gov (United States)

    Shi, Yinan; Guo, Yanzhi; Hu, Yayun; Li, Menglong

    2015-07-23

    Protein methylation plays vital roles in many biological processes and has been implicated in various human diseases. To fully understand the mechanisms underlying methylation for use in drug design and work in methylation-related diseases, an initial but crucial step is to identify methylation sites. The use of high-throughput bioinformatics methods has become imperative to predict methylation sites. In this study, we developed a novel method that is based only on sequence conservation to predict protein methylation sites. Conservation difference profiles between methylated and non-methylated peptides were constructed by the information entropy (IE) in a wider neighbor interval around the methylation sites that fully incorporated all of the environmental information. Then, the distinctive neighbor residues were identified by the importance scores of information gain (IG). The most representative model was constructed by support vector machine (SVM) for Arginine and Lysine methylation, respectively. This model yielded a promising result on both the benchmark dataset and independent test set. The model was used to screen the entire human proteome, and many unknown substrates were identified. These results indicate that our method can serve as a useful supplement to elucidate the mechanism of protein methylation and facilitate hypothesis-driven experimental design and validation.

  12. Archaeology Through Computational Linguistics: Inscription Statistics Predict Excavation Sites of Indus Valley Artifacts.

    Science.gov (United States)

    Recchia, Gabriel L; Louwerse, Max M

    2016-11-01

    Computational techniques comparing co-occurrences of city names in texts allow the relative longitudes and latitudes of cities to be estimated algorithmically. However, these techniques have not been applied to estimate the provenance of artifacts with unknown origins. Here, we estimate the geographic origin of artifacts from the Indus Valley Civilization, applying methods commonly used in cognitive science to the Indus script. We show that these methods can accurately predict the relative locations of archeological sites on the basis of artifacts of known provenance, and we further apply these techniques to determine the most probable excavation sites of four sealings of unknown provenance. These findings suggest that inscription statistics reflect historical interactions among locations in the Indus Valley region, and they illustrate how computational methods can help localize inscribed archeological artifacts of unknown origin. The success of this method offers opportunities for the cognitive sciences in general and for computational anthropology specifically. Copyright © 2015 Cognitive Science Society, Inc.

  13. Methods of Usability Testing in Libraries Web Sites

    Directory of Open Access Journals (Sweden)

    Eman Fawzy

    2006-03-01

    Full Text Available A Study about libraries' web sites evaluation, that is the Usability, the study talking about methods of usability testing and define it, and its important in web sites evaluation, then details the methods of usability: questionnaire, core groups, testing experimental model, cards arrangement, and composed evaluation.

  14. On - Site Assessment Methods For Environmental Radioactivity

    International Nuclear Information System (INIS)

    Petrinec, B.; Babic, D.; Bituh, T.

    2015-01-01

    A method for the rapid determination of radioactivity in cases of release into the environment as well as in cases of nuclear/radiological accidents is described. These measurements would enable a direct risk assessment for humans and biota, without any sampling and at a considerably larger number of locations than in previous studies. Thus obtained, the substantially expanded dataset is expected to shed more light on the properties of environmental radioactivity both in the region studied and in other similar areas. Field measurements will be performed and samples of soil and biota will be collected in order to compare field results with laboratory measurements. Once the method has been validated, previously unexplored locations will be included in the study. Our measurements at numerous locations will also provide control values for comparison in cases of any unplanned or accidental radiological event. An assessment of the possible effects of radionuclide concentrations on the human food chain and biota will be performed within the appropriate models used worldwide exactly for this purpose. In this way, the project should contribute to regional, European, and global efforts towards understanding the radiological impact on ecosystems. Field measurements will also address certain issues in the environmental metrology of radioactive substances, e.g., simultaneous determination of activity concentrations and related dose rates. This will serve as a tool for rapid risk assessment in emergency situations. (author).

  15. High DNA melting temperature predicts transcription start site location in human and mouse.

    LENUS (Irish Health Repository)

    Dineen, David G

    2009-12-01

    The accurate computational prediction of transcription start sites (TSS) in vertebrate genomes is a difficult problem. The physicochemical properties of DNA can be computed in various ways and a many combinations of DNA features have been tested in the past for use as predictors of transcription. We looked in detail at melting temperature, which measures the temperature, at which two strands of DNA separate, considering the cooperative nature of this process. We find that peaks in melting temperature correspond closely to experimentally determined transcription start sites in human and mouse chromosomes. Using melting temperature alone, and with simple thresholding, we can predict TSS with accuracy that is competitive with the most accurate state-of-the-art TSS prediction methods. Accuracy is measured using both experimentally and manually determined TSS. The method works especially well with CpG island containing promoters, but also works when CpG islands are absent. This result is clear evidence of the important role of the physical properties of DNA in the process of transcription. It also points to the importance for TSS prediction methods to include melting temperature as prior information.

  16. Seminal quality prediction using data mining methods.

    Science.gov (United States)

    Sahoo, Anoop J; Kumar, Yugal

    2014-01-01

    Now-a-days, some new classes of diseases have come into existences which are known as lifestyle diseases. The main reasons behind these diseases are changes in the lifestyle of people such as alcohol drinking, smoking, food habits etc. After going through the various lifestyle diseases, it has been found that the fertility rates (sperm quantity) in men has considerably been decreasing in last two decades. Lifestyle factors as well as environmental factors are mainly responsible for the change in the semen quality. The objective of this paper is to identify the lifestyle and environmental features that affects the seminal quality and also fertility rate in man using data mining methods. The five artificial intelligence techniques such as Multilayer perceptron (MLP), Decision Tree (DT), Navie Bayes (Kernel), Support vector machine+Particle swarm optimization (SVM+PSO) and Support vector machine (SVM) have been applied on fertility dataset to evaluate the seminal quality and also to predict the person is either normal or having altered fertility rate. While the eight feature selection techniques such as support vector machine (SVM), neural network (NN), evolutionary logistic regression (LR), support vector machine plus particle swarm optimization (SVM+PSO), principle component analysis (PCA), chi-square test, correlation and T-test methods have been used to identify more relevant features which affect the seminal quality. These techniques are applied on fertility dataset which contains 100 instances with nine attribute with two classes. The experimental result shows that SVM+PSO provides higher accuracy and area under curve (AUC) rate (94% & 0.932) among multi-layer perceptron (MLP) (92% & 0.728), Support Vector Machines (91% & 0.758), Navie Bayes (Kernel) (89% & 0.850) and Decision Tree (89% & 0.735) for some of the seminal parameters. This paper also focuses on the feature selection process i.e. how to select the features which are more important for prediction of

  17. Geophysical Exploration. New site exploration method

    Energy Technology Data Exchange (ETDEWEB)

    Imai, Tsuneo; Otomo, Hideo; Sakayama, Toshihiko

    1988-07-25

    Geophysical exploration is used for geologic survey to serve purposes in civil engineering. New methods are being developed inside and outside Japan and are used to serve various purposes. This paper discusses recently developed techniques based on the measurement of seismic waves and electric potential. It also explains seismic tomography, radar tomography, and resistivity tomography which are included in the category of geotomography. At present, effort is being made to apply geophysical exploration technology to problems which were considered to be unsuitable for conventional exploration techniques. When such effort proceeds successfully, it is necessary to develop technology for presenting results quickly and exploration equipment which can work in various conditions. (10 figs, 15 refs)

  18. SPEER-SERVER: a web server for prediction of protein specificity determining sites.

    Science.gov (United States)

    Chakraborty, Abhijit; Mandloi, Sapan; Lanczycki, Christopher J; Panchenko, Anna R; Chakrabarti, Saikat

    2012-07-01

    Sites that show specific conservation patterns within subsets of proteins in a protein family are likely to be involved in the development of functional specificity. These sites, generally termed specificity determining sites (SDS), might play a crucial role in binding to a specific substrate or proteins. Identification of SDS through experimental techniques is a slow, difficult and tedious job. Hence, it is very important to develop efficient computational methods that can more expediently identify SDS. Herein, we present Specificity prediction using amino acids' Properties, Entropy and Evolution Rate (SPEER)-SERVER, a web server that predicts SDS by analyzing quantitative measures of the conservation patterns of protein sites based on their physico-chemical properties and the heterogeneity of evolutionary changes between and within the protein subfamilies. This web server provides an improved representation of results, adds useful input and output options and integrates a wide range of analysis and data visualization tools when compared with the original standalone version of the SPEER algorithm. Extensive benchmarking finds that SPEER-SERVER exhibits sensitivity and precision performance that, on average, meets or exceeds that of other currently available methods. SPEER-SERVER is available at http://www.hpppi.iicb.res.in/ss/.

  19. Development of subsurface characterization method for decommissioning site remediation

    Energy Technology Data Exchange (ETDEWEB)

    Hong, Sang Bum; Nam, Jong Soo; Choi, Yong Suk; Seo, Bum Kyoung; Moon, Jei Kwon; Choi, Jong Won [KAERI, Daejeon (Korea, Republic of)

    2016-05-15

    In situ measurement of peak to valley method based on the ratio of counting rate between the full energy peak and Compton region was applied to identify the depth distribution of 137Cs. The In situ measurement and sampling results were applied to evaluate a residual radioactivity before and after remediation in decommissioning KRR site. Spatial analysis based on the Geostatistics method provides a reliable estimating the volume of contaminated soil with a graphical analysis, which was applied to the site characterization in the decommissioning KRR site. The in situ measurement and spatial analysis results for characterization of subsurface contamination are presented. The objective of a remedial action is to reduce risks to human health to acceptable levels by removing the source of contamination. Site characterization of the subsurface contamination is an important factor for planning and implementation of site remediation. Radiological survey and evaluation technology are required to ensure the reliability of the results, and the process must be easily applied during field measurements. In situ gamma-ray spectrometry is a powerful method for site characterization that can be used to identify the depth distribution and quantify radionuclides directly at the measurement site. The in situ measurement and Geostatistics method was applied to the site characterization for remediation and final status survey in decommissioning KRR site.

  20. A risk assessment method for multi-site damage

    Science.gov (United States)

    Millwater, Harry Russell, Jr.

    This research focused on developing probabilistic methods suitable for computing small probabilities of failure, e.g., 10sp{-6}, of structures subject to multi-site damage (MSD). MSD is defined as the simultaneous development of fatigue cracks at multiple sites in the same structural element such that the fatigue cracks may coalesce to form one large crack. MSD is modeled as an array of collinear cracks with random initial crack lengths with the centers of the initial cracks spaced uniformly apart. The data used was chosen to be representative of aluminum structures. The structure is considered failed whenever any two adjacent cracks link up. A fatigue computer model is developed that can accurately and efficiently grow a collinear array of arbitrary length cracks from initial size until failure. An algorithm is developed to compute the stress intensity factors of all cracks considering all interaction effects. The probability of failure of two to 100 cracks is studied. Lower bounds on the probability of failure are developed based upon the probability of the largest crack exceeding a critical crack size. The critical crack size is based on the initial crack size that will grow across the ligament when the neighboring crack has zero length. The probability is evaluated using extreme value theory. An upper bound is based on the probability of the maximum sum of initial cracks being greater than a critical crack size. A weakest link sampling approach is developed that can accurately and efficiently compute small probabilities of failure. This methodology is based on predicting the weakest link, i.e., the two cracks to link up first, for a realization of initial crack sizes, and computing the cycles-to-failure using these two cracks. Criteria to determine the weakest link are discussed. Probability results using the weakest link sampling method are compared to Monte Carlo-based benchmark results. The results indicate that very small probabilities can be computed

  1. AutoSite: an automated approach for pseudo-ligands prediction—from ligand-binding sites identification to predicting key ligand atoms

    Science.gov (United States)

    Ravindranath, Pradeep Anand; Sanner, Michel F.

    2016-01-01

    Motivation: The identification of ligand-binding sites from a protein structure facilitates computational drug design and optimization, and protein function assignment. We introduce AutoSite: an efficient software tool for identifying ligand-binding sites and predicting pseudo ligand corresponding to each binding site identified. Binding sites are reported as clusters of 3D points called fills in which every point is labelled as hydrophobic or as hydrogen bond donor or acceptor. From these fills AutoSite derives feature points: a set of putative positions of hydrophobic-, and hydrogen-bond forming ligand atoms. Results: We show that AutoSite identifies ligand-binding sites with higher accuracy than other leading methods, and produces fills that better matches the ligand shape and properties, than the fills obtained with a software program with similar capabilities, AutoLigand. In addition, we demonstrate that for the Astex Diverse Set, the feature points identify 79% of hydrophobic ligand atoms, and 81% and 62% of the hydrogen acceptor and donor hydrogen ligand atoms interacting with the receptor, and predict 81.2% of water molecules mediating interactions between ligand and receptor. Finally, we illustrate potential uses of the predicted feature points in the context of lead optimization in drug discovery projects. Availability and Implementation: http://adfr.scripps.edu/AutoDockFR/autosite.html Contact: sanner@scripps.edu Supplementary information: Supplementary data are available at Bioinformatics online. PMID:27354702

  2. Method for assigning sites to projected generic nuclear power plants

    Energy Technology Data Exchange (ETDEWEB)

    Holter, G.M.; Purcell, W.L.; Shutz, M.E.; Young, J.R.

    1986-07-01

    Pacific Northwest Laboratory developed a method for forecasting potential locations and startup sequences of nuclear power plants that will be required in the future but have not yet been specifically identified by electric utilities. Use of the method results in numerical ratings for potential nuclear power plant sites located in each of the 10 federal energy regions. The rating for each potential site is obtained from numerical factors assigned to each of 5 primary siting characteristics: (1) cooling water availability, (2) site land area, (3) power transmission land area, (4) proximity to metropolitan areas, and (5) utility plans for the site. The sequence of plant startups in each federal energy region is obtained by use of the numerical ratings and the forecasts of generic nuclear power plant startups obtained from the EIA Middle Case electricity forecast. Sites are assigned to generic plants in chronological order according to startup date.

  3. Method for assigning sites to projected generic nuclear power plants

    International Nuclear Information System (INIS)

    Holter, G.M.; Purcell, W.L.; Shutz, M.E.; Young, J.R.

    1986-07-01

    Pacific Northwest Laboratory developed a method for forecasting potential locations and startup sequences of nuclear power plants that will be required in the future but have not yet been specifically identified by electric utilities. Use of the method results in numerical ratings for potential nuclear power plant sites located in each of the 10 federal energy regions. The rating for each potential site is obtained from numerical factors assigned to each of 5 primary siting characteristics: (1) cooling water availability, (2) site land area, (3) power transmission land area, (4) proximity to metropolitan areas, and (5) utility plans for the site. The sequence of plant startups in each federal energy region is obtained by use of the numerical ratings and the forecasts of generic nuclear power plant startups obtained from the EIA Middle Case electricity forecast. Sites are assigned to generic plants in chronological order according to startup date

  4. Site-specific data confirm arsenic exposure predicted by the U.S. Environmental Protection Agency.

    OpenAIRE

    Walker, S; Griffin, S

    1998-01-01

    The EPA uses an exposure assessment model to estimate daily intake to chemicals of potential concern. At the Anaconda Superfund site in Montana, the EPA exposure assessment model was used to predict total and speciated urinary arsenic concentrations. Predicted concentrations were then compared to concentrations measured in children living near the site. When site-specific information on concentrations of arsenic in soil, interior dust, and diet, site-specific ingestion rates, and arsenic abso...

  5. NBA-Palm: prediction of palmitoylation site implemented in Naïve Bayes algorithm

    Directory of Open Access Journals (Sweden)

    Jin Changjiang

    2006-10-01

    Full Text Available Abstract Background Protein palmitoylation, an essential and reversible post-translational modification (PTM, has been implicated in cellular dynamics and plasticity. Although numerous experimental studies have been performed to explore the molecular mechanisms underlying palmitoylation processes, the intrinsic feature of substrate specificity has remained elusive. Thus, computational approaches for palmitoylation prediction are much desirable for further experimental design. Results In this work, we present NBA-Palm, a novel computational method based on Naïve Bayes algorithm for prediction of palmitoylation site. The training data is curated from scientific literature (PubMed and includes 245 palmitoylated sites from 105 distinct proteins after redundancy elimination. The proper window length for a potential palmitoylated peptide is optimized as six. To evaluate the prediction performance of NBA-Palm, 3-fold cross-validation, 8-fold cross-validation and Jack-Knife validation have been carried out. Prediction accuracies reach 85.79% for 3-fold cross-validation, 86.72% for 8-fold cross-validation and 86.74% for Jack-Knife validation. Two more algorithms, RBF network and support vector machine (SVM, also have been employed and compared with NBA-Palm. Conclusion Taken together, our analyses demonstrate that NBA-Palm is a useful computational program that provides insights for further experimentation. The accuracy of NBA-Palm is comparable with our previously described tool CSS-Palm. The NBA-Palm is freely accessible from: http://www.bioinfo.tsinghua.edu.cn/NBA-Palm.

  6. NBA-Palm: prediction of palmitoylation site implemented in Naïve Bayes algorithm.

    Science.gov (United States)

    Xue, Yu; Chen, Hu; Jin, Changjiang; Sun, Zhirong; Yao, Xuebiao

    2006-10-17

    Protein palmitoylation, an essential and reversible post-translational modification (PTM), has been implicated in cellular dynamics and plasticity. Although numerous experimental studies have been performed to explore the molecular mechanisms underlying palmitoylation processes, the intrinsic feature of substrate specificity has remained elusive. Thus, computational approaches for palmitoylation prediction are much desirable for further experimental design. In this work, we present NBA-Palm, a novel computational method based on Naïve Bayes algorithm for prediction of palmitoylation site. The training data is curated from scientific literature (PubMed) and includes 245 palmitoylated sites from 105 distinct proteins after redundancy elimination. The proper window length for a potential palmitoylated peptide is optimized as six. To evaluate the prediction performance of NBA-Palm, 3-fold cross-validation, 8-fold cross-validation and Jack-Knife validation have been carried out. Prediction accuracies reach 85.79% for 3-fold cross-validation, 86.72% for 8-fold cross-validation and 86.74% for Jack-Knife validation. Two more algorithms, RBF network and support vector machine (SVM), also have been employed and compared with NBA-Palm. Taken together, our analyses demonstrate that NBA-Palm is a useful computational program that provides insights for further experimentation. The accuracy of NBA-Palm is comparable with our previously described tool CSS-Palm. The NBA-Palm is freely accessible from: http://www.bioinfo.tsinghua.edu.cn/NBA-Palm.

  7. Establishing a predictive maintenance program at the Hanford Site

    International Nuclear Information System (INIS)

    Winslow, R.W.

    1994-05-01

    This document contains information about a new Predictive Maintenance Program being developed and implemented at the Hanford Reservation. Details of the document include: background on persons developing the program, history of predictive maintenance, implementation of new program, engineering task analysis, network development and new software, issues to be resolved, benefits expected, and appendix gives information about the symposium from which this paper is based

  8. Prediction of P53 mutants (multiple sites transcriptional activity based on structural (2D&3D properties.

    Directory of Open Access Journals (Sweden)

    R Geetha Ramani

    Full Text Available Prediction of secondary site mutations that reinstate mutated p53 to normalcy has been the focus of intense research in the recent past owing to the fact that p53 mutants have been implicated in more than half of all human cancers and restoration of p53 causes tumor regression. However laboratory investigations are more often laborious and resource intensive but computational techniques could well surmount these drawbacks. In view of this, we formulated a novel approach utilizing computational techniques to predict the transcriptional activity of multiple site (one-site to five-site p53 mutants. The optimal MCC obtained by the proposed approach on prediction of one-site, two-site, three-site, four-site and five-site mutants were 0.775,0.341,0.784,0.916 and 0.655 respectively, the highest reported thus far in literature. We have also demonstrated that 2D and 3D features generate higher prediction accuracy of p53 activity and our findings revealed the optimal results for prediction of p53 status, reported till date. We believe detection of the secondary site mutations that suppress tumor growth may facilitate better understanding of the relationship between p53 structure and function and further knowledge on the molecular mechanisms and biological activity of p53, a targeted source for cancer therapy. We expect that our prediction methods and reported results may provide useful insights on p53 functional mechanisms and generate more avenues for utilizing computational techniques in biological data analysis.

  9. Utilizing knowledge base of amino acids structural neighborhoods to predict protein-protein interaction sites.

    Science.gov (United States)

    Jelínek, Jan; Škoda, Petr; Hoksza, David

    2017-12-06

    Protein-protein interactions (PPI) play a key role in an investigation of various biochemical processes, and their identification is thus of great importance. Although computational prediction of which amino acids take part in a PPI has been an active field of research for some time, the quality of in-silico methods is still far from perfect. We have developed a novel prediction method called INSPiRE which benefits from a knowledge base built from data available in Protein Data Bank. All proteins involved in PPIs were converted into labeled graphs with nodes corresponding to amino acids and edges to pairs of neighboring amino acids. A structural neighborhood of each node was then encoded into a bit string and stored in the knowledge base. When predicting PPIs, INSPiRE labels amino acids of unknown proteins as interface or non-interface based on how often their structural neighborhood appears as interface or non-interface in the knowledge base. We evaluated INSPiRE's behavior with respect to different types and sizes of the structural neighborhood. Furthermore, we examined the suitability of several different features for labeling the nodes. Our evaluations showed that INSPiRE clearly outperforms existing methods with respect to Matthews correlation coefficient. In this paper we introduce a new knowledge-based method for identification of protein-protein interaction sites called INSPiRE. Its knowledge base utilizes structural patterns of known interaction sites in the Protein Data Bank which are then used for PPI prediction. Extensive experiments on several well-established datasets show that INSPiRE significantly surpasses existing PPI approaches.

  10. Predicting on-site environmental impacts of municipal engineering works

    OpenAIRE

    Gangolells Solanellas, Marta; Casals Casanova, Miquel; Forcada Matheu, Núria; Macarulla Martí, Marcel

    2014-01-01

    The research findings fill a gap in the body of knowledge by presenting an effective way to evaluate the significance of on-site environmental impacts of municipal engineering works prior to the construction stage. First, 42 on-site environmental impacts of municipal engineering works were identified by means of a process-oriented approach. Then, 46 indicators and their corresponding significance limits were determined on the basis of a statistical analysis of 25 new-build and remodelling mun...

  11. FACTORS PREDICTING CONSUMERS' ASSESSMENT OF ADVERTISEMENTS ON SOCIAL NETWORKING SITES

    OpenAIRE

    Hossam Deraz; Gabriel Baffour Awuah; Desalegn Abraha Gebrekidan

    2015-01-01

    Marketers act on social networking sites (SNSs) in order to be more efficient in merchandising their products and/or services. Even so, the scope of the published studies regarding the assessment of advertisements on social networking sites (SNAs) is limited. Consequently, the present study aimed to consider credibility and interactivity, in addition to information, entertainment and irritation values, as main factors for consumers’ assessment of SNAs, as perceived by SNSs’ users. An analysis...

  12. Prediction of citrullination sites by incorporating k-spaced amino acid pairs into Chou's general pseudo amino acid composition.

    Science.gov (United States)

    Ju, Zhe; Wang, Shi-Yun

    2018-04-22

    As one of the most important and common protein post-translational modifications, citrullination plays a key role in regulating various biological processes and is associated with several human diseases. The accurate identification of citrullination sites is crucial for elucidating the underlying molecular mechanisms of citrullination and designing drugs for related human diseases. In this study, a novel bioinformatics tool named CKSAAP_CitrSite is developed for the prediction of citrullination sites. With the assistance of support vector machine algorithm, the highlight of CKSAAP_CitrSite is to adopt the composition of k-spaced amino acid pairs surrounding a query site as input. As illustrated by 10-fold cross-validation, CKSAAP_CitrSite achieves a satisfactory performance with a Sensitivity of 77.59%, a Specificity of 95.26%, an Accuracy of 89.37% and a Matthew's correlation coefficient of 0.7566, which is much better than those of the existing prediction method. Feature analysis shows that the N-terminal space containing pairs may play an important role in the prediction of citrullination sites, and the arginines close to N-terminus tend to be citrullinated. The conclusions derived from this study could offer useful information for elucidating the molecular mechanisms of citrullination and related experimental validations. A user-friendly web-server for CKSAAP_CitrSite is available at 123.206.31.171/CKSAAP_CitrSite/. Copyright © 2017. Published by Elsevier B.V.

  13. New methods for fall risk prediction.

    Science.gov (United States)

    Ejupi, Andreas; Lord, Stephen R; Delbaere, Kim

    2014-09-01

    Accidental falls are the leading cause of injury-related death and hospitalization in old age, with over one-third of the older adults experiencing at least one fall or more each year. Because of limited healthcare resources, regular objective fall risk assessments are not possible in the community on a large scale. New methods for fall prediction are necessary to identify and monitor those older people at high risk of falling who would benefit from participating in falls prevention programmes. Technological advances have enabled less expensive ways to quantify physical fall risk in clinical practice and in the homes of older people. Recently, several studies have demonstrated that sensor-based fall risk assessments of postural sway, functional mobility, stepping and walking can discriminate between fallers and nonfallers. Recent research has used low-cost, portable and objective measuring instruments to assess fall risk in older people. Future use of these technologies holds promise for assessing fall risk accurately in an unobtrusive manner in clinical and daily life settings.

  14. Prediction of functional sites in proteins using conserved functional group analysis.

    Science.gov (United States)

    Innis, C Axel; Anand, A Prem; Sowdhamini, R

    2004-04-02

    A detailed knowledge of a protein's functional site is an absolute prerequisite for understanding its mode of action at the molecular level. However, the rapid pace at which sequence and structural information is being accumulated for proteins greatly exceeds our ability to determine their biochemical roles experimentally. As a result, computational methods are required which allow for the efficient processing of the evolutionary information contained in this wealth of data, in particular that related to the nature and location of functionally important sites and residues. The method presented here, referred to as conserved functional group (CFG) analysis, relies on a simplified representation of the chemical groups found in amino acid side-chains to identify functional sites from a single protein structure and a number of its sequence homologues. We show that CFG analysis can fully or partially predict the location of functional sites in approximately 96% of the 470 cases tested and that, unlike other methods available, it is able to tolerate wide variations in sequence identity. In addition, we discuss its potential in a structural genomics context, where automation, scalability and efficiency are critical, and an increasing number of protein structures are determined with no prior knowledge of function. This is exemplified by our analysis of the hypothetical protein Ydde_Ecoli, whose structure was recently solved by members of the North East Structural Genomics consortium. Although the proposed active site for this protein needs to be validated experimentally, this example illustrates the scope of CFG analysis as a general tool for the identification of residues likely to play an important role in a protein's biochemical function. Thus, our method offers a convenient solution to rapidly and automatically process the vast amounts of data that are beginning to emerge from structural genomics projects.

  15. Musite, a tool for global prediction of general and kinase-specific phosphorylation sites.

    Science.gov (United States)

    Gao, Jianjiong; Thelen, Jay J; Dunker, A Keith; Xu, Dong

    2010-12-01

    Reversible protein phosphorylation is one of the most pervasive post-translational modifications, regulating diverse cellular processes in various organisms. High throughput experimental studies using mass spectrometry have identified many phosphorylation sites, primarily from eukaryotes. However, the vast majority of phosphorylation sites remain undiscovered, even in well studied systems. Because mass spectrometry-based experimental approaches for identifying phosphorylation events are costly, time-consuming, and biased toward abundant proteins and proteotypic peptides, in silico prediction of phosphorylation sites is potentially a useful alternative strategy for whole proteome annotation. Because of various limitations, current phosphorylation site prediction tools were not well designed for comprehensive assessment of proteomes. Here, we present a novel software tool, Musite, specifically designed for large scale predictions of both general and kinase-specific phosphorylation sites. We collected phosphoproteomics data in multiple organisms from several reliable sources and used them to train prediction models by a comprehensive machine-learning approach that integrates local sequence similarities to known phosphorylation sites, protein disorder scores, and amino acid frequencies. Application of Musite on several proteomes yielded tens of thousands of phosphorylation site predictions at a high stringency level. Cross-validation tests show that Musite achieves some improvement over existing tools in predicting general phosphorylation sites, and it is at least comparable with those for predicting kinase-specific phosphorylation sites. In Musite V1.0, we have trained general prediction models for six organisms and kinase-specific prediction models for 13 kinases or kinase families. Although the current pretrained models were not correlated with any particular cellular conditions, Musite provides a unique functionality for training customized prediction models

  16. Prediction of glutathionylation sites in proteins using minimal sequence information and their experimental validation.

    Science.gov (United States)

    Pal, Debojyoti; Sharma, Deepak; Kumar, Mukesh; Sandur, Santosh K

    2016-09-01

    S-glutathionylation of proteins plays an important role in various biological processes and is known to be protective modification during oxidative stress. Since, experimental detection of S-glutathionylation is labor intensive and time consuming, bioinformatics based approach is a viable alternative. Available methods require relatively longer sequence information, which may prevent prediction if sequence information is incomplete. Here, we present a model to predict glutathionylation sites from pentapeptide sequences. It is based upon differential association of amino acids with glutathionylated and non-glutathionylated cysteines from a database of experimentally verified sequences. This data was used to calculate position dependent F-scores, which measure how a particular amino acid at a particular position may affect the likelihood of glutathionylation event. Glutathionylation-score (G-score), indicating propensity of a sequence to undergo glutathionylation, was calculated using position-dependent F-scores for each amino-acid. Cut-off values were used for prediction. Our model returned an accuracy of 58% with Matthew's correlation-coefficient (MCC) value of 0.165. On an independent dataset, our model outperformed the currently available model, in spite of needing much less sequence information. Pentapeptide motifs having high abundance among glutathionylated proteins were identified. A list of potential glutathionylation hotspot sequences were obtained by assigning G-scores and subsequent Protein-BLAST analysis revealed a total of 254 putative glutathionable proteins, a number of which were already known to be glutathionylated. Our model predicted glutathionylation sites in 93.93% of experimentally verified glutathionylated proteins. Outcome of this study may assist in discovering novel glutathionylation sites and finding candidate proteins for glutathionylation.

  17. Predicting Protein-Protein Interaction Sites with a Novel Membership Based Fuzzy SVM Classifier.

    Science.gov (United States)

    Sriwastava, Brijesh K; Basu, Subhadip; Maulik, Ujjwal

    2015-01-01

    Predicting residues that participate in protein-protein interactions (PPI) helps to identify, which amino acids are located at the interface. In this paper, we show that the performance of the classical support vector machine (SVM) algorithm can further be improved with the use of a custom-designed fuzzy membership function, for the partner-specific PPI interface prediction problem. We evaluated the performances of both classical SVM and fuzzy SVM (F-SVM) on the PPI databases of three different model proteomes of Homo sapiens, Escherichia coli and Saccharomyces Cerevisiae and calculated the statistical significance of the developed F-SVM over classical SVM algorithm. We also compared our performance with the available state-of-the-art fuzzy methods in this domain and observed significant performance improvements. To predict interaction sites in protein complexes, local composition of amino acids together with their physico-chemical characteristics are used, where the F-SVM based prediction method exploits the membership function for each pair of sequence fragments. The average F-SVM performance (area under ROC curve) on the test samples in 10-fold cross validation experiment are measured as 77.07, 78.39, and 74.91 percent for the aforementioned organisms respectively. Performances on independent test sets are obtained as 72.09, 73.24 and 82.74 percent respectively. The software is available for free download from http://code.google.com/p/cmater-bioinfo.

  18. Computational methods in sequence and structure prediction

    Science.gov (United States)

    Lang, Caiyi

    This dissertation is organized into two parts. In the first part, we will discuss three computational methods for cis-regulatory element recognition in three different gene regulatory networks as the following: (a) Using a comprehensive "Phylogenetic Footprinting Comparison" method, we will investigate the promoter sequence structures of three enzymes (PAL, CHS and DFR) that catalyze sequential steps in the pathway from phenylalanine to anthocyanins in plants. Our result shows there exists a putative cis-regulatory element "AC(C/G)TAC(C)" in the upstream of these enzyme genes. We propose this cis-regulatory element to be responsible for the genetic regulation of these three enzymes and this element, might also be the binding site for MYB class transcription factor PAP1. (b) We will investigate the role of the Arabidopsis gene glutamate receptor 1.1 (AtGLR1.1) in C and N metabolism by utilizing the microarray data we obtained from AtGLR1.1 deficient lines (antiAtGLR1.1). We focus our investigation on the putatively co-regulated transcript profile of 876 genes we have collected in antiAtGLR1.1 lines. By (a) scanning the occurrence of several groups of known abscisic acid (ABA) related cisregulatory elements in the upstream regions of 876 Arabidopsis genes; and (b) exhaustive scanning of all possible 6-10 bps motif occurrence in the upstream regions of the same set of genes, we are able to make a quantative estimation on the enrichment level of each of the cis-regulatory element candidates. We finally conclude that one specific cis-regulatory element group, called "ABRE" elements, are statistically highly enriched within the 876-gene group as compared to their occurrence within the genome. (c) We will introduce a new general purpose algorithm, called "fuzzy REDUCE1", which we have developed recently for automated cis-regulatory element identification. In the second part, we will discuss our newly devised protein design framework. With this framework we have developed

  19. Predictability of bone density at posterior mandibular implant sites using cone-beam computed tomography intensity values

    OpenAIRE

    Alkhader, Mustafa; Hudieb, Malik; Khader, Yousef

    2017-01-01

    Objective: The aim of this study was to investigate the predictability of bone density at posterior mandibular implant sites using cone-beam computed tomography (CBCT) intensity values. Materials and Methods: CBCT cross-sectional images for 436 posterior mandibular implant sites were selected for the study. Using Invivo software (Anatomage, San Jose, California, USA), two observers classified the bone density into three categories: low, intermediate, and high, and CBCT intensity values were g...

  20. LIGSITEcsc: predicting ligand binding sites using the Connolly surface and degree of conservation

    Directory of Open Access Journals (Sweden)

    Schroeder Michael

    2006-09-01

    Full Text Available Abstract Background Identifying pockets on protein surfaces is of great importance for many structure-based drug design applications and protein-ligand docking algorithms. Over the last ten years, many geometric methods for the prediction of ligand-binding sites have been developed. Results We present LIGSITEcsc, an extension and implementation of the LIGSITE algorithm. LIGSITEcsc is based on the notion of surface-solvent-surface events and the degree of conservation of the involved surface residues. We compare our algorithm to four other approaches, LIGSITE, CAST, PASS, and SURFNET, and evaluate all on a dataset of 48 unbound/bound structures and 210 bound-structures. LIGSITEcsc performs slightly better than the other tools and achieves a success rate of 71% and 75%, respectively. Conclusion The use of the Connolly surface leads to slight improvements, the prediction re-ranking by conservation to significant improvements of the binding site predictions. A web server for LIGSITEcsc and its source code is available at scoppi.biotec.tu-dresden.de/pocket.

  1. Methods for assessing mine site rehabilitation design for erosion impact

    International Nuclear Information System (INIS)

    Evans, K. G.

    2000-01-01

    Erosion of rehabilitated mines may result in landform instability, which in turn may result in exposure of encapsulated contaminants, elevated sediment delivery at catchment outlets, and subsequent degradation of downstream water quality. Rehabilitation design can be assessed using erosion and hydrology models calibrated to mine site conditions. Incision rates in containment structures can be quantified using 3-dimensional landform evolution simulation techniques. Sediment delivery at catchment outlets for various landform amelioration techniques can be predicted using process-based and empirical erosion-prediction models and sediment delivery ratios. The predicted sediment delivery can be used to estimate an average annual stream sediment load that can, in turn, be used to assess water quality impacts. Application of these techniques is demonstrated through a case study applied to a proposed rehabilitation design option for the Energy Resources of Australia Ltd (ERA) Ranger Mine in the Northern Territory of Australia. Copyright (2000) CSIRO Australia

  2. Hanford Site groundwater monitoring: Setting, sources and methods

    International Nuclear Information System (INIS)

    Hartman, M.J.

    2000-01-01

    Groundwater monitoring is conducted on the Hanford Site to meet the requirements of the Resource Conservation and Recovery Act of 1976 (RCRA); Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA); U.S. Department of Energy (DOE) orders; and the Washington Administrative Code. Results of monitoring are published annually (e.g., PNNL-11989). To reduce the redundancy of these annual reports, background information that does not change significantly from year to year has been extracted from the annual report and published in this companion volume. This report includes a description of groundwater monitoring requirements, site hydrogeology, and waste sites that have affected groundwater quality or that require groundwater monitoring. Monitoring networks and methods for sampling, analysis, and interpretation are summarized. Vadose zone monitoring methods and statistical methods also are described. Whenever necessary, updates to information contained in this document will be published in future groundwater annual reports

  3. Hanford Site groundwater monitoring: Setting, sources and methods

    Energy Technology Data Exchange (ETDEWEB)

    M.J. Hartman

    2000-04-11

    Groundwater monitoring is conducted on the Hanford Site to meet the requirements of the Resource Conservation and Recovery Act of 1976 (RCRA); Comprehensive Environmental Response, Compensation, and Liability Act of 1980 (CERCLA); U.S. Department of Energy (DOE) orders; and the Washington Administrative Code. Results of monitoring are published annually (e.g., PNNL-11989). To reduce the redundancy of these annual reports, background information that does not change significantly from year to year has been extracted from the annual report and published in this companion volume. This report includes a description of groundwater monitoring requirements, site hydrogeology, and waste sites that have affected groundwater quality or that require groundwater monitoring. Monitoring networks and methods for sampling, analysis, and interpretation are summarized. Vadose zone monitoring methods and statistical methods also are described. Whenever necessary, updates to information contained in this document will be published in future groundwater annual reports.

  4. POLYAR, a new computer program for prediction of poly(A sites in human sequences

    Directory of Open Access Journals (Sweden)

    Qamar Raheel

    2010-11-01

    Full Text Available Abstract Background mRNA polyadenylation is an essential step of pre-mRNA processing in eukaryotes. Accurate prediction of the pre-mRNA 3'-end cleavage/polyadenylation sites is important for defining the gene boundaries and understanding gene expression mechanisms. Results 28761 human mapped poly(A sites have been classified into three classes containing different known forms of polyadenylation signal (PAS or none of them (PAS-strong, PAS-weak and PAS-less, respectively and a new computer program POLYAR for the prediction of poly(A sites of each class was developed. In comparison with polya_svm (till date the most accurate computer program for prediction of poly(A sites while searching for PAS-strong poly(A sites in human sequences, POLYAR had a significantly higher prediction sensitivity (80.8% versus 65.7% and specificity (66.4% versus 51.7% However, when a similar sort of search was conducted for PAS-weak and PAS-less poly(A sites, both programs had a very low prediction accuracy, which indicates that our knowledge about factors involved in the determination of the poly(A sites is not sufficient to identify such polyadenylation regions. Conclusions We present a new classification of polyadenylation sites into three classes and a novel computer program POLYAR for prediction of poly(A sites/regions of each of the class. In tests, POLYAR shows high accuracy of prediction of the PAS-strong poly(A sites, though this program's efficiency in searching for PAS-weak and PAS-less poly(A sites is not very high but is comparable to other available programs. These findings suggest that additional characteristics of such poly(A sites remain to be elucidated. POLYAR program with a stand-alone version for downloading is available at http://cub.comsats.edu.pk/polyapredict.htm.

  5. Expeditious Methods for Site Characterization and Risk Assessment at Department of Defense Hazardous Waste Sites in the Republic of Korea

    National Research Council Canada - National Science Library

    Hartman, Dean

    1999-01-01

    ...) with preferred innovative site characterization technologies and risk assessment methods to meet their needs in obtaining hazardous waste site data and then prioritizing those sites for remediation based upon risk...

  6. Analytical methods for predicting contaminant transport

    International Nuclear Information System (INIS)

    Pigford, T.H.

    1989-09-01

    This paper summarizes some of the previous and recent work at the University of California on analytical solutions for predicting contaminate transport in porous and fractured geologic media. Emphasis is given here to the theories for predicting near-field transport, needed to derive the time-dependent source term for predicting far-field transport and overall repository performance. New theories summarized include solubility-limited release rate with flow backfill in rock, near-field transport of radioactive decay chains, interactive transport of colloid and solute, transport of carbon-14 as carbon dioxide in unsaturated rock, and flow of gases out of and a waste container through cracks and penetrations. 28 refs., 4 figs

  7. Evaluation of two surveillance methods for surgical site infection

    Directory of Open Access Journals (Sweden)

    M. Haji Abdolbaghi

    2006-08-01

    Full Text Available Background: Surgical wound infection surveillance is an important facet of hospital infection control processes. There are several surveillance methods for surgical site infections. The objective of this study is to evaluate the accuracy of two different surgical site infection surveillance methods. Methods: In this prospective cross sectional study 3020 undergoing surgey in general surgical wards of Imam Khomeini hospital were included. Surveillance methods consisted of review of medical records for postoperative fever and review of nursing daily note for prescription of antibiotics postoperatively and during patient’s discharge. Review of patient’s history and daily records and interview with patient’s surgeon and the head-nurse of the ward considered as a gold standard for surveillance. Results: The postoperative antibiotic consumption especially when considering its duration is a proper method for surgical wound infection surveillance. Accomplishments of a prospective study with postdischarge follow up until 30 days after surgery is recommended. Conclusion: The result of this study showed that postoperative antibiotic surveillance method specially with consideration of the antibiotic usage duration is a proper method for surgical site infection surveillance in general surgery wards. Accomplishments of a prospective study with post discharge follow up until 30 days after surgery is recommended.

  8. Machine Learning Methods to Predict Diabetes Complications.

    Science.gov (United States)

    Dagliati, Arianna; Marini, Simone; Sacchi, Lucia; Cogni, Giulia; Teliti, Marsida; Tibollo, Valentina; De Cata, Pasquale; Chiovato, Luca; Bellazzi, Riccardo

    2018-03-01

    One of the areas where Artificial Intelligence is having more impact is machine learning, which develops algorithms able to learn patterns and decision rules from data. Machine learning algorithms have been embedded into data mining pipelines, which can combine them with classical statistical strategies, to extract knowledge from data. Within the EU-funded MOSAIC project, a data mining pipeline has been used to derive a set of predictive models of type 2 diabetes mellitus (T2DM) complications based on electronic health record data of nearly one thousand patients. Such pipeline comprises clinical center profiling, predictive model targeting, predictive model construction and model validation. After having dealt with missing data by means of random forest (RF) and having applied suitable strategies to handle class imbalance, we have used Logistic Regression with stepwise feature selection to predict the onset of retinopathy, neuropathy, or nephropathy, at different time scenarios, at 3, 5, and 7 years from the first visit at the Hospital Center for Diabetes (not from the diagnosis). Considered variables are gender, age, time from diagnosis, body mass index (BMI), glycated hemoglobin (HbA1c), hypertension, and smoking habit. Final models, tailored in accordance with the complications, provided an accuracy up to 0.838. Different variables were selected for each complication and time scenario, leading to specialized models easy to translate to the clinical practice.

  9. Integrated account of method, site selection and programme prior to the site investigation phase

    International Nuclear Information System (INIS)

    2000-12-01

    applications and have these applications reviewed by the appropriate authorities. An analysis of conceivable alternatives for managing and disposing of spent nuclear fuel has confirmed that deep geological disposal according to the KBS-3 method has the best prospects of meeting all requirements. The alternative of putting off a decision until some future time (the zero alternative) does not appear tenable. The assessment of long-term safety shows that the prospects of building a safe deep repository in the Swedish bedrock are good. Independent Swedish and international review of the safety assessment confirm that the body of data in this respect is adequate for the siting process to proceed to the site investigation phase. A fuller summary is given below of the account given in this report of method as well as site selection and programme for the site investigation phase. The point of departure for the account is the review comments made by the regulatory authorities and the Government's decision regarding RD and D-Programme 98. In its decision, the Government stipulated conditions for SKB's continued research and development programme. The analysis of alternative system designs was to be supplemented, mainly with regard to the zero alternative and very deep boreholes. Furthermore, the Government decided that SKB shall submit an integrated evaluation of completed feasibility studies and other background material for selection of sites for site investigations and present a clear programme for site investigations

  10. Seismic assessment of a site using the time series method

    International Nuclear Information System (INIS)

    Krutzik, N.J.; Rotaru, I.; Bobei, M.; Mingiuc, C.; Serban, V.; Androne, M.

    1997-01-01

    To increase the safety of a NPP located on a seismic site, the seismic acceleration level to which the NPP should be qualified must be as representative as possible for that site, with a conservative degree of safety but not too exaggerated. The consideration of the seismic events affecting the site as independent events and the use of statistic methods to define some safety levels with very low annual occurrence probability (10 -4 ) may lead to some exaggerations of the seismic safety level. The use of some very high value for the seismic acceleration imposed by the seismic safety levels required by the hazard analysis may lead to very costly technical solutions that can make the plant operation more difficult and increase maintenance costs. The considerations of seismic events as a time series with dependence among the events produced, may lead to a more representative assessment of a NPP site seismic activity and consequently to a prognosis on the seismic level values to which the NPP would be ensured throughout its life-span. That prognosis should consider the actual seismic activity (including small earthquakes in real time) of the focuses that affect the plant site. The paper proposes the applications of Autoregressive Time Series to issue a prognosis on the seismic activity of a focus and presents the analysis on Vrancea focus that affects NPP Cernavoda site, by this method. The paper also presents the manner to analyse the focus activity as per the new approach and it assesses the maximum seismic acceleration that may affect NPP Cernavoda throughout its life-span (∼ 30 years). Development and applications of new mathematical analysis method, both for long - and short - time intervals, may lead to important contributions in the process of foretelling the seismic events in the future. (authors)

  11. Different Methods of Predicting Permeability in Shale

    DEFF Research Database (Denmark)

    Mbia, Ernest Ncha; Fabricius, Ida Lykke; Krogsbøll, Anette

    by two to five orders of magnitudes at lower vertical effective stress below 40 MPa as the content of clay minerals increases causing heterogeneity in shale material. Indirect permeability from consolidation can give maximum and minimum values of shale permeability needed in simulating fluid flow......Permeability is often very difficult to measure or predict in shale lithology. In this work we are determining shale permeability from consolidation tests data using Wissa et al., (1971) approach and comparing the results with predicted permeability from Kozeny’s model. Core and cuttings materials...... effective stress to 9 μD at high vertical effective stress of 100 MPa. The indirect permeability calculated from consolidation tests falls in the same magnitude at higher vertical effective stress, above 40 MPa, as that of the Kozeny model for shale samples with high non-clay content ≥ 70% but are higher...

  12. Seismic assessment of a site using the time series method

    International Nuclear Information System (INIS)

    Krutzik, N.J.; Rotaru, I.; Bobei, M.; Mingiuc, C.; Serban, V.; Androne, M.

    2001-01-01

    1. To increase the safety of a NPP located on a seismic site, the seismic acceleration level to which the NPP should be qualified must be as representative as possible for that site, with a conservative degree of safety but not too exaggerated. 2. The consideration of the seismic events affecting the site as independent events and the use of statistic methods to define some safety levels with very low annual occurrence probabilities (10 -4 ) may lead to some exaggerations of the seismic safety level. 3. The use of some very high values for the seismic accelerations imposed by the seismic safety levels required by the hazard analysis may lead to very expensive technical solutions that can make the plant operation more difficult and increase the maintenance costs. 4. The consideration of seismic events as a time series with dependence among the events produced may lead to a more representative assessment of a NPP site seismic activity and consequently to a prognosis on the seismic level values to which the NPP would be ensured throughout its life-span. That prognosis should consider the actual seismic activity (including small earthquakes in real time) of the focuses that affect the plant site. The method is useful for two purposes: a) research, i.e. homogenizing the history data basis by the generation of earthquakes during periods lacking information and correlation of the information with the existing information. The aim is to perform the hazard analysis using a homogeneous data set in order to determine the seismic design data for a site; b) operation, i.e. the performance of a prognosis on the seismic activity on a certain site and consideration of preventive measures to minimize the possible effects of an earthquake. 5. The paper proposes the application of Autoregressive Time Series to issue a prognosis on the seismic activity of a focus and presents the analysis on Vrancea focus that affects Cernavoda NPP site by this method. 6. The paper also presents the

  13. Soil characterization methods for unsaturated low-level waste sites

    International Nuclear Information System (INIS)

    Wierenga, P.J.; Young, M.H.; Hills, R.G.

    1993-01-01

    To support a license application for the disposal of low-level radioactive waste (LLW), applicants must characterize the unsaturated zone and demonstrate that waste will not migrate from the facility boundary. This document provides a strategy for developing this characterization plan. It describes principles of contaminant flow and transport, site characterization and monitoring strategies, and data management. It also discusses methods and practices that are currently used to monitor properties and conditions in the soil profile, how these properties influence water and waste migration, and why they are important to the license application. The methods part of the document is divided into sections on laboratory and field-based properties, then further subdivided into the description of methods for determining 18 physical, flow, and transport properties. Because of the availability of detailed procedures in many texts and journal articles, the reader is often directed for details to the available literature. References are made to experiments performed at the Las Cruces Trench site, New Mexico, that support LLW site characterization activities. A major contribution from the Las Cruces study is the experience gained in handling data sets for site characterization and the subsequent use of these data sets in modeling studies

  14. Connecting clinical and actuarial prediction with rule-based methods

    NARCIS (Netherlands)

    Fokkema, M.; Smits, N.; Kelderman, H.; Penninx, B.W.J.H.

    2015-01-01

    Meta-analyses comparing the accuracy of clinical versus actuarial prediction have shown actuarial methods to outperform clinical methods, on average. However, actuarial methods are still not widely used in clinical practice, and there has been a call for the development of actuarial prediction

  15. Prediction of Water Binding to Protein Hydration Sites with a Discrete, Semiexplicit Solvent Model.

    Science.gov (United States)

    Setny, Piotr

    2015-12-08

    Buried water molecules are ubiquitous in protein structures and are found at the interface of most protein-ligand complexes. Determining their distribution and thermodynamic effect is a challenging yet important task, of great of practical value for the modeling of biomolecular structures and their interactions. In this study, we present a novel method aimed at the prediction of buried water molecules in protein structures and estimation of their binding free energies. It is based on a semiexplicit, discrete solvation model, which we previously introduced in the context of small molecule hydration. The method is applicable to all macromolecular structures described by a standard all-atom force field, and predicts complete solvent distribution within a single run with modest computational cost. We demonstrate that it indicates positions of buried hydration sites, including those filled by more than one water molecule, and accurately differentiates them from sterically accessible to water but void regions. The obtained estimates of water binding free energies are in fair agreement with reference results determined with the double decoupling method.

  16. Can Morphing Methods Predict Intermediate Structures?

    Science.gov (United States)

    Weiss, Dahlia R.; Levitt, Michael

    2009-01-01

    Movement is crucial to the biological function of many proteins, yet crystallographic structures of proteins can give us only a static snapshot. The protein dynamics that are important to biological function often happen on a timescale that is unattainable through detailed simulation methods such as molecular dynamics as they often involve crossing high-energy barriers. To address this coarse-grained motion, several methods have been implemented as web servers in which a set of coordinates is usually linearly interpolated from an initial crystallographic structure to a final crystallographic structure. We present a new morphing method that does not extrapolate linearly and can therefore go around high-energy barriers and which can produce different trajectories between the same two starting points. In this work, we evaluate our method and other established coarse-grained methods according to an objective measure: how close a coarse-grained dynamics method comes to a crystallographically determined intermediate structure when calculating a trajectory between the initial and final crystal protein structure. We test this with a set of five proteins with at least three crystallographically determined on-pathway high-resolution intermediate structures from the Protein Data Bank. For simple hinging motions involving a small conformational change, segmentation of the protein into two rigid sections outperforms other more computationally involved methods. However, large-scale conformational change is best addressed using a nonlinear approach and we suggest that there is merit in further developing such methods. PMID:18996395

  17. PROSPER: an integrated feature-based tool for predicting protease substrate cleavage sites.

    Directory of Open Access Journals (Sweden)

    Jiangning Song

    Full Text Available The ability to catalytically cleave protein substrates after synthesis is fundamental for all forms of life. Accordingly, site-specific proteolysis is one of the most important post-translational modifications. The key to understanding the physiological role of a protease is to identify its natural substrate(s. Knowledge of the substrate specificity of a protease can dramatically improve our ability to predict its target protein substrates, but this information must be utilized in an effective manner in order to efficiently identify protein substrates by in silico approaches. To address this problem, we present PROSPER, an integrated feature-based server for in silico identification of protease substrates and their cleavage sites for twenty-four different proteases. PROSPER utilizes established specificity information for these proteases (derived from the MEROPS database with a machine learning approach to predict protease cleavage sites by using different, but complementary sequence and structure characteristics. Features used by PROSPER include local amino acid sequence profile, predicted secondary structure, solvent accessibility and predicted native disorder. Thus, for proteases with known amino acid specificity, PROSPER provides a convenient, pre-prepared tool for use in identifying protein substrates for the enzymes. Systematic prediction analysis for the twenty-four proteases thus far included in the database revealed that the features we have included in the tool strongly improve performance in terms of cleavage site prediction, as evidenced by their contribution to performance improvement in terms of identifying known cleavage sites in substrates for these enzymes. In comparison with two state-of-the-art prediction tools, PoPS and SitePrediction, PROSPER achieves greater accuracy and coverage. To our knowledge, PROSPER is the first comprehensive server capable of predicting cleavage sites of multiple proteases within a single substrate

  18. NetOglyc: prediction of mucin type O-glycosylation sites based on sequence context and surface accessibility

    DEFF Research Database (Denmark)

    Hansen, Jan Erik; Lund, Ole; Tolstrup, Niels

    1998-01-01

    -glycosylated serine and threonine residues in independent test sets, thus proving more accurate than matrix statistics and vector projection methods. Predicition of O-glycosylation sites in the envelope glycoprotein gp120 from the primate lentiviruses HIV-1, HIV-2 and SIV are presented. The most conserved O...... structure and surface accessibility. The sequence context of glycosylated threonines was found to differ from that of serine, and the sites were found to cluster. Non-clustered sites had a sequence context different from that of clustered sites. charged residues were disfavoured at postition -1 and +3......-glycosylation signals in these evolutionary-related glycoproteins were found in their first hypervariable loop, V1. However, the strain variation for HIV-1 gp120 was significant. A computer server, available through WWW or E-mail, has been developed for prediction of mucin type O-glycosylation sites in proteins based...

  19. Using TESS to predict transcription factor binding sites in DNA sequence.

    Science.gov (United States)

    Schug, Jonathan

    2008-03-01

    This unit describes how to use the Transcription Element Search System (TESS). This Web site predicts transcription factor binding sites (TFBS) in DNA sequence using two different kinds of models of sites, strings and positional weight matrices. The binding of transcription factors to DNA is a major part of the control of gene expression. Transcription factors exhibit sequence-specific binding; they form stronger bonds to some DNA sequences than to others. Identification of a good binding site in the promoter for a gene suggests the possibility that the corresponding factor may play a role in the regulation of that gene. However, the sequences transcription factors recognize are typically short and allow for some amount of mismatch. Because of this, binding sites for a factor can typically be found at random every few hundred to a thousand base pairs. TESS has features to help sort through and evaluate the significance of predicted sites.

  20. Prediction, conservation analysis, and structural characterization of mammalian mucin-type O-glycosylation sites

    DEFF Research Database (Denmark)

    Julenius, Karin; Mølgaard, Anne; Gupta, Ramneek

    2005-01-01

    could be predicted from averaged properties together with the fact that glycosylation sites are not precisely conserved indicates that mucin-type glycosylation in most cases is a bulk property and not a very site-specific one. NetOGlyc 3.1 is made available at www.cbs.dtu.dk/services/netoglyc....

  1. Prediction Methods in Science and Technology

    DEFF Research Database (Denmark)

    Høskuldsson, Agnar

    Presents the H-principle, the Heisenberg modelling principle. General properties of the Heisenberg modelling procedure is developed. The theory is applied to principal component analysis and linear regression analysis. It is shown that the H-principle leads to PLS regression in case the task...... is linear regression analysis. The book contains different methods to find the dimensions of linear models, to carry out sensitivity analysis in latent structure models, variable selection methods and presentation of results from analysis....

  2. Prediction of Active Site and Distal Residues in E. coli DNA Polymerase III alpha Polymerase Activity.

    Science.gov (United States)

    Parasuram, Ramya; Coulther, Timothy A; Hollander, Judith M; Keston-Smith, Elise; Ondrechen, Mary Jo; Beuning, Penny J

    2018-02-20

    The process of DNA replication is carried out with high efficiency and accuracy by DNA polymerases. The replicative polymerase in E. coli is DNA Pol III, which is a complex of 10 different subunits that coordinates simultaneous replication on the leading and lagging strands. The 1160-residue Pol III alpha subunit is responsible for the polymerase activity and copies DNA accurately, making one error per 10 5 nucleotide incorporations. The goal of this research is to determine the residues that contribute to the activity of the polymerase subunit. Homology modeling and the computational methods of THEMATICS and POOL were used to predict functionally important amino acid residues through their computed chemical properties. Site-directed mutagenesis and biochemical assays were used to validate these predictions. Primer extension, steady-state single-nucleotide incorporation kinetics, and thermal denaturation assays were performed to understand the contribution of these residues to the function of the polymerase. This work shows that the top 15 residues predicted by POOL, a set that includes the three previously known catalytic aspartate residues, seven remote residues, plus five previously unexplored first-layer residues, are important for function. Six previously unidentified residues, R362, D405, K553, Y686, E688, and H760, are each essential to Pol III activity; three additional residues, Y340, R390, and K758, play important roles in activity.

  3. Conserved Functional Motifs and Homology Modeling to Predict Hidden Moonlighting Functional Sites

    KAUST Repository

    Wong, Aloysius Tze; Gehring, Christoph A; Irving, Helen R.

    2015-01-01

    Moonlighting functional centers within proteins can provide them with hitherto unrecognized functions. Here, we review how hidden moonlighting functional centers, which we define as binding sites that have catalytic activity or regulate protein function in a novel manner, can be identified using targeted bioinformatic searches. Functional motifs used in such searches include amino acid residues that are conserved across species and many of which have been assigned functional roles based on experimental evidence. Molecules that were identified in this manner seeking cyclic mononucleotide cyclases in plants are used as examples. The strength of this computational approach is enhanced when good homology models can be developed to test the functionality of the predicted centers in silico, which, in turn, increases confidence in the ability of the identified candidates to perform the predicted functions. Computational characterization of moonlighting functional centers is not diagnostic for catalysis but serves as a rapid screening method, and highlights testable targets from a potentially large pool of candidates for subsequent in vitro and in vivo experiments required to confirm the functionality of the predicted moonlighting centers.

  4. Conserved Functional Motifs and Homology Modeling to Predict Hidden Moonlighting Functional Sites

    KAUST Repository

    Wong, Aloysius Tze

    2015-06-09

    Moonlighting functional centers within proteins can provide them with hitherto unrecognized functions. Here, we review how hidden moonlighting functional centers, which we define as binding sites that have catalytic activity or regulate protein function in a novel manner, can be identified using targeted bioinformatic searches. Functional motifs used in such searches include amino acid residues that are conserved across species and many of which have been assigned functional roles based on experimental evidence. Molecules that were identified in this manner seeking cyclic mononucleotide cyclases in plants are used as examples. The strength of this computational approach is enhanced when good homology models can be developed to test the functionality of the predicted centers in silico, which, in turn, increases confidence in the ability of the identified candidates to perform the predicted functions. Computational characterization of moonlighting functional centers is not diagnostic for catalysis but serves as a rapid screening method, and highlights testable targets from a potentially large pool of candidates for subsequent in vitro and in vivo experiments required to confirm the functionality of the predicted moonlighting centers.

  5. Evaluation of the site effect with Heuristic Methods

    Science.gov (United States)

    Torres, N. N.; Ortiz-Aleman, C.

    2017-12-01

    The seismic site response in an area depends mainly on the local geological and topographical conditions. Estimation of variations in ground motion can lead to significant contributions on seismic hazard assessment, in order to reduce human and economic losses. Site response estimation can be posed as a parameterized inversion approach which allows separating source and path effects. The generalized inversion (Field and Jacob, 1995) represents one of the alternative methods to estimate the local seismic response, which involves solving a strongly non-linear multiparametric problem. In this work, local seismic response was estimated using global optimization methods (Genetic Algorithms and Simulated Annealing) which allowed us to increase the range of explored solutions in a nonlinear search, as compared to other conventional linear methods. By using the VEOX Network velocity records, collected from August 2007 to March 2009, source, path and site parameters corresponding to the amplitude spectra of the S wave of the velocity seismic records are estimated. We can establish that inverted parameters resulting from this simultaneous inversion approach, show excellent agreement, not only in terms of adjustment between observed and calculated spectra, but also when compared to previous work from several authors.

  6. Workshop on methods for siting groundwater monitoring wells: Proceedings

    International Nuclear Information System (INIS)

    Jacobson, E.

    1992-02-01

    The primary purpose of this workshop was to identify methods for the optimum siting of groundwater monitoring wells to minimize the number required that will provide statistically and physically representative samples. In addition, the workshop served to identify information and data gaps, stimulated discussion and provided an opportunity for exchange of ideas between regulators and scientists interested in siting groundwater monitoring wells. These proceedings should serve these objectives and provide a source of relevant information which may be used to evaluate the current state of development of methods for siting groundwater monitoring wells and the additional research needs. The proceedings contain the agenda and list of attendees in the first section. The abstract and viewgraphs for each presentation are given in the second section. For several presentations, abstracts and viewgraphs were not received. After the presentations, four working groups were organized and met for approximately a day. The working group leaders then gave a verbal summary of their sessions. This material was transcribed and is included in the next section of these proceedings. The appendices contain forms describing various methods discussed in the working groups

  7. Generic methods for aero-engine exhaust emission prediction

    NARCIS (Netherlands)

    Shakariyants, S.A.

    2008-01-01

    In the thesis, generic methods have been developed for aero-engine combustor performance, combustion chemistry, as well as airplane aerodynamics, airplane and engine performance. These methods specifically aim to support diverse emission prediction studies coupled with airplane and engine

  8. Force prediction in cold rolling mills by polynomial methods

    Directory of Open Access Journals (Sweden)

    Nicu ROMAN

    2007-12-01

    Full Text Available A method for steel and aluminium strip thickness control is provided including a new technique for predictive rolling force estimation method by statistic model based on polynomial techniques.

  9. An Approximate Method for Pitch-Damping Prediction

    National Research Council Canada - National Science Library

    Danberg, James

    2003-01-01

    ...) method for predicting the pitch-damping coefficients has been employed. The CFD method provides important details necessary to derive the correlation functions that are unavailable from the current experimental database...

  10. Bouguer density analysis using nettleton method at Banten NPP site

    International Nuclear Information System (INIS)

    Yuliastuti; Hadi Suntoko; Yarianto SBS

    2017-01-01

    Sub-surface information become crucial in determining a feasible NPP site that safe from external hazards. Gravity survey which result as density information, is essential to understand the sub-surface structure. Nevertheless, overcorrected or under corrected will lead to a false interpretation. Therefore, density correction in term of near-surface average density or Bouguer density is necessary to be calculated. The objective of this paper is to estimate and analyze Bouguer density using Nettleton method at Banten NPP Site. Methodology used in this paper is Nettleton method that applied in three different slices (A-B, A-C and A-D) with density assumption range between 1700 and 3300 kg/m"3. Nettleton method is based on minimum correlation between gravity anomaly and topography to determine density correction. The result shows that slice A-B which covers rough topography difference, Nettleton method failed. While using the other two slices, Nettleton method yield with a different density value, 2700 kg/m"3 for A-C and 2300 kg/m"3 for A-D. A-C provides the lowest correlation value which represents the Upper Banten tuff and Gede Mt. volcanic rocks in accordance with Quartenary rocks exist in the studied area. (author)

  11. Prediction of protein hydration sites from sequence by modular neural networks

    DEFF Research Database (Denmark)

    Ehrlich, L.; Reczko, M.; Bohr, Henrik

    1998-01-01

    The hydration properties of a protein are important determinants of its structure and function. Here, modular neural networks are employed to predict ordered hydration sites using protein sequence information. First, secondary structure and solvent accessibility are predicted from sequence with two...... separate neural networks. These predictions are used as input together with protein sequences for networks predicting hydration of residues, backbone atoms and sidechains. These networks are teined with protein crystal structures. The prediction of hydration is improved by adding information on secondary...... structure and solvent accessibility and, using actual values of these properties, redidue hydration can be predicted to 77% accuracy with a Metthews coefficient of 0.43. However, predicted property data with an accuracy of 60-70% result in less than half the improvement in predictive performance observed...

  12. Electron microscopy methods in studies of cultural heritage sites

    Energy Technology Data Exchange (ETDEWEB)

    Vasiliev, A. L., E-mail: a.vasiliev56@gmail.com; Kovalchuk, M. V.; Yatsishina, E. B. [National Research Centre “Kurchatov Institute” (Russian Federation)

    2016-11-15

    The history of the development and application of scanning electron microscopy (SEM), transmission electron microscopy (TEM), and energy-dispersive X-ray microanalysis (EDXMA) in studies of cultural heritage sites is considered. In fact, investigations based on these methods began when electron microscopes became a commercial product. Currently, these methods, being developed and improved, help solve many historical enigmas. To date, electron microscopy combined with microanalysis makes it possible to investigate any object, from parchment and wooden articles to pigments, tools, and objects of art. Studies by these methods have revealed that some articles were made by ancient masters using ancient “nanotechnologies”; hence, their comprehensive analysis calls for the latest achievements in the corresponding instrumental methods and sample preparation techniques.

  13. Electron microscopy methods in studies of cultural heritage sites

    Science.gov (United States)

    Vasiliev, A. L.; Kovalchuk, M. V.; Yatsishina, E. B.

    2016-11-01

    The history of the development and application of scanning electron microscopy (SEM), transmission electron microscopy (TEM), and energy-dispersive X-ray microanalysis (EDXMA) in studies of cultural heritage sites is considered. In fact, investigations based on these methods began when electron microscopes became a commercial product. Currently, these methods, being developed and improved, help solve many historical enigmas. To date, electron microscopy combined with microanalysis makes it possible to investigate any object, from parchment and wooden articles to pigments, tools, and objects of art. Studies by these methods have revealed that some articles were made by ancient masters using ancient "nanotechnologies"; hence, their comprehensive analysis calls for the latest achievements in the corresponding instrumental methods and sample preparation techniques.

  14. Electron microscopy methods in studies of cultural heritage sites

    International Nuclear Information System (INIS)

    Vasiliev, A. L.; Kovalchuk, M. V.; Yatsishina, E. B.

    2016-01-01

    The history of the development and application of scanning electron microscopy (SEM), transmission electron microscopy (TEM), and energy-dispersive X-ray microanalysis (EDXMA) in studies of cultural heritage sites is considered. In fact, investigations based on these methods began when electron microscopes became a commercial product. Currently, these methods, being developed and improved, help solve many historical enigmas. To date, electron microscopy combined with microanalysis makes it possible to investigate any object, from parchment and wooden articles to pigments, tools, and objects of art. Studies by these methods have revealed that some articles were made by ancient masters using ancient “nanotechnologies”; hence, their comprehensive analysis calls for the latest achievements in the corresponding instrumental methods and sample preparation techniques.

  15. A Versatile Nonlinear Method for Predictive Modeling

    Science.gov (United States)

    Liou, Meng-Sing; Yao, Weigang

    2015-01-01

    As computational fluid dynamics techniques and tools become widely accepted for realworld practice today, it is intriguing to ask: what areas can it be utilized to its potential in the future. Some promising areas include design optimization and exploration of fluid dynamics phenomena (the concept of numerical wind tunnel), in which both have the common feature where some parameters are varied repeatedly and the computation can be costly. We are especially interested in the need for an accurate and efficient approach for handling these applications: (1) capturing complex nonlinear dynamics inherent in a system under consideration and (2) versatility (robustness) to encompass a range of parametric variations. In our previous paper, we proposed to use first-order Taylor expansion collected at numerous sampling points along a trajectory and assembled together via nonlinear weighting functions. The validity and performance of this approach was demonstrated for a number of problems with a vastly different input functions. In this study, we are especially interested in enhancing the method's accuracy; we extend it to include the second-orer Taylor expansion, which however requires a complicated evaluation of Hessian matrices for a system of equations, like in fluid dynamics. We propose a method to avoid these Hessian matrices, while maintaining the accuracy. Results based on the method are presented to confirm its validity.

  16. Characterization of hazardous waste sites: a methods manual. Volume 2. Available sampling methods (second edition)

    International Nuclear Information System (INIS)

    Ford, P.J.; Turina, P.J.; Seely, D.E.

    1984-12-01

    Investigations at hazardous waste sites and sites of chemical spills often require on-site measurements and sampling activities to assess the type and extent of contamination. This document is a compilation of sampling methods and materials suitable to address most needs that arise during routine waste site and hazardous spill investigations. The sampling methods presented in this document are compiled by media, and were selected on the basis of practicality, economics, representativeness, compatability with analytical considerations, and safety, as well as other criteria. In addition to sampling procedures, sample handling and shipping, chain-of-custody procedures, instrument certification, equipment fabrication, and equipment decontamination procedures are described. Sampling methods for soil, sludges, sediments, and bulk materials cover the solids medium. Ten methods are detailed for surface waters, groundwater and containerized liquids; twelve are presented for ambient air, soil gases and vapors, and headspace gases. A brief discussion of ionizing radiation survey instruments is also provided

  17. Site-specific data confirm arsenic exposure predicted by the U.S. Environmental Protection Agency.

    Science.gov (United States)

    Walker, S; Griffin, S

    1998-03-01

    The EPA uses an exposure assessment model to estimate daily intake to chemicals of potential concern. At the Anaconda Superfund site in Montana, the EPA exposure assessment model was used to predict total and speciated urinary arsenic concentrations. Predicted concentrations were then compared to concentrations measured in children living near the site. When site-specific information on concentrations of arsenic in soil, interior dust, and diet, site-specific ingestion rates, and arsenic absorption rates were used, measured and predicted urinary arsenic concentrations were in reasonable agreement. The central tendency exposure assessment model successfully described the measured urinary arsenic concentration for the majority of children at the site. The reasonable maximum exposure assessment model successfully identified the uppermost exposed population. While the agreement between measured and predicted urinary arsenic is good, it is not exact. The variables that were identified which influenced agreement included soil and dust sample collection methodology, daily urinary volume, soil ingestion rate, and the ability to define the exposure unit. The concentration of arsenic in food affected agreement between measured and predicted total urinary arsenic, but was not considered when comparing measured and predicted speciated urinary arsenic. Speciated urinary arsenic is the recommended biomarker for recent inorganic arsenic exposure. By using site-specific data in the exposure assessment model, predicted risks from exposure to arsenic were less than predicted risks would have been if the EPA's default values had been used in the exposure assessment model. This difference resulted in reduced magnitude and cost of remediation while still protecting human health.

  18. DASPfind: new efficient method to predict drug–target interactions

    KAUST Repository

    Ba Alawi, Wail

    2016-03-16

    Background Identification of novel drug–target interactions (DTIs) is important for drug discovery. Experimental determination of such DTIs is costly and time consuming, hence it necessitates the development of efficient computational methods for the accurate prediction of potential DTIs. To-date, many computational methods have been proposed for this purpose, but they suffer the drawback of a high rate of false positive predictions. Results Here, we developed a novel computational DTI prediction method, DASPfind. DASPfind uses simple paths of particular lengths inferred from a graph that describes DTIs, similarities between drugs, and similarities between the protein targets of drugs. We show that on average, over the four gold standard DTI datasets, DASPfind significantly outperforms other existing methods when the single top-ranked predictions are considered, resulting in 46.17 % of these predictions being correct, and it achieves 49.22 % correct single top ranked predictions when the set of all DTIs for a single drug is tested. Furthermore, we demonstrate that our method is best suited for predicting DTIs in cases of drugs with no known targets or with few known targets. We also show the practical use of DASPfind by generating novel predictions for the Ion Channel dataset and validating them manually. Conclusions DASPfind is a computational method for finding reliable new interactions between drugs and proteins. We show over six different DTI datasets that DASPfind outperforms other state-of-the-art methods when the single top-ranked predictions are considered, or when a drug with no known targets or with few known targets is considered. We illustrate the usefulness and practicality of DASPfind by predicting novel DTIs for the Ion Channel dataset. The validated predictions suggest that DASPfind can be used as an efficient method to identify correct DTIs, thus reducing the cost of necessary experimental verifications in the process of drug discovery. DASPfind

  19. A Novel Computational Method for Detecting DNA Methylation Sites with DNA Sequence Information and Physicochemical Properties.

    Science.gov (United States)

    Pan, Gaofeng; Jiang, Limin; Tang, Jijun; Guo, Fei

    2018-02-08

    DNA methylation is an important biochemical process, and it has a close connection with many types of cancer. Research about DNA methylation can help us to understand the regulation mechanism and epigenetic reprogramming. Therefore, it becomes very important to recognize the methylation sites in the DNA sequence. In the past several decades, many computational methods-especially machine learning methods-have been developed since the high-throughout sequencing technology became widely used in research and industry. In order to accurately identify whether or not a nucleotide residue is methylated under the specific DNA sequence context, we propose a novel method that overcomes the shortcomings of previous methods for predicting methylation sites. We use k -gram, multivariate mutual information, discrete wavelet transform, and pseudo amino acid composition to extract features, and train a sparse Bayesian learning model to do DNA methylation prediction. Five criteria-area under the receiver operating characteristic curve (AUC), Matthew's correlation coefficient (MCC), accuracy (ACC), sensitivity (SN), and specificity-are used to evaluate the prediction results of our method. On the benchmark dataset, we could reach 0.8632 on AUC, 0.8017 on ACC, 0.5558 on MCC, and 0.7268 on SN. Additionally, the best results on two scBS-seq profiled mouse embryonic stem cells datasets were 0.8896 and 0.9511 by AUC, respectively. When compared with other outstanding methods, our method surpassed them on the accuracy of prediction. The improvement of AUC by our method compared to other methods was at least 0.0399 . For the convenience of other researchers, our code has been uploaded to a file hosting service, and can be downloaded from: https://figshare.com/s/0697b692d802861282d3.

  20. Nuclear method applied in archaeological sites at the Amazon basin

    International Nuclear Information System (INIS)

    Nicoli, Ieda Gomes; Bernedo, Alfredo Victor Bellido; Latini, Rose Mary

    2002-01-01

    The aim of this work was to use the nuclear methodology to character pottery discovered inside archaeological sites recognized with circular earth structure in Acre State - Brazil which may contribute to the research in the reconstruction of part of the pre-history of the Amazonic Basin. The sites are located mainly in the Hydrographic Basin of High Purus River. Three of them were strategic chosen to collect the ceramics: Lobao, in Sena Madureira County at north; Alto Alegre in Rio Branco County at east and Xipamanu I, in Xapuri County at south. Neutron Activation Analysis in conjunction with multivariate statistical methods were used for the ceramic characterization and classification. An homogeneous group was established by all the sherds collected from Alto Alegre and was distinct from the other two groups analyzed. Some of the sherds collected from Xipamunu I appeared in Lobao's urns, probably because they had the same fabrication process. (author)

  1. Site investigations. Investigation methods and general execution programme

    International Nuclear Information System (INIS)

    2001-01-01

    at least one deep chemistry-prioritized cored borehole, and start of long-term monitoring of chemical parameters in new selected sampling points. Fracture-filling mineral investigations are initiated during the final phase of the initial site investigation. The transport properties of the rock are estimated mainly on the basis of the hydrogeological and hydrogeochemical description, combined with generic, non-site-specific information. Furthermore, supplementary measurements of groundwater flow are performed in one of the first deep boreholes. In cases where mineralogy and/or groundwater chemistry differs significantly from the generic database, certain time-consuming laboratory investigations such as through diffusion measurements will be initiated. A large number of methods are used to investigate the geology of the site. They can be described in general terms under the headings: geophysics, surface geology, soil geology, bedrock geology, borehole investigations and geodetic measurements. The thermal properties of the rock are determined primarily on the basis of mineral composition and by means of laboratory studies of recovered rock cores. The determination of the transport properties of the rock is based on generic, non-site-specific data, combined with he hydrogeological and hydrogeochemical description of the rock. Laboratory measurements on drill cores and rock material are used to determine/verify sorption values and diffusivities. Diffusion tests and tracer tests in and between boreholes can be used to verify the reasonableness of estimated parameter values. The discipline surface ecosystems includes both the living environment, i.e. animals and plants, and their interactions with the non-living environment, e.g. climate and water. The characterization includes both hydrogeological and hydrogeochemical characterization of soil strata and surface waters and inventory/characterization of flora and fauna

  2. Cross-prediction of the groundwater chemistry at the SKB sites in Sweden. Pilot study

    International Nuclear Information System (INIS)

    Skaarman, C.; Laaksoharju, M.

    1997-08-01

    The possibility to perform a large scale prediction throughout Sweden was tested. The aim of the work was: to collect data and create a groundwater database for current and future use; to see if there is any correlation between data at different sites; to perform a modelling where the groundwater composition at different regions in Sweden is predicted. The outcome of the predictions were compared with the measured data at different sites. The results show that it is possible but more work needs to be done to improve the prediction models. More measurements at depth are needed to enable the use of 3D models. It is also important to include hydrogeological parameters in the groundwater chemical prediction models that are used

  3. Exploring the role of water in molecular recognition: predicting protein ligandability using a combinatorial search of surface hydration sites

    Science.gov (United States)

    Vukovic, Sinisa; Brennan, Paul E.; Huggins, David J.

    2016-09-01

    The interaction between any two biological molecules must compete with their interaction with water molecules. This makes water the most important molecule in medicine, as it controls the interactions of every therapeutic with its target. A small molecule binding to a protein is able to recognize a unique binding site on a protein by displacing bound water molecules from specific hydration sites. Quantifying the interactions of these water molecules allows us to estimate the potential of the protein to bind a small molecule. This is referred to as ligandability. In the study, we describe a method to predict ligandability by performing a search of all possible combinations of hydration sites on protein surfaces. We predict ligandability as the summed binding free energy for each of the constituent hydration sites, computed using inhomogeneous fluid solvation theory. We compared the predicted ligandability with the maximum observed binding affinity for 20 proteins in the human bromodomain family. Based on this comparison, it was determined that effective inhibitors have been developed for the majority of bromodomains, in the range from 10 to 100 nM. However, we predict that more potent inhibitors can be developed for the bromodomains BPTF and BRD7 with relative ease, but that further efforts to develop inhibitors for ATAD2 will be extremely challenging. We have also made predictions for the 14 bromodomains with no reported small molecule K d values by isothermal titration calorimetry. The calculations predict that PBRM1(1) will be a challenging target, while others such as TAF1L(2), PBRM1(4) and TAF1(2), should be highly ligandable. As an outcome of this work, we assembled a database of experimental maximal K d that can serve as a community resource assisting medicinal chemistry efforts focused on BRDs. Effective prediction of ligandability would be a very useful tool in the drug discovery process.

  4. Exploring the role of water in molecular recognition: predicting protein ligandability using a combinatorial search of surface hydration sites.

    Science.gov (United States)

    Vukovic, Sinisa; Brennan, Paul E; Huggins, David J

    2016-09-01

    The interaction between any two biological molecules must compete with their interaction with water molecules. This makes water the most important molecule in medicine, as it controls the interactions of every therapeutic with its target. A small molecule binding to a protein is able to recognize a unique binding site on a protein by displacing bound water molecules from specific hydration sites. Quantifying the interactions of these water molecules allows us to estimate the potential of the protein to bind a small molecule. This is referred to as ligandability. In the study, we describe a method to predict ligandability by performing a search of all possible combinations of hydration sites on protein surfaces. We predict ligandability as the summed binding free energy for each of the constituent hydration sites, computed using inhomogeneous fluid solvation theory. We compared the predicted ligandability with the maximum observed binding affinity for 20 proteins in the human bromodomain family. Based on this comparison, it was determined that effective inhibitors have been developed for the majority of bromodomains, in the range from 10 to 100 nM. However, we predict that more potent inhibitors can be developed for the bromodomains BPTF and BRD7 with relative ease, but that further efforts to develop inhibitors for ATAD2 will be extremely challenging. We have also made predictions for the 14 bromodomains with no reported small molecule K d values by isothermal titration calorimetry. The calculations predict that PBRM1(1) will be a challenging target, while others such as TAF1L(2), PBRM1(4) and TAF1(2), should be highly ligandable. As an outcome of this work, we assembled a database of experimental maximal K d that can serve as a community resource assisting medicinal chemistry efforts focused on BRDs. Effective prediction of ligandability would be a very useful tool in the drug discovery process.

  5. Deep learning methods for protein torsion angle prediction.

    Science.gov (United States)

    Li, Haiou; Hou, Jie; Adhikari, Badri; Lyu, Qiang; Cheng, Jianlin

    2017-09-18

    Deep learning is one of the most powerful machine learning methods that has achieved the state-of-the-art performance in many domains. Since deep learning was introduced to the field of bioinformatics in 2012, it has achieved success in a number of areas such as protein residue-residue contact prediction, secondary structure prediction, and fold recognition. In this work, we developed deep learning methods to improve the prediction of torsion (dihedral) angles of proteins. We design four different deep learning architectures to predict protein torsion angles. The architectures including deep neural network (DNN) and deep restricted Boltzmann machine (DRBN), deep recurrent neural network (DRNN) and deep recurrent restricted Boltzmann machine (DReRBM) since the protein torsion angle prediction is a sequence related problem. In addition to existing protein features, two new features (predicted residue contact number and the error distribution of torsion angles extracted from sequence fragments) are used as input to each of the four deep learning architectures to predict phi and psi angles of protein backbone. The mean absolute error (MAE) of phi and psi angles predicted by DRNN, DReRBM, DRBM and DNN is about 20-21° and 29-30° on an independent dataset. The MAE of phi angle is comparable to the existing methods, but the MAE of psi angle is 29°, 2° lower than the existing methods. On the latest CASP12 targets, our methods also achieved the performance better than or comparable to a state-of-the art method. Our experiment demonstrates that deep learning is a valuable method for predicting protein torsion angles. The deep recurrent network architecture performs slightly better than deep feed-forward architecture, and the predicted residue contact number and the error distribution of torsion angles extracted from sequence fragments are useful features for improving prediction accuracy.

  6. Prediction of Lunar- and Martian-Based Intra- and Site-to-Site Task Performance.

    Science.gov (United States)

    Ade, Carl J; Broxterman, Ryan M; Craig, Jesse C; Schlup, Susanna J; Wilcox, Samuel L; Warren, Steve; Kuehl, Phillip; Gude, Dana; Jia, Chen; Barstow, Thomas J

    2016-04-01

    This study aimed to investigate the feasibility of determining the physiological parameters associated with the ability to complete simulated exploration type tasks at metabolic rates which might be expected for lunar and Martian ambulation. Running V̇O2max and gas exchange threshold (GET) were measured in 21 volunteers. Two simulated extravehicular activity field tests were completed in 1 G in regular athletic apparel at two intensities designed to elicit metabolic rates of ∼20.0 and ∼30.0 ml · kg(-1) · min(-1), which are similar to those previously reported for ambulation in simulated lunar- and Martian-based environments, respectively. All subjects were able to complete the field test at the lunar intensity, but 28% were unable to complete the field test at the Martian intensity (non-Finishers). During the Martian field test there were no differences in V̇O2 between Finishers and non-Finishers, but the non-Finishers achieved a greater %V̇O2max compared to Finishers (78.4 ± 4.6% vs. 64.9 ± 9.6%). Logistic regression analysis revealed fitness thresholds for a predicted probability of 0.5, at which Finishing and non-Finishing are equally likely, and 0.75, at which an individual has a 75% chance of Finishing, to be a V̇O2max of 38.4 ml · kg(-1) · min(-1) and 40.0 ml · kg(-1) · min(-1) or a GET of 20.1 ml · kg(-1) · min(-1) and 25.1 ml · kg(-1) · min(-1), respectively (χ(2) = 10.2). Logistic regression analysis also revealed that the expected %V̇O2max required to complete a field test could be used to successfully predict performance (χ(2) = 19.3). The results of the present investigation highlight the potential utility of V̇O2max, particularly as it relates to the metabolic demands of a surface ambulation, in defining successful completion of planetary-based exploration field tests.

  7. Site investigations. Investigation methods and general execution programme

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2001-01-01

    characterization of at least one deep chemistry-prioritized cored borehole, and start of long-term monitoring of chemical parameters in new selected sampling points. Fracture-filling mineral investigations are initiated during the final phase of the initial site investigation. The transport properties of the rock are estimated mainly on the basis of the hydrogeological and hydrogeochemical description, combined with generic, non-site-specific information. Furthermore, supplementary measurements of groundwater flow are performed in one of the first deep boreholes. In cases where mineralogy and/or groundwater chemistry differs significantly from the generic database, certain time-consuming laboratory investigations such as through diffusion measurements will be initiated. A large number of methods are used to investigate the geology of the site. (abstract truncated)

  8. Novel immunohistochemistry-based signatures to predict metastatic site of triple-negative breast cancers.

    Science.gov (United States)

    Klimov, Sergey; Rida, Padmashree Cg; Aleskandarany, Mohammed A; Green, Andrew R; Ellis, Ian O; Janssen, Emiel Am; Rakha, Emad A; Aneja, Ritu

    2017-09-05

    Although distant metastasis (DM) in breast cancer (BC) is the most lethal form of recurrence and the most common underlying cause of cancer related deaths, the outcome following the development of DM is related to the site of metastasis. Triple negative BC (TNBC) is an aggressive form of BC characterised by early recurrences and high mortality. Athough multiple variables can be used to predict the risk of metastasis, few markers can predict the specific site of metastasis. This study aimed at identifying a biomarker signature to predict particular sites of DM in TNBC. A clinically annotated series of 322 TNBC were immunohistochemically stained with 133 biomarkers relevant to BC, to develop multibiomarker models for predicting metastasis to the bone, liver, lung and brain. Patients who experienced metastasis to each site were compared with those who did not, by gradually filtering the biomarker set via a two-tailed t-test and Cox univariate analyses. Biomarker combinations were finally ranked based on statistical significance, and evaluated in multivariable analyses. Our final models were able to stratify TNBC patients into high risk groups that showed over 5, 6, 7 and 8 times higher risk of developing metastasis to the bone, liver, lung and brain, respectively, than low-risk subgroups. These models for predicting site-specific metastasis retained significance following adjustment for tumour size, patient age and chemotherapy status. Our novel IHC-based biomarkers signatures, when assessed in primary TNBC tumours, enable prediction of specific sites of metastasis, and potentially unravel biomarkers previously unknown in site tropism.

  9. Prediction of Poly(A Sites by Poly(A Read Mapping.

    Directory of Open Access Journals (Sweden)

    Thomas Bonfert

    Full Text Available RNA-seq reads containing part of the poly(A tail of transcripts (denoted as poly(A reads provide the most direct evidence for the position of poly(A sites in the genome. However, due to reduced coverage of poly(A tails by reads, poly(A reads are not routinely identified during RNA-seq mapping. Nevertheless, recent studies for several herpesviruses successfully employed mapping of poly(A reads to identify herpesvirus poly(A sites using different strategies and customized programs. To more easily allow such analyses without requiring additional programs, we integrated poly(A read mapping and prediction of poly(A sites into our RNA-seq mapping program ContextMap 2. The implemented approach essentially generalizes previously used poly(A read mapping approaches and combines them with the context-based approach of ContextMap 2 to take into account information provided by other reads aligned to the same location. Poly(A read mapping using ContextMap 2 was evaluated on real-life data from the ENCODE project and compared against a competing approach based on transcriptome assembly (KLEAT. This showed high positive predictive value for our approach, evidenced also by the presence of poly(A signals, and considerably lower runtime than KLEAT. Although sensitivity is low for both methods, we show that this is in part due to a high extent of spurious results in the gold standard set derived from RNA-PET data. Sensitivity improves for poly(A sites of known transcripts or determined with a more specific poly(A sequencing protocol and increases with read coverage on transcript ends. Finally, we illustrate the usefulness of the approach in a high read coverage scenario by a re-analysis of published data for herpes simplex virus 1. Thus, with current trends towards increasing sequencing depth and read length, poly(A read mapping will prove to be increasingly useful and can now be performed automatically during RNA-seq mapping with ContextMap 2.

  10. Intercomparison of Different Energy Prediction Methods Within the European Project "Performance" - Results of the 1st Round Robin

    NARCIS (Netherlands)

    Friesen, G.; Gottschalg, R.; Beyer, H.G.; Williams, S.R.; van Sark, W.G.J.H.M.; Guérin de Montgareuil, A.; van der Borg, N; Huld, T.; Müller, B.; de Keizer, A.C.; Niu, Y.

    2007-01-01

    Eight separate energy prediction methods, developed independently across European Universities and Research Centres, have been compared with respect to their estimated DC energy generation for five different photovoltaic (PV) module technologies and 7 different sites distributed over whole Europe.

  11. The Bolmen tunnel project - evaluation of geophysical site investigation methods

    International Nuclear Information System (INIS)

    Stanfors, R.

    1987-12-01

    The report presents geophysical measurements along and adjacent to the tunnel and an evaluation of the ability of the various methods to permit prediction of rock mass parameters of significance to stability and water bearing ability. The evaluation shows that, using airborne electro-magnetic surveys, it was possible to indicate about 80% of alla the zones of weakness more than 50 m wide in the tunnel. Airborne magnetic surveys located about 90% of all dolerite dykes more than 10 m wide. Ground-level VLF and Slingram methods of electro-magnetic measurement indicated 75% and 85% respectively of all zones of weakness more than 50 m wide. Resistivity methods were successfully used to locate clay filled and water-bearing fracture zones. About 75% of the length of tunnel over which resistivity values below 500 ohm m were measured required shotcrete support and pre-grouting. (orig./DG)

  12. SITE-94. Discrete-feature modelling of the Aespoe Site: 3. Predictions of hydrogeological parameters for performance assessment

    International Nuclear Information System (INIS)

    Geier, J.E.

    1996-12-01

    A 3-dimensional, discrete-feature hydrological model is developed. The model integrates structural and hydrologic data for the Aespoe site, on scales ranging from semi regional fracture zones to individual fractures in the vicinity of the nuclear waste canisters. Predicted parameters for the near field include fracture spacing, fracture aperture, and Darcy velocity at each of forty canister deposition holes. Parameters for the far field include discharge location, Darcy velocity, effective longitudinal dispersion coefficient and head gradient, flow porosity, and flow wetted surface, for each canister source that discharges to the biosphere. Results are presented in the form of statistical summaries for a total of 42 calculation cases, which treat a set of 25 model variants in various combinations. The variants for the SITE-94 Reference Case model address conceptual and parametric uncertainty related to the site-scale hydrogeologic model and its properties, the fracture network within the repository, effective semi regional boundary conditions for the model, and the disturbed-rock zone around the repository tunnels and shafts. Two calculation cases simulate hydrologic conditions that are predicted to occur during future glacial episodes. 30 refs

  13. Selection of Prediction Methods for Thermophysical Properties for Process Modeling and Product Design of Biodiesel Manufacturing

    DEFF Research Database (Denmark)

    Su, Yung-Chieh; Liu, Y. A.; Díaz Tovar, Carlos Axel

    2011-01-01

    To optimize biodiesel manufacturing, many reported studies have built simulation models to quantify the relationship between operating conditions and process performance. For mass and energy balance simulations, it is essential to know the four fundamental thermophysical properties of the feed oil...... prediction methods on our group Web site (www.design.che.vt.edu) for the reader to download without charge....

  14. A Novel Computational Method for Detecting DNA Methylation Sites with DNA Sequence Information and Physicochemical Properties

    Directory of Open Access Journals (Sweden)

    Gaofeng Pan

    2018-02-01

    Full Text Available DNA methylation is an important biochemical process, and it has a close connection with many types of cancer. Research about DNA methylation can help us to understand the regulation mechanism and epigenetic reprogramming. Therefore, it becomes very important to recognize the methylation sites in the DNA sequence. In the past several decades, many computational methods—especially machine learning methods—have been developed since the high-throughout sequencing technology became widely used in research and industry. In order to accurately identify whether or not a nucleotide residue is methylated under the specific DNA sequence context, we propose a novel method that overcomes the shortcomings of previous methods for predicting methylation sites. We use k-gram, multivariate mutual information, discrete wavelet transform, and pseudo amino acid composition to extract features, and train a sparse Bayesian learning model to do DNA methylation prediction. Five criteria—area under the receiver operating characteristic curve (AUC, Matthew’s correlation coefficient (MCC, accuracy (ACC, sensitivity (SN, and specificity—are used to evaluate the prediction results of our method. On the benchmark dataset, we could reach 0.8632 on AUC, 0.8017 on ACC, 0.5558 on MCC, and 0.7268 on SN. Additionally, the best results on two scBS-seq profiled mouse embryonic stem cells datasets were 0.8896 and 0.9511 by AUC, respectively. When compared with other outstanding methods, our method surpassed them on the accuracy of prediction. The improvement of AUC by our method compared to other methods was at least 0.0399 . For the convenience of other researchers, our code has been uploaded to a file hosting service, and can be downloaded from: https://figshare.com/s/0697b692d802861282d3.

  15. Efficacy of GPS cluster analysis for predicting carnivory sites of a wide-ranging omnivore: the American black bear

    Science.gov (United States)

    Kindschuh, Sarah R.; Cain, James W.; Daniel, David; Peyton, Mark A.

    2016-01-01

    The capacity to describe and quantify predation by large carnivores expanded considerably with the advent of GPS technology. Analyzing clusters of GPS locations formed by carnivores facilitates the detection of predation events by identifying characteristics which distinguish predation sites. We present a performance assessment of GPS cluster analysis as applied to the predation and scavenging of an omnivore, the American black bear (Ursus americanus), on ungulate prey and carrion. Through field investigations of 6854 GPS locations from 24 individual bears, we identified 54 sites where black bears formed a cluster of locations while predating or scavenging elk (Cervus elaphus), mule deer (Odocoileus hemionus), or cattle (Bos spp.). We developed models for three data sets to predict whether a GPS cluster was formed at a carnivory site vs. a non-carnivory site (e.g., bed sites or non-ungulate foraging sites). Two full-season data sets contained GPS locations logged at either 3-h or 30-min intervals from April to November, and a third data set contained 30-min interval data from April through July corresponding to the calving period for elk. Longer fix intervals resulted in the detection of fewer carnivory sites. Clusters were more likely to be carnivory sites if they occurred in open or edge habitats, if they occurred in the early season, if the mean distance between all pairs of GPS locations within the cluster was less, and if the cluster endured for a longer period of time. Clusters were less likely to be carnivory sites if they were initiated in the morning or night compared to the day. The top models for each data set performed well and successfully predicted 71–96% of field-verified carnivory events, 55–75% of non–carnivory events, and 58–76% of clusters overall. Refinement of this method will benefit from further application across species and ecological systems.

  16. Applicability of deterministic methods in seismic site effects modeling

    International Nuclear Information System (INIS)

    Cioflan, C.O.; Radulian, M.; Apostol, B.F.; Ciucu, C.

    2005-01-01

    The up-to-date information related to local geological structure in the Bucharest urban area has been integrated in complex analyses of the seismic ground motion simulation using deterministic procedures. The data recorded for the Vrancea intermediate-depth large earthquakes are supplemented with synthetic computations all over the city area. The hybrid method with a double-couple seismic source approximation and a relatively simple regional and local structure models allows a satisfactory reproduction of the strong motion records in the frequency domain (0.05-1)Hz. The new geological information and a deterministic analytical method which combine the modal summation technique, applied to model the seismic wave propagation between the seismic source and the studied sites, with the mode coupling approach used to model the seismic wave propagation through the local sedimentary structure of the target site, allows to extend the modelling to higher frequencies of earthquake engineering interest. The results of these studies (synthetic time histories of the ground motion parameters, absolute and relative response spectra etc) for the last 3 Vrancea strong events (August 31,1986 M w =7.1; May 30,1990 M w = 6.9 and October 27, 2004 M w = 6.0) can complete the strong motion database used for the microzonation purposes. Implications and integration of the deterministic results into the urban planning and disaster management strategies are also discussed. (authors)

  17. Life prediction methods for the combined creep-fatigue endurance

    International Nuclear Information System (INIS)

    Wareing, J.; Lloyd, G.J.

    1980-09-01

    The basis and current status of development of the various approaches to the prediction of the combined creep-fatigue endurance are reviewed. It is concluded that an inadequate materials data base makes it difficult to draw sensible conclusions about the prediction capabilities of each of the available methods. Correlation with data for stainless steel 304 and 316 is presented. (U.K.)

  18. AthMethPre: a web server for the prediction and query of mRNA m6A sites in Arabidopsis thaliana.

    Science.gov (United States)

    Xiang, Shunian; Yan, Zhangming; Liu, Ke; Zhang, Yaou; Sun, Zhirong

    2016-10-18

    N 6 -Methyladenosine (m 6 A) is the most prevalent and abundant modification in mRNA that has been linked to many key biological processes. High-throughput experiments have generated m 6 A-peaks across the transcriptome of A. thaliana, but the specific methylated sites were not assigned, which impedes the understanding of m 6 A functions in plants. Therefore, computational prediction of mRNA m 6 A sites becomes emergently important. Here, we present a method to predict the m 6 A sites for A. thaliana mRNA sequence(s). To predict the m 6 A sites of an mRNA sequence, we employed the support vector machine to build a classifier using the features of the positional flanking nucleotide sequence and position-independent k-mer nucleotide spectrum. Our method achieved good performance and was applied to a web server to provide service for the prediction of A. thaliana m 6 A sites. The server also provides a comprehensive database of predicted transcriptome-wide m 6 A sites and curated m 6 A-seq peaks from the literature for query and visualization. The AthMethPre web server is the first web server that provides a user-friendly tool for the prediction and query of A. thaliana mRNA m 6 A sites, which is freely accessible for public use at .

  19. Establishing a predictive maintenance (PdM) program at the Hanford Site

    International Nuclear Information System (INIS)

    Murray, W.A.; Winslow, R.G.

    1994-02-01

    The production reactors have been shut down for some time. But for the rest of the site, there is currently about 16,000 people engaged in a multi-billion dollar effort to safely process wastes which have been stored at the site since the 1940's. This effort also includes demolition of some older facilities and environmental restoration of much of the site. This is expected to take approximately 30 to 40 years. The concept of a site-wide predictive maintenance (PdM) program began to form in early 1993. Several informal studies showed that the stand alone predictive maintenance groups which had prevailed on site to date were less than 15% effective at trending equipment conditions and predicting failures. To improve the effectiveness of PdM within the company, an engineering analysis by Rick Winslow confirmed that utilization of software networking technology which was now available would significantly overcome many of these built in handicaps. A site-wide predictive maintenance network would make PdM technology accessible to all of the areas and facilities at the site regardless of geographical distances and company division lines. Site resident vibration experts can easily be located and provide consultations on the network. However, it was recognized that strong leadership and management skills would be required within each of the two organizations for effective implementation. To start this process, a letter of understanding and agreement between the facilities and Tank Farm divisions was drafted and endorsed by company management. The agreement assigned the primary responsibility of acquiring the network software and licensee to the Tank Farms division. The acquisition and installation of the network server would be the responsibility of the facilities division. This paper describes the rest of the network development and implementation process

  20. Prediction of suspended-sediment concentrations at selected sites in the Fountain Creek watershed, Colorado, 2008-09

    Science.gov (United States)

    Stogner, Sr., Robert W.; Nelson, Jonathan M.; McDonald, Richard R.; Kinzel, Paul J.; Mau, David P.

    2013-01-01

    In 2008, the U.S. Geological Survey (USGS), in cooperation with Pikes Peak Area Council of Governments, Colorado Water Conservation Board, Colorado Springs City Engineering, and the Lower Arkansas Valley Water Conservancy District, began a small-scale pilot study to evaluate the effectiveness of the use of a computational model of streamflow and suspended-sediment transport for predicting suspended-sediment concentrations and loads in the Fountain Creek watershed in Colorado. Increased erosion and sedimentation damage have been identified by the Fountain Creek Watershed Plan as key problems within the watershed. A recommendation in the Fountain Creek Watershed plan for management of the basin is to establish measurable criteria to determine if progress in reducing erosion and sedimentation damage is being made. The major objective of this study was to test a computational method to predict local suspended-sediment loads at two sites with different geomorphic characteristics in order to evaluate the feasibility of using such an approach to predict local suspended-sediment loads throughout the entire watershed. Detailed topographic surveys, particle-size data, and suspended-sediment samples were collected at two gaged sites: Monument Creek above Woodmen Road at Colorado Springs, Colorado (USGS gage 07103970), and Sand Creek above mouth at Colorado Springs, Colorado (USGS gage 07105600). These data were used to construct three-dimensional computational models of relatively short channel reaches at each site. The streamflow component of these models predicted a spatially distributed field of water-surface elevation, water velocity, and bed shear stress for a range of stream discharges. Using the model predictions, along with measured particle sizes, the sediment-transport component of the model predicted the suspended-sediment concentration throughout the reach of interest. These computed concentrations were used with predicted flow patterns and channel morphology to

  1. What Predicts Use of Learning-Centered, Interactive Engagement Methods?

    Science.gov (United States)

    Madson, Laura; Trafimow, David; Gray, Tara; Gutowitz, Michael

    2014-01-01

    What makes some faculty members more likely to use interactive engagement methods than others? We use the theory of reasoned action to predict faculty members' use of interactive engagement methods. Results indicate that faculty members' beliefs about the personal positive consequences of using these methods (e.g., "Using interactive…

  2. Method for Predicting Solubilities of Solids in Mixed Solvents

    DEFF Research Database (Denmark)

    Ellegaard, Martin Dela; Abildskov, Jens; O'Connell, J. P.

    2009-01-01

    A method is presented for predicting solubilities of solid solutes in mixed solvents, based on excess Henry's law constants. The basis is statistical mechanical fluctuation solution theory for composition derivatives of solute/solvent infinite dilution activity coefficients. Suitable approximatio...

  3. Fast Prediction Method for Steady-State Heat Convection

    KAUST Repository

    Wá ng, Yì ; Yu, Bo; Sun, Shuyu

    2012-01-01

    , the nonuniform POD-Galerkin projection method exhibits high accuracy, good suitability, and fast computation. It has universal significance for accurate and fast prediction. Also, the methodology can be applied to more complex modeling in chemical engineering

  4. Development of motion image prediction method using principal component analysis

    International Nuclear Information System (INIS)

    Chhatkuli, Ritu Bhusal; Demachi, Kazuyuki; Kawai, Masaki; Sakakibara, Hiroshi; Kamiaka, Kazuma

    2012-01-01

    Respiratory motion can induce the limit in the accuracy of area irradiated during lung cancer radiation therapy. Many methods have been introduced to minimize the impact of healthy tissue irradiation due to the lung tumor motion. The purpose of this research is to develop an algorithm for the improvement of image guided radiation therapy by the prediction of motion images. We predict the motion images by using principal component analysis (PCA) and multi-channel singular spectral analysis (MSSA) method. The images/movies were successfully predicted and verified using the developed algorithm. With the proposed prediction method it is possible to forecast the tumor images over the next breathing period. The implementation of this method in real time is believed to be significant for higher level of tumor tracking including the detection of sudden abdominal changes during radiation therapy. (author)

  5. PROXIMAL: a method for Prediction of Xenobiotic Metabolism.

    Science.gov (United States)

    Yousofshahi, Mona; Manteiga, Sara; Wu, Charmian; Lee, Kyongbum; Hassoun, Soha

    2015-12-22

    Contamination of the environment with bioactive chemicals has emerged as a potential public health risk. These substances that may cause distress or disease in humans can be found in air, water and food supplies. An open question is whether these chemicals transform into potentially more active or toxic derivatives via xenobiotic metabolizing enzymes expressed in the body. We present a new prediction tool, which we call PROXIMAL (Prediction of Xenobiotic Metabolism) for identifying possible transformation products of xenobiotic chemicals in the liver. Using reaction data from DrugBank and KEGG, PROXIMAL builds look-up tables that catalog the sites and types of structural modifications performed by Phase I and Phase II enzymes. Given a compound of interest, PROXIMAL searches for substructures that match the sites cataloged in the look-up tables, applies the corresponding modifications to generate a panel of possible transformation products, and ranks the products based on the activity and abundance of the enzymes involved. PROXIMAL generates transformations that are specific for the chemical of interest by analyzing the chemical's substructures. We evaluate the accuracy of PROXIMAL's predictions through case studies on two environmental chemicals with suspected endocrine disrupting activity, bisphenol A (BPA) and 4-chlorobiphenyl (PCB3). Comparisons with published reports confirm 5 out of 7 and 17 out of 26 of the predicted derivatives for BPA and PCB3, respectively. We also compare biotransformation predictions generated by PROXIMAL with those generated by METEOR and Metaprint2D-react, two other prediction tools. PROXIMAL can predict transformations of chemicals that contain substructures recognizable by human liver enzymes. It also has the ability to rank the predicted metabolites based on the activity and abundance of enzymes involved in xenobiotic transformation.

  6. An assessment on epitope prediction methods for protozoa genomes

    Directory of Open Access Journals (Sweden)

    Resende Daniela M

    2012-11-01

    Full Text Available Abstract Background Epitope prediction using computational methods represents one of the most promising approaches to vaccine development. Reduction of time, cost, and the availability of completely sequenced genomes are key points and highly motivating regarding the use of reverse vaccinology. Parasites of genus Leishmania are widely spread and they are the etiologic agents of leishmaniasis. Currently, there is no efficient vaccine against this pathogen and the drug treatment is highly toxic. The lack of sufficiently large datasets of experimentally validated parasites epitopes represents a serious limitation, especially for trypanomatids genomes. In this work we highlight the predictive performances of several algorithms that were evaluated through the development of a MySQL database built with the purpose of: a evaluating individual algorithms prediction performances and their combination for CD8+ T cell epitopes, B-cell epitopes and subcellular localization by means of AUC (Area Under Curve performance and a threshold dependent method that employs a confusion matrix; b integrating data from experimentally validated and in silico predicted epitopes; and c integrating the subcellular localization predictions and experimental data. NetCTL, NetMHC, BepiPred, BCPred12, and AAP12 algorithms were used for in silico epitope prediction and WoLF PSORT, Sigcleave and TargetP for in silico subcellular localization prediction against trypanosomatid genomes. Results A database-driven epitope prediction method was developed with built-in functions that were capable of: a removing experimental data redundancy; b parsing algorithms predictions and storage experimental validated and predict data; and c evaluating algorithm performances. Results show that a better performance is achieved when the combined prediction is considered. This is particularly true for B cell epitope predictors, where the combined prediction of AAP12 and BCPred12 reached an AUC value

  7. Refinement of a Method for Identifying Probable Archaeological Sites from Remotely Sensed Data

    Science.gov (United States)

    Tilton, James C.; Comer, Douglas C.; Priebe, Carey E.; Sussman, Daniel; Chen, Li

    2012-01-01

    To facilitate locating archaeological sites before they are compromised or destroyed, we are developing approaches for generating maps of probable archaeological sites, through detecting subtle anomalies in vegetative cover, soil chemistry, and soil moisture by analyzing remotely sensed data from multiple sources. We previously reported some success in this effort with a statistical analysis of slope, radar, and Ikonos data (including tasseled cap and NDVI transforms) with Student's t-test. We report here on new developments in our work, performing an analysis of 8-band multispectral Worldview-2 data. The Worldview-2 analysis begins by computing medians and median absolute deviations for the pixels in various annuli around each site of interest on the 28 band difference ratios. We then use principle components analysis followed by linear discriminant analysis to train a classifier which assigns a posterior probability that a location is an archaeological site. We tested the procedure using leave-one-out cross validation with a second leave-one-out step to choose parameters on a 9,859x23,000 subset of the WorldView-2 data over the western portion of Ft. Irwin, CA, USA. We used 100 known non-sites and trained one classifier for lithic sites (n=33) and one classifier for habitation sites (n=16). We then analyzed convex combinations of scores from the Archaeological Predictive Model (APM) and our scores. We found that that the combined scores had a higher area under the ROC curve than either individual method, indicating that including WorldView-2 data in analysis improved the predictive power of the provided APM.

  8. Assessment of a method for the prediction of mandibular rotation.

    Science.gov (United States)

    Lee, R S; Daniel, F J; Swartz, M; Baumrind, S; Korn, E L

    1987-05-01

    A new method to predict mandibular rotation developed by Skieller and co-workers on a sample of 21 implant subjects with extreme growth patterns has been tested against an alternative sample of 25 implant patients with generally similar mean values, but with less extreme facial patterns. The method, which had been highly successful in retrospectively predicting changes in the sample of extreme subjects, was much less successful in predicting individual patterns of mandibular rotation in the new, less extreme sample. The observation of a large difference in the strength of the predictions for these two samples, even though their mean values were quite similar, should serve to increase our awareness of the complexity of the problem of predicting growth patterns in individual cases.

  9. Pharmacological considerations for predicting PK/PD at the site of action for therapeutic proteins.

    Science.gov (United States)

    Wang, Weirong; Zhou, Honghui

    For therapeutic proteins whose sites of action are distal to the systemic circulation, both drug and target concentrations at the tissue sites are not necessarily proportional to those in systemic circulation, highlighting the importance of understanding pharmacokinetic/pharmacodynamic (PK/PD) relationship at the sites of action. This review summarizes the pharmacological considerations for predicting local PK/PD and the importance of measuring PK and PD at site of action. Three case examples are presented to show how mechanistic and physiologically based PK/PD (PBPK/PD) models which incorporated the PK and PD at the tissue site can be used to facilitate understanding the exposure-response relationship for therapeutic proteins. Copyright © 2016. Published by Elsevier Ltd.

  10. Predicting transcription factor binding sites using local over-representation and comparative genomics

    Directory of Open Access Journals (Sweden)

    Touzet Hélène

    2006-08-01

    Full Text Available Abstract Background Identifying cis-regulatory elements is crucial to understanding gene expression, which highlights the importance of the computational detection of overrepresented transcription factor binding sites (TFBSs in coexpressed or coregulated genes. However, this is a challenging problem, especially when considering higher eukaryotic organisms. Results We have developed a method, named TFM-Explorer, that searches for locally overrepresented TFBSs in a set of coregulated genes, which are modeled by profiles provided by a database of position weight matrices. The novelty of the method is that it takes advantage of spatial conservation in the sequence and supports multiple species. The efficiency of the underlying algorithm and its robustness to noise allow weak regulatory signals to be detected in large heterogeneous data sets. Conclusion TFM-Explorer provides an efficient way to predict TFBS overrepresentation in related sequences. Promising results were obtained in a variety of examples in human, mouse, and rat genomes. The software is publicly available at http://bioinfo.lifl.fr/TFM-Explorer.

  11. Prediction of transcriptional regulatory sites in the complete genome sequence of Escherichia coli K-12.

    Science.gov (United States)

    Thieffry, D; Salgado, H; Huerta, A M; Collado-Vides, J

    1998-06-01

    As one of the best-characterized free-living organisms, Escherichia coli and its recently completed genomic sequence offer a special opportunity to exploit systematically the variety of regulatory data available in the literature in order to make a comprehensive set of regulatory predictions in the whole genome. The complete genome sequence of E.coli was analyzed for the binding of transcriptional regulators upstream of coding sequences. The biological information contained in RegulonDB (Huerta, A.M. et al., Nucleic Acids Res.,26,55-60, 1998) for 56 different transcriptional proteins was the support to implement a stringent strategy combining string search and weight matrices. We estimate that our search included representatives of 15-25% of the total number of regulatory binding proteins in E.coli. This search was performed on the set of 4288 putative regulatory regions, each 450 bp long. Within the regions with predicted sites, 89% are regulated by one protein and 81% involve only one site. These numbers are reasonably consistent with the distribution of experimental regulatory sites. Regulatory sites are found in 603 regions corresponding to 16% of operon regions and 10% of intra-operonic regions. Additional evidence gives stronger support to some of these predictions, including the position of the site, biological consistency with the function of the downstream gene, as well as genetic evidence for the regulatory interaction. The predictions described here were incorporated into the map presented in the paper describing the complete E.coli genome (Blattner,F.R. et al., Science, 277, 1453-1461, 1997). The complete set of predictions in GenBank format is available at the url: http://www. cifn.unam.mx/Computational_Biology/E.coli-predictions ecoli-reg@cifn.unam.mx, collado@cifn.unam.mx

  12. Assessing the model transferability for prediction of transcription factor binding sites based on chromatin accessibility.

    Science.gov (United States)

    Liu, Sheng; Zibetti, Cristina; Wan, Jun; Wang, Guohua; Blackshaw, Seth; Qian, Jiang

    2017-07-27

    Computational prediction of transcription factor (TF) binding sites in different cell types is challenging. Recent technology development allows us to determine the genome-wide chromatin accessibility in various cellular and developmental contexts. The chromatin accessibility profiles provide useful information in prediction of TF binding events in various physiological conditions. Furthermore, ChIP-Seq analysis was used to determine genome-wide binding sites for a range of different TFs in multiple cell types. Integration of these two types of genomic information can improve the prediction of TF binding events. We assessed to what extent a model built upon on other TFs and/or other cell types could be used to predict the binding sites of TFs of interest. A random forest model was built using a set of cell type-independent features such as specific sequences recognized by the TFs and evolutionary conservation, as well as cell type-specific features derived from chromatin accessibility data. Our analysis suggested that the models learned from other TFs and/or cell lines performed almost as well as the model learned from the target TF in the cell type of interest. Interestingly, models based on multiple TFs performed better than single-TF models. Finally, we proposed a universal model, BPAC, which was generated using ChIP-Seq data from multiple TFs in various cell types. Integrating chromatin accessibility information with sequence information improves prediction of TF binding.The prediction of TF binding is transferable across TFs and/or cell lines suggesting there are a set of universal "rules". A computational tool was developed to predict TF binding sites based on the universal "rules".

  13. Performance prediction method for a multi-stage Knudsen pump

    Science.gov (United States)

    Kugimoto, K.; Hirota, Y.; Kizaki, Y.; Yamaguchi, H.; Niimi, T.

    2017-12-01

    In this study, the novel method to predict the performance of a multi-stage Knudsen pump is proposed. The performance prediction method is carried out in two steps numerically with the assistance of a simple experimental result. In the first step, the performance of a single-stage Knudsen pump was measured experimentally under various pressure conditions, and the relationship of the mass flow rate was obtained with respect to the average pressure between the inlet and outlet of the pump and the pressure difference between them. In the second step, the performance of a multi-stage pump was analyzed by a one-dimensional model derived from the mass conservation law. The performances predicted by the 1D-model of 1-stage, 2-stage, 3-stage, and 4-stage pumps were validated by the experimental results for the corresponding number of stages. It was concluded that the proposed prediction method works properly.

  14. Splice site prediction in Arabidopsis thaliana pre-mRNA by combining local and global sequence information

    DEFF Research Database (Denmark)

    Hebsgaard, Stefan M.; Korning, Peter G.; Tolstrup, Niels

    1996-01-01

    Artificial neural networks have been combined with a rule based system to predict intron splice sites in the dicot plant Arabidopsis thaliana. A two step prediction scheme, where a global prediction of the coding potential regulates a cutoff level for a local predicition of splice sites, is refin...

  15. Predicting ambient aerosol thermal-optical reflectance (TOR) measurements from infrared spectra: extending the predictions to different years and different sites

    Science.gov (United States)

    Reggente, Matteo; Dillner, Ann M.; Takahama, Satoshi

    2016-02-01

    Organic carbon (OC) and elemental carbon (EC) are major components of atmospheric particulate matter (PM), which has been associated with increased morbidity and mortality, climate change, and reduced visibility. Typically OC and EC concentrations are measured using thermal-optical methods such as thermal-optical reflectance (TOR) from samples collected on quartz filters. In this work, we estimate TOR OC and EC using Fourier transform infrared (FT-IR) absorbance spectra from polytetrafluoroethylene (PTFE Teflon) filters using partial least square regression (PLSR) calibrated to TOR OC and EC measurements for a wide range of samples. The proposed method can be integrated with analysis of routinely collected PTFE filter samples that, in addition to OC and EC concentrations, can concurrently provide information regarding the functional group composition of the organic aerosol. We have used the FT-IR absorbance spectra and TOR OC and EC concentrations collected in the Interagency Monitoring of PROtected Visual Environments (IMPROVE) network (USA). We used 526 samples collected in 2011 at seven sites to calibrate the models, and more than 2000 samples collected in 2013 at 17 sites to test the models. Samples from six sites are present both in the calibration and test sets. The calibrations produce accurate predictions both for samples collected at the same six sites present in the calibration set (R2 = 0.97 and R2 = 0.95 for OC and EC respectively), and for samples from 9 of the 11 sites not included in the calibration set (R2 = 0.96 and R2 = 0.91 for OC and EC respectively). Samples collected at the other two sites require a different calibration model to achieve accurate predictions. We also propose a method to anticipate the prediction error; we calculate the squared Mahalanobis distance in the feature space (scores determined by PLSR) between new spectra and spectra in the calibration set. The squared Mahalanobis distance provides a crude method for assessing the

  16. Neural Network Prediction of Translation Initiation Sites in Eukaryotes: Perspectives for EST and Genome analysis

    DEFF Research Database (Denmark)

    Pedersen, Anders Gorm; Nielsen, Henrik

    1997-01-01

    Translation in eukaryotes does not always start at the first AUG in an mRNA, implying that context information also plays a role.This makes prediction of translation initiation sites a non-trivial task, especially when analysing EST and genome data where the entire mature mRNA sequence is not known...

  17. Mapping the heparin-binding site of the BMP antagonist gremlin by site-directed mutagenesis based on predictive modelling.

    Science.gov (United States)

    Tatsinkam, Arnold Junior; Mulloy, Barbara; Rider, Christopher C

    2015-08-15

    Gremlin is a member of the CAN (cerberus and DAN) family of secreted BMP (bone morphogenetic protein) antagonists and also an agonist of VEGF (vascular endothelial growth factor) receptor-2. It is critical in limb skeleton and kidney development and is re-expressed during tissue fibrosis. Gremlin binds strongly to heparin and heparan sulfate and, in the present study, we sought to investigate its heparin-binding site. In order to explore a putative non-contiguous binding site predicted by computational molecular modelling, we substituted a total of 11 key arginines and lysines located in three basic residue sequence clusters with homologous sequences from cerberus and DAN (differential screening selected gene abberative in neuroblastoma), CAN proteins which lack basic residues in these positions. A panel of six Myc-tagged gremlin mutants, MGR-1-MGR-6 (MGR, mutant gremlin), each containing different combinations of targeted substitutions, all showed markedly reduced affinity for heparin as demonstrated by their NaCl elution on heparin affinity chromatography, thus verifying our predictions. Both MGR-5 and MGR-6 retained BMP-4-binding activity comparable to that of wild-type gremlin. Low-molecular-mass heparin neither promoted nor inhibited BMP-4 binding. Finally, glutaraldehyde cross-linking demonstrated that gremlin forms non-covalent dimers, similar behaviour to that of DAN and also PRDC (protein related to cerberus and DAN), another CAN protein. The resulting dimer would possess two heparin-binding sites, each running along an exposed surface on the second β-strand finger loop of one of the monomers. © 2015 Authors; published by Portland Press Limited.

  18. Probabilistic Assessment of Above Zone Pressure Predictions at a Geologic Carbon Storage Site

    Energy Technology Data Exchange (ETDEWEB)

    Namhata, Argha; Oladyshkin, Sergey; Dilmore, Robert M.; Zhang, Liwei; Nakles, David V.

    2016-12-01

    Carbon dioxide (CO2) storage into geological formations is regarded as an important mitigation strategy for anthropogenic CO2 emissions to the atmosphere. This study first simulates the leakage of CO2 and brine from a storage reservoir through the caprock. Then, we estimate the resulting pressure changes at the zone overlying the caprock also known as Above Zone Monitoring Interval (AZMI). A data-driven approach of arbitrary Polynomial Chaos (aPC) Expansion is then used to quantify the uncertainty in the above zone pressure prediction based on the uncertainties in different geologic parameters. Finally, a global sensitivity analysis is performed with Sobol indices based on the aPC technique to determine the relative importance of different parameters on pressure prediction. The results indicate that there can be uncertainty in pressure prediction locally around the leakage zones. The degree of such uncertainty in prediction depends on the quality of site specific information available for analysis. The scientific results from this study provide substantial insight that there is a need for site-specific data for efficient predictions of risks associated with storage activities. The presented approach can provide a basis of optimized pressure based monitoring network design at carbon storage sites.

  19. Analysis and prediction of gene splice sites in four Aspergillus genomes

    DEFF Research Database (Denmark)

    Wang, Kai; Ussery, David; Brunak, Søren

    2009-01-01

    Several Aspergillus fungal genomic sequences have been published, with many more in progress. Obviously, it is essential to have high-quality, consistently annotated sets of proteins from each of the genomes, in order to make meaningful comparisons. We have developed a dedicated, publicly available......, splice site prediction program called NetAspGene, for the genus Aspergillus. Gene sequences from Aspergillus fumigatus, the most common mould pathogen, were used to build and test our model. Compared to many animals and plants, Aspergillus contains smaller introns; thus we have applied a larger window...... better splice site prediction than other available tools. NetAspGene will be very helpful for the study in Aspergillus splice sites and especially in alternative splicing. A webpage for NetAspGene is publicly available at http://www.cbs.dtu.dk/services/NetAspGene....

  20. Connecting clinical and actuarial prediction with rule-based methods.

    Science.gov (United States)

    Fokkema, Marjolein; Smits, Niels; Kelderman, Henk; Penninx, Brenda W J H

    2015-06-01

    Meta-analyses comparing the accuracy of clinical versus actuarial prediction have shown actuarial methods to outperform clinical methods, on average. However, actuarial methods are still not widely used in clinical practice, and there has been a call for the development of actuarial prediction methods for clinical practice. We argue that rule-based methods may be more useful than the linear main effect models usually employed in prediction studies, from a data and decision analytic as well as a practical perspective. In addition, decision rules derived with rule-based methods can be represented as fast and frugal trees, which, unlike main effects models, can be used in a sequential fashion, reducing the number of cues that have to be evaluated before making a prediction. We illustrate the usability of rule-based methods by applying RuleFit, an algorithm for deriving decision rules for classification and regression problems, to a dataset on prediction of the course of depressive and anxiety disorders from Penninx et al. (2011). The RuleFit algorithm provided a model consisting of 2 simple decision rules, requiring evaluation of only 2 to 4 cues. Predictive accuracy of the 2-rule model was very similar to that of a logistic regression model incorporating 20 predictor variables, originally applied to the dataset. In addition, the 2-rule model required, on average, evaluation of only 3 cues. Therefore, the RuleFit algorithm appears to be a promising method for creating decision tools that are less time consuming and easier to apply in psychological practice, and with accuracy comparable to traditional actuarial methods. (c) 2015 APA, all rights reserved).

  1. Prediction of mucin-type O-glycosylation sites in mammalian proteins using the composition of k-spaced amino acid pairs

    Directory of Open Access Journals (Sweden)

    Sheng Zhi-Ya

    2008-02-01

    Full Text Available Abstract Background As one of the most common protein post-translational modifications, glycosylation is involved in a variety of important biological processes. Computational identification of glycosylation sites in protein sequences becomes increasingly important in the post-genomic era. A new encoding scheme was employed to improve the prediction of mucin-type O-glycosylation sites in mammalian proteins. Results A new protein bioinformatics tool, CKSAAP_OGlySite, was developed to predict mucin-type O-glycosylation serine/threonine (S/T sites in mammalian proteins. Using the composition of k-spaced amino acid pairs (CKSAAP based encoding scheme, the proposed method was trained and tested in a new and stringent O-glycosylation dataset with the assistance of Support Vector Machine (SVM. When the ratio of O-glycosylation to non-glycosylation sites in training datasets was set as 1:1, 10-fold cross-validation tests showed that the proposed method yielded a high accuracy of 83.1% and 81.4% in predicting O-glycosylated S and T sites, respectively. Based on the same datasets, CKSAAP_OGlySite resulted in a higher accuracy than the conventional binary encoding based method (about +5.0%. When trained and tested in 1:5 datasets, the CKSAAP encoding showed a more significant improvement than the binary encoding. We also merged the training datasets of S and T sites and integrated the prediction of S and T sites into one single predictor (i.e. S+T predictor. Either in 1:1 or 1:5 datasets, the performance of this S+T predictor was always slightly better than those predictors where S and T sites were independently predicted, suggesting that the molecular recognition of O-glycosylated S/T sites seems to be similar and the increase of the S+T predictor's accuracy may be a result of expanded training datasets. Moreover, CKSAAP_OGlySite was also shown to have better performance when benchmarked against two existing predictors. Conclusion Because of CKSAAP

  2. The trajectory prediction of spacecraft by grey method

    International Nuclear Information System (INIS)

    Wang, Qiyue; Wang, Zhongyu; Zhang, Zili; Wang, Yanqing; Zhou, Weihu

    2016-01-01

    The real-time and high-precision trajectory prediction of a moving object is a core technology in the field of aerospace engineering. The real-time monitoring and tracking technology are also significant guarantees of aerospace equipment. A dynamic trajectory prediction method called grey dynamic filter (GDF) which combines the dynamic measurement theory and grey system theory is proposed. GDF can use coordinates of the current period to extrapolate coordinates of the following period. At meantime, GDF can also keep the instantaneity of measured coordinates by the metabolism model. In this paper the optimal model length of GDF is firstly selected to improve the prediction accuracy. Then the simulation for uniformly accelerated motion and variably accelerated motion is conducted. The simulation results indicate that the mean composite position error of GDF prediction is one-fifth to that of Kalman filter (KF). By using a spacecraft landing experiment, the prediction accuracy of GDF is compared with the KF method and the primitive grey method (GM). The results show that the motion trajectory of spacecraft predicted by GDF is much closer to actual trajectory than the other two methods. The mean composite position error calculated by GDF is one-eighth to KF and one-fifth to GM respectively. (paper)

  3. Predicting chaos in memristive oscillator via harmonic balance method.

    Science.gov (United States)

    Wang, Xin; Li, Chuandong; Huang, Tingwen; Duan, Shukai

    2012-12-01

    This paper studies the possible chaotic behaviors in a memristive oscillator with cubic nonlinearities via harmonic balance method which is also called the method of describing function. This method was proposed to detect chaos in classical Chua's circuit. We first transform the considered memristive oscillator system into Lur'e model and present the prediction of the existence of chaotic behaviors. To ensure the prediction result is correct, the distortion index is also measured. Numerical simulations are presented to show the effectiveness of theoretical results.

  4. Evaluation and comparison of mammalian subcellular localization prediction methods

    Directory of Open Access Journals (Sweden)

    Fink J Lynn

    2006-12-01

    Full Text Available Abstract Background Determination of the subcellular location of a protein is essential to understanding its biochemical function. This information can provide insight into the function of hypothetical or novel proteins. These data are difficult to obtain experimentally but have become especially important since many whole genome sequencing projects have been finished and many resulting protein sequences are still lacking detailed functional information. In order to address this paucity of data, many computational prediction methods have been developed. However, these methods have varying levels of accuracy and perform differently based on the sequences that are presented to the underlying algorithm. It is therefore useful to compare these methods and monitor their performance. Results In order to perform a comprehensive survey of prediction methods, we selected only methods that accepted large batches of protein sequences, were publicly available, and were able to predict localization to at least nine of the major subcellular locations (nucleus, cytosol, mitochondrion, extracellular region, plasma membrane, Golgi apparatus, endoplasmic reticulum (ER, peroxisome, and lysosome. The selected methods were CELLO, MultiLoc, Proteome Analyst, pTarget and WoLF PSORT. These methods were evaluated using 3763 mouse proteins from SwissProt that represent the source of the training sets used in development of the individual methods. In addition, an independent evaluation set of 2145 mouse proteins from LOCATE with a bias towards the subcellular localization underrepresented in SwissProt was used. The sensitivity and specificity were calculated for each method and compared to a theoretical value based on what might be observed by random chance. Conclusion No individual method had a sufficient level of sensitivity across both evaluation sets that would enable reliable application to hypothetical proteins. All methods showed lower performance on the LOCATE

  5. Univariate Time Series Prediction of Solar Power Using a Hybrid Wavelet-ARMA-NARX Prediction Method

    Energy Technology Data Exchange (ETDEWEB)

    Nazaripouya, Hamidreza; Wang, Yubo; Chu, Chi-Cheng; Pota, Hemanshu; Gadh, Rajit

    2016-05-02

    This paper proposes a new hybrid method for super short-term solar power prediction. Solar output power usually has a complex, nonstationary, and nonlinear characteristic due to intermittent and time varying behavior of solar radiance. In addition, solar power dynamics is fast and is inertia less. An accurate super short-time prediction is required to compensate for the fluctuations and reduce the impact of solar power penetration on the power system. The objective is to predict one step-ahead solar power generation based only on historical solar power time series data. The proposed method incorporates discrete wavelet transform (DWT), Auto-Regressive Moving Average (ARMA) models, and Recurrent Neural Networks (RNN), while the RNN architecture is based on Nonlinear Auto-Regressive models with eXogenous inputs (NARX). The wavelet transform is utilized to decompose the solar power time series into a set of richer-behaved forming series for prediction. ARMA model is employed as a linear predictor while NARX is used as a nonlinear pattern recognition tool to estimate and compensate the error of wavelet-ARMA prediction. The proposed method is applied to the data captured from UCLA solar PV panels and the results are compared with some of the common and most recent solar power prediction methods. The results validate the effectiveness of the proposed approach and show a considerable improvement in the prediction precision.

  6. Available Prediction Methods for Corrosion under Insulation (CUI): A Review

    OpenAIRE

    Burhani Nurul Rawaida Ain; Muhammad Masdi; Ismail Mokhtar Che

    2014-01-01

    Corrosion under insulation (CUI) is an increasingly important issue for the piping in industries especially petrochemical and chemical plants due to its unexpected catastrophic disaster. Therefore, attention towards the maintenance and prediction of CUI occurrence, particularly in the corrosion rates, has grown in recent years. In this study, a literature review in determining the corrosion rates by using various prediction models and method of the corrosion occurrence between the external su...

  7. Methods, apparatus and system for notification of predictable memory failure

    Energy Technology Data Exchange (ETDEWEB)

    Cher, Chen-Yong; Andrade Costa, Carlos H.; Park, Yoonho; Rosenburg, Bryan S.; Ryu, Kyung D.

    2017-01-03

    A method for providing notification of a predictable memory failure includes the steps of: obtaining information regarding at least one condition associated with a memory; calculating a memory failure probability as a function of the obtained information; calculating a failure probability threshold; and generating a signal when the memory failure probability exceeds the failure probability threshold, the signal being indicative of a predicted future memory failure.

  8. Three-dimensional protein structure prediction: Methods and computational strategies.

    Science.gov (United States)

    Dorn, Márcio; E Silva, Mariel Barbachan; Buriol, Luciana S; Lamb, Luis C

    2014-10-12

    A long standing problem in structural bioinformatics is to determine the three-dimensional (3-D) structure of a protein when only a sequence of amino acid residues is given. Many computational methodologies and algorithms have been proposed as a solution to the 3-D Protein Structure Prediction (3-D-PSP) problem. These methods can be divided in four main classes: (a) first principle methods without database information; (b) first principle methods with database information; (c) fold recognition and threading methods; and (d) comparative modeling methods and sequence alignment strategies. Deterministic computational techniques, optimization techniques, data mining and machine learning approaches are typically used in the construction of computational solutions for the PSP problem. Our main goal with this work is to review the methods and computational strategies that are currently used in 3-D protein prediction. Copyright © 2014 Elsevier Ltd. All rights reserved.

  9. Methods and techniques for prediction of environmental impact

    International Nuclear Information System (INIS)

    1992-04-01

    Environmental impact assessment (EIA) is the procedure that helps decision makers understand the environmental implications of their decisions. The prediction of environmental effects or impact is an extremely important part of the EIA procedure and improvements in existing capabilities are needed. Considerable attention is paid within environmental impact assessment and in handbooks on EIA to methods for identifying and evaluating environmental impacts. However, little attention is given to the issue distribution of information on impact prediction methods. The quantitative or qualitative methods for the prediction of environmental impacts appear to be the two basic approaches for incorporating environmental concerns into the decision-making process. Depending on the nature of the proposed activity and the environment likely to be affected, a combination of both quantitative and qualitative methods is used. Within environmental impact assessment, the accuracy of methods for the prediction of environmental impacts is of major importance while it provides for sound and well-balanced decision making. Pertinent and effective action to deal with the problems of environmental protection and the rational use of natural resources and sustainable development is only possible given objective methods and techniques for the prediction of environmental impact. Therefore, the Senior Advisers to ECE Governments on Environmental and Water Problems, decided to set up a task force, with the USSR as lead country, on methods and techniques for the prediction of environmental impacts in order to undertake a study to review and analyse existing methodological approaches and to elaborate recommendations to ECE Governments. The work of the task force was completed in 1990 and the resulting report, with all relevant background material, was approved by the Senior Advisers to ECE Governments on Environmental and Water Problems in 1991. The present report reflects the situation, state of

  10. Modified-Fibonacci-Dual-Lucas method for earthquake prediction

    Science.gov (United States)

    Boucouvalas, A. C.; Gkasios, M.; Tselikas, N. T.; Drakatos, G.

    2015-06-01

    The FDL method makes use of Fibonacci, Dual and Lucas numbers and has shown considerable success in predicting earthquake events locally as well as globally. Predicting the location of the epicenter of an earthquake is one difficult challenge the other being the timing and magnitude. One technique for predicting the onset of earthquakes is the use of cycles, and the discovery of periodicity. Part of this category is the reported FDL method. The basis of the reported FDL method is the creation of FDL future dates based on the onset date of significant earthquakes. The assumption being that each occurred earthquake discontinuity can be thought of as a generating source of FDL time series The connection between past earthquakes and future earthquakes based on FDL numbers has also been reported with sample earthquakes since 1900. Using clustering methods it has been shown that significant earthquakes (conjunct Sun, Moon opposite Sun, Moon conjunct or opposite North or South Modes. In order to test improvement of the method we used all +8R earthquakes recorded since 1900, (86 earthquakes from USGS data). We have developed the FDL numbers for each of those seeds, and examined the earthquake hit rates (for a window of 3, i.e. +-1 day of target date) and for <6.5R. The successes are counted for each one of the 86 earthquake seeds and we compare the MFDL method with the FDL method. In every case we find improvement when the starting seed date is on the planetary trigger date prior to the earthquake. We observe no improvement only when a planetary trigger coincided with the earthquake date and in this case the FDL method coincides with the MFDL. Based on the MDFL method we present the prediction method capable of predicting global events or localized earthquakes and we will discuss the accuracy of the method in as far as the prediction and location parts of the method. We show example calendar style predictions for global events as well as for the Greek region using

  11. Methods for early prediction of lactation flow in Holstein heifers

    Directory of Open Access Journals (Sweden)

    Vesna Gantner

    2010-12-01

    Full Text Available The aim of this research was to define methods for early prediction (based on I. milk control record of lactation flow in Holstein heifers as well as to choose optimal one in terms of prediction fit and application simplicity. Total of 304,569 daily yield records automatically recorded on a 1,136 first lactation Holstein cows, from March 2003 till August 2008., were included in analysis. According to the test date, calving date, the age at first calving, lactation stage when I. milk control occurred and to the average milk yield in first 25th, T1 (and 25th-45th, T2 lactation days, measuring monthcalving month-age-production-time-period subgroups were formed. The parameters of analysed nonlinear and linear methods were estimated for each defined subgroup. As models evaluation measures,adjusted coefficient of determination, and average and standard deviation of error were used. Considering obtained results, in terms of total variance explanation (R2 adj, the nonlinear Wood’s method showed superiority above the linear ones (Wilmink’s, Ali-Schaeffer’s and Guo-Swalve’s method in both time-period subgroups (T1 - 97.5 % of explained variability; T2 - 98.1 % of explained variability. Regarding the evaluation measures based on prediction error amount (eavg±eSD, the lowest average error of daily milk yield prediction (less than 0.005 kg/day, as well as of lactation milk yield prediction (less than 50 kg/lactation (T1 time-period subgroup and less than 30 kg/lactation (T2 time-period subgroup; were determined when Wood’s nonlinear prediction method were applied. Obtained results indicate that estimated Wood’s regression parameters could be used in routine work for early prediction of Holstein heifer’s lactation flow.

  12. Prediction of protein post-translational modifications: main trends and methods

    Science.gov (United States)

    Sobolev, B. N.; Veselovsky, A. V.; Poroikov, V. V.

    2014-02-01

    The review summarizes main trends in the development of methods for the prediction of protein post-translational modifications (PTMs) by considering the three most common types of PTMs — phosphorylation, acetylation and glycosylation. Considerable attention is given to general characteristics of regulatory interactions associated with PTMs. Different approaches to the prediction of PTMs are analyzed. Most of the methods are based only on the analysis of the neighbouring environment of modification sites. The related software is characterized by relatively low accuracy of PTM predictions, which may be due both to the incompleteness of training data and the features of PTM regulation. Advantages and limitations of the phylogenetic approach are considered. The prediction of PTMs using data on regulatory interactions, including the modular organization of interacting proteins, is a promising field, provided that a more carefully selected training data will be used. The bibliography includes 145 references.

  13. Development of a regional ensemble prediction method for probabilistic weather prediction

    International Nuclear Information System (INIS)

    Nohara, Daisuke; Tamura, Hidetoshi; Hirakuchi, Hiromaru

    2015-01-01

    A regional ensemble prediction method has been developed to provide probabilistic weather prediction using a numerical weather prediction model. To obtain consistent perturbations with the synoptic weather pattern, both of initial and lateral boundary perturbations were given by differences between control and ensemble member of the Japan Meteorological Agency (JMA)'s operational one-week ensemble forecast. The method provides a multiple ensemble member with a horizontal resolution of 15 km for 48-hour based on a downscaling of the JMA's operational global forecast accompanied with the perturbations. The ensemble prediction was examined in the case of heavy snow fall event in Kanto area on January 14, 2013. The results showed that the predictions represent different features of high-resolution spatiotemporal distribution of precipitation affected by intensity and location of extra-tropical cyclone in each ensemble member. Although the ensemble prediction has model bias of mean values and variances in some variables such as wind speed and solar radiation, the ensemble prediction has a potential to append a probabilistic information to a deterministic prediction. (author)

  14. Modelling for the Stripa site characterization and validation drift inflow: prediction of flow through fractured rock

    International Nuclear Information System (INIS)

    Herbert, A.; Gale, J.; MacLeod, R.; Lanyon, G.

    1991-12-01

    We present our approach to predicting flow through a fractured rock site; the site characterization and validation region in the Stripa mine. Our approach is based on discrete fracture network modelling using the NAPSAC computer code. We describe the conceptual models and assumptions that we have used to interpret the geometry and flow properties of the fracture networks, from measurements at the site. These are used to investigate large scale properties of the network and we show that for flows on scales larger than about 10 m, porous medium approximation should be used. The porous medium groundwater flow code CFEST is used to predict the large scale flows through the mine and the SCV region. This, in turn, is used to provide boundary conditions for more detailed models, which predict the details of flow, using a discrete fracture network model, on scales of less than 10 m. We conclude that a fracture network approach is feasible and that it provides a better understanding of details of flow than conventional porous medium approaches and a quantification of the uncertainty associated with predictive flow modelling characterised from field measurement in fractured rock. (au)

  15. Towards a unified fatigue life prediction method for marine structures

    CERN Document Server

    Cui, Weicheng; Wang, Fang

    2014-01-01

    In order to apply the damage tolerance design philosophy to design marine structures, accurate prediction of fatigue crack growth under service conditions is required. Now, more and more people have realized that only a fatigue life prediction method based on fatigue crack propagation (FCP) theory has the potential to explain various fatigue phenomena observed. In this book, the issues leading towards the development of a unified fatigue life prediction (UFLP) method based on FCP theory are addressed. Based on the philosophy of the UFLP method, the current inconsistency between fatigue design and inspection of marine structures could be resolved. This book presents the state-of-the-art and recent advances, including those by the authors, in fatigue studies. It is designed to lead the future directions and to provide a useful tool in many practical applications. It is intended to address to engineers, naval architects, research staff, professionals and graduates engaged in fatigue prevention design and survey ...

  16. DASPfind: new efficient method to predict drug–target interactions

    KAUST Repository

    Ba Alawi, Wail; Soufan, Othman; Essack, Magbubah; Kalnis, Panos; Bajic, Vladimir B.

    2016-01-01

    DASPfind is a computational method for finding reliable new interactions between drugs and proteins. We show over six different DTI datasets that DASPfind outperforms other state-of-the-art methods when the single top-ranked predictions are considered, or when a drug with no known targets or with few known targets is considered. We illustrate the usefulness and practicality of DASPfind by predicting novel DTIs for the Ion Channel dataset. The validated predictions suggest that DASPfind can be used as an efficient method to identify correct DTIs, thus reducing the cost of necessary experimental verifications in the process of drug discovery. DASPfind can be accessed online at: http://​www.​cbrc.​kaust.​edu.​sa/​daspfind.

  17. Prediction of Protein–Protein Interactions by Evidence Combining Methods

    Directory of Open Access Journals (Sweden)

    Ji-Wei Chang

    2016-11-01

    Full Text Available Most cellular functions involve proteins’ features based on their physical interactions with other partner proteins. Sketching a map of protein–protein interactions (PPIs is therefore an important inception step towards understanding the basics of cell functions. Several experimental techniques operating in vivo or in vitro have made significant contributions to screening a large number of protein interaction partners, especially high-throughput experimental methods. However, computational approaches for PPI predication supported by rapid accumulation of data generated from experimental techniques, 3D structure definitions, and genome sequencing have boosted the map sketching of PPIs. In this review, we shed light on in silico PPI prediction methods that integrate evidence from multiple sources, including evolutionary relationship, function annotation, sequence/structure features, network topology and text mining. These methods are developed for integration of multi-dimensional evidence, for designing the strategies to predict novel interactions, and for making the results consistent with the increase of prediction coverage and accuracy.

  18. Predicting Metabolic Syndrome Using the Random Forest Method

    Directory of Open Access Journals (Sweden)

    Apilak Worachartcheewan

    2015-01-01

    Full Text Available Aims. This study proposes a computational method for determining the prevalence of metabolic syndrome (MS and to predict its occurrence using the National Cholesterol Education Program Adult Treatment Panel III (NCEP ATP III criteria. The Random Forest (RF method is also applied to identify significant health parameters. Materials and Methods. We used data from 5,646 adults aged between 18–78 years residing in Bangkok who had received an annual health check-up in 2008. MS was identified using the NCEP ATP III criteria. The RF method was applied to predict the occurrence of MS and to identify important health parameters surrounding this disorder. Results. The overall prevalence of MS was 23.70% (34.32% for males and 17.74% for females. RF accuracy for predicting MS in an adult Thai population was 98.11%. Further, based on RF, triglyceride levels were the most important health parameter associated with MS. Conclusion. RF was shown to predict MS in an adult Thai population with an accuracy >98% and triglyceride levels were identified as the most informative variable associated with MS. Therefore, using RF to predict MS may be potentially beneficial in identifying MS status for preventing the development of diabetes mellitus and cardiovascular diseases.

  19. Prediction of polymer flooding performance using an analytical method

    International Nuclear Information System (INIS)

    Tan Czek Hoong; Mariyamni Awang; Foo Kok Wai

    2001-01-01

    The study investigated the applicability of an analytical method developed by El-Khatib in polymer flooding. Results from a simulator UTCHEM and experiments were compared with the El-Khatib prediction method. In general, by assuming a constant viscosity polymer injection, the method gave much higher recovery values than the simulation runs and the experiments. A modification of the method gave better correlation, albeit only oil production. Investigation is continuing on modifying the method so that a better overall fit can be obtained for polymer flooding. (Author)

  20. Preface to the Focus Issue: Chaos Detection Methods and Predictability

    International Nuclear Information System (INIS)

    Gottwald, Georg A.; Skokos, Charalampos

    2014-01-01

    This Focus Issue presents a collection of papers originating from the workshop Methods of Chaos Detection and Predictability: Theory and Applications held at the Max Planck Institute for the Physics of Complex Systems in Dresden, June 17–21, 2013. The main aim of this interdisciplinary workshop was to review comprehensively the theory and numerical implementation of the existing methods of chaos detection and predictability, as well as to report recent applications of these techniques to different scientific fields. The collection of twelve papers in this Focus Issue represents the wide range of applications, spanning mathematics, physics, astronomy, particle accelerator physics, meteorology and medical research. This Preface surveys the papers of this Issue

  1. Preface to the Focus Issue: chaos detection methods and predictability.

    Science.gov (United States)

    Gottwald, Georg A; Skokos, Charalampos

    2014-06-01

    This Focus Issue presents a collection of papers originating from the workshop Methods of Chaos Detection and Predictability: Theory and Applications held at the Max Planck Institute for the Physics of Complex Systems in Dresden, June 17-21, 2013. The main aim of this interdisciplinary workshop was to review comprehensively the theory and numerical implementation of the existing methods of chaos detection and predictability, as well as to report recent applications of these techniques to different scientific fields. The collection of twelve papers in this Focus Issue represents the wide range of applications, spanning mathematics, physics, astronomy, particle accelerator physics, meteorology and medical research. This Preface surveys the papers of this Issue.

  2. [Prediction of 137Cs accumulation in animal products in the territory of Semipalatinsk test site].

    Science.gov (United States)

    Spiridonov, S I; Gontarenko, I A; Mukusheva, M K; Fesenko, S V; Semioshkina, N A

    2005-01-01

    The paper describes mathematical models for 137Cs behavior in the organism of horses and sheep pasturing on the bording area to the testing area "Ground Zero" of the Semipalatinsk Test Site. The models are parameterized on the base of the data from an experiment with the breeds of animals now commonly encountered within the Semipalatinsk Test Site. The predictive calculations with the models devised have shown that 137Cs concentrations in milk of horses and sheep pasturingon the testing area to "Ground Zero" can exceed the adopted standards during a long period of time.

  3. The energetic cost of walking: a comparison of predictive methods.

    Directory of Open Access Journals (Sweden)

    Patricia Ann Kramer

    Full Text Available BACKGROUND: The energy that animals devote to locomotion has been of intense interest to biologists for decades and two basic methodologies have emerged to predict locomotor energy expenditure: those based on metabolic and those based on mechanical energy. Metabolic energy approaches share the perspective that prediction of locomotor energy expenditure should be based on statistically significant proxies of metabolic function, while mechanical energy approaches, which derive from many different perspectives, focus on quantifying the energy of movement. Some controversy exists as to which mechanical perspective is "best", but from first principles all mechanical methods should be equivalent if the inputs to the simulation are of similar quality. Our goals in this paper are 1 to establish the degree to which the various methods of calculating mechanical energy are correlated, and 2 to investigate to what degree the prediction methods explain the variation in energy expenditure. METHODOLOGY/PRINCIPAL FINDINGS: We use modern humans as the model organism in this experiment because their data are readily attainable, but the methodology is appropriate for use in other species. Volumetric oxygen consumption and kinematic and kinetic data were collected on 8 adults while walking at their self-selected slow, normal and fast velocities. Using hierarchical statistical modeling via ordinary least squares and maximum likelihood techniques, the predictive ability of several metabolic and mechanical approaches were assessed. We found that all approaches are correlated and that the mechanical approaches explain similar amounts of the variation in metabolic energy expenditure. Most methods predict the variation within an individual well, but are poor at accounting for variation between individuals. CONCLUSION: Our results indicate that the choice of predictive method is dependent on the question(s of interest and the data available for use as inputs. Although we

  4. The energetic cost of walking: a comparison of predictive methods.

    Science.gov (United States)

    Kramer, Patricia Ann; Sylvester, Adam D

    2011-01-01

    The energy that animals devote to locomotion has been of intense interest to biologists for decades and two basic methodologies have emerged to predict locomotor energy expenditure: those based on metabolic and those based on mechanical energy. Metabolic energy approaches share the perspective that prediction of locomotor energy expenditure should be based on statistically significant proxies of metabolic function, while mechanical energy approaches, which derive from many different perspectives, focus on quantifying the energy of movement. Some controversy exists as to which mechanical perspective is "best", but from first principles all mechanical methods should be equivalent if the inputs to the simulation are of similar quality. Our goals in this paper are 1) to establish the degree to which the various methods of calculating mechanical energy are correlated, and 2) to investigate to what degree the prediction methods explain the variation in energy expenditure. We use modern humans as the model organism in this experiment because their data are readily attainable, but the methodology is appropriate for use in other species. Volumetric oxygen consumption and kinematic and kinetic data were collected on 8 adults while walking at their self-selected slow, normal and fast velocities. Using hierarchical statistical modeling via ordinary least squares and maximum likelihood techniques, the predictive ability of several metabolic and mechanical approaches were assessed. We found that all approaches are correlated and that the mechanical approaches explain similar amounts of the variation in metabolic energy expenditure. Most methods predict the variation within an individual well, but are poor at accounting for variation between individuals. Our results indicate that the choice of predictive method is dependent on the question(s) of interest and the data available for use as inputs. Although we used modern humans as our model organism, these results can be extended

  5. Probabilistic Seismic Hazard Assessment Method for Nonlinear Soil Sites based on the Hazard Spectrum of Bedrock Sites

    International Nuclear Information System (INIS)

    Hahm, Dae Gi; Seo, Jeong Moon; Choi, In Kil

    2011-01-01

    For the probabilistic safety assessment of the nuclear power plants (NPP) under seismic events, the rational probabilistic seismic hazard estimation should be performed. Generally, the probabilistic seismic hazard of NPP site is represented by the uniform hazard spectrum (UHS) for the specific annual frequency. In most case, since that the attenuation equations were defined for the bedrock sites, the standard attenuation laws cannot be applied to the general soft soil sites. Hence, for the probabilistic estimation of the seismic hazard of soft soil sites, a methodology of probabilistic seismic hazard analysis (PSHA) coupled with nonlinear dynamic analyses of the soil column are required. Two methods are commonly used for the site response analysis considering the nonlinearity of sites. The one is the deterministic method and another is the probabilistic method. In the analysis of site response, there exist many uncertainty factors such as the variation of the magnitude and frequency contents of input ground motion, and material properties of soil deposits. Hence, nowadays, it is recommended that the adoption of the probabilistic method for the PSHA of soft soil deposits considering such uncertainty factors. In this study, we estimated the amplification factor of the surface of the soft soil deposits with considering the uncertainties of the input ground motions and the soil material properties. Then, we proposed the probabilistic methodology to evaluate the UHS of the soft soil site by multiplying the amplification factor to that of the bedrock site. The proposed method was applied to four typical target sites of KNGR and APR1400 NPP site categories

  6. Methods for the analysis and remediation of contaminated sites

    International Nuclear Information System (INIS)

    Mariani, M.; Bemporad, E.; Berardi, S.; Marino, A.; Paglietti, F.

    2008-01-01

    In Italy, in recent years, the number of contaminated sites has multiplied disproportionately. In essence, contamination is caused by accidental spills or intentional discharge of pollutants into the soils or waters from industrial activities, or non-controlled deposits of urban and/or industrial waste, mostly part toxic and harmful. Contaminated sites clearly pose risks to human health and the environment; hence the need to remediate these sites. The remediation of soil and water and the restoration of degraded areas are complex operations requiring specific technical and scientific know-how, including knowledge of the methodologies and tools required to tackle problems arising during the different phases of the remediation process. These include, in particular: - health and environmental risk assessment procedures for the quantification of risks to human health (general population and workers) and the environment from a contaminated site; - remote sensing and the Geographical Information Systems (GIS), which are a fundamentally important IT support for each phase of planning and management of remediation interventions; - criteria for the management of sites contaminated by asbestos, a highly carcinogenic and therefore hazardous substance that was widely used in the past due to its particular mechanical and thermal characteristics; - analysis of the issues relating to waste management in contaminated sites; - relationship between safety procedures for workers and the general population. Identification of the best available techniques for an efficient, integrated management of contaminated sites, which will also take into account the health protection of workers and of the general population living near such sites

  7. Combining gene prediction methods to improve metagenomic gene annotation

    Directory of Open Access Journals (Sweden)

    Rosen Gail L

    2011-01-01

    Full Text Available Abstract Background Traditional gene annotation methods rely on characteristics that may not be available in short reads generated from next generation technology, resulting in suboptimal performance for metagenomic (environmental samples. Therefore, in recent years, new programs have been developed that optimize performance on short reads. In this work, we benchmark three metagenomic gene prediction programs and combine their predictions to improve metagenomic read gene annotation. Results We not only analyze the programs' performance at different read-lengths like similar studies, but also separate different types of reads, including intra- and intergenic regions, for analysis. The main deficiencies are in the algorithms' ability to predict non-coding regions and gene edges, resulting in more false-positives and false-negatives than desired. In fact, the specificities of the algorithms are notably worse than the sensitivities. By combining the programs' predictions, we show significant improvement in specificity at minimal cost to sensitivity, resulting in 4% improvement in accuracy for 100 bp reads with ~1% improvement in accuracy for 200 bp reads and above. To correctly annotate the start and stop of the genes, we find that a consensus of all the predictors performs best for shorter read lengths while a unanimous agreement is better for longer read lengths, boosting annotation accuracy by 1-8%. We also demonstrate use of the classifier combinations on a real dataset. Conclusions To optimize the performance for both prediction and annotation accuracies, we conclude that the consensus of all methods (or a majority vote is the best for reads 400 bp and shorter, while using the intersection of GeneMark and Orphelia predictions is the best for reads 500 bp and longer. We demonstrate that most methods predict over 80% coding (including partially coding reads on a real human gut sample sequenced by Illumina technology.

  8. Trends of Abutment-Scour Prediction Equations Applied to 144 Field Sites in South Carolina

    Science.gov (United States)

    Benedict, Stephen T.; Deshpande, Nikhil; Aziz, Nadim M.; Conrads, Paul

    2006-01-01

    The U.S. Geological Survey conducted a study in cooperation with the Federal Highway Administration in which predicted abutment-scour depths computed with selected predictive equations were compared with field measurements of abutment-scour depth made at 144 bridges in South Carolina. The assessment used five equations published in the Fourth Edition of 'Evaluating Scour at Bridges,' (Hydraulic Engineering Circular 18), including the original Froehlich, the modified Froehlich, the Sturm, the Maryland, and the HIRE equations. An additional unpublished equation also was assessed. Comparisons between predicted and observed scour depths are intended to illustrate general trends and order-of-magnitude differences for the prediction equations. Field measurements were taken during non-flood conditions when the hydraulic conditions that caused the scour generally are unknown. The predicted scour depths are based on hydraulic conditions associated with the 100-year flow at all sites and the flood of record for 35 sites. Comparisons showed that predicted scour depths frequently overpredict observed scour and at times were excessive. The comparison also showed that underprediction occurred, but with less frequency. The performance of these equations indicates that they are poor predictors of abutment-scour depth in South Carolina, and it is probable that poor performance will occur when the equations are applied in other geographic regions. Extensive data and graphs used to compare predicted and observed scour depths in this study were compiled into spreadsheets and are included in digital format with this report. In addition to the equation-comparison data, Water-Surface Profile Model tube-velocity data, soil-boring data, and selected abutment-scour data are included in digital format with this report. The digital database was developed as a resource for future researchers and is especially valuable for evaluating the reasonableness of future equations that may be developed.

  9. Orthology prediction methods: a quality assessment using curated protein families.

    Science.gov (United States)

    Trachana, Kalliopi; Larsson, Tomas A; Powell, Sean; Chen, Wei-Hua; Doerks, Tobias; Muller, Jean; Bork, Peer

    2011-10-01

    The increasing number of sequenced genomes has prompted the development of several automated orthology prediction methods. Tests to evaluate the accuracy of predictions and to explore biases caused by biological and technical factors are therefore required. We used 70 manually curated families to analyze the performance of five public methods in Metazoa. We analyzed the strengths and weaknesses of the methods and quantified the impact of biological and technical challenges. From the latter part of the analysis, genome annotation emerged as the largest single influencer, affecting up to 30% of the performance. Generally, most methods did well in assigning orthologous group but they failed to assign the exact number of genes for half of the groups. The publicly available benchmark set (http://eggnog.embl.de/orthobench/) should facilitate the improvement of current orthology assignment protocols, which is of utmost importance for many fields of biology and should be tackled by a broad scientific community. Copyright © 2011 WILEY Periodicals, Inc.

  10. Fast Prediction Method for Steady-State Heat Convection

    KAUST Repository

    Wáng, Yì

    2012-03-14

    A reduced model by proper orthogonal decomposition (POD) and Galerkin projection methods for steady-state heat convection is established on a nonuniform grid. It was verified by thousands of examples that the results are in good agreement with the results obtained from the finite volume method. This model can also predict the cases where model parameters far exceed the sample scope. Moreover, the calculation time needed by the model is much shorter than that needed for the finite volume method. Thus, the nonuniform POD-Galerkin projection method exhibits high accuracy, good suitability, and fast computation. It has universal significance for accurate and fast prediction. Also, the methodology can be applied to more complex modeling in chemical engineering and technology, such as reaction and turbulence. © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  11. Predicting human splicing branchpoints by combining sequence-derived features and multi-label learning methods.

    Science.gov (United States)

    Zhang, Wen; Zhu, Xiaopeng; Fu, Yu; Tsuji, Junko; Weng, Zhiping

    2017-12-01

    Alternative splicing is the critical process in a single gene coding, which removes introns and joins exons, and splicing branchpoints are indicators for the alternative splicing. Wet experiments have identified a great number of human splicing branchpoints, but many branchpoints are still unknown. In order to guide wet experiments, we develop computational methods to predict human splicing branchpoints. Considering the fact that an intron may have multiple branchpoints, we transform the branchpoint prediction as the multi-label learning problem, and attempt to predict branchpoint sites from intron sequences. First, we investigate a variety of intron sequence-derived features, such as sparse profile, dinucleotide profile, position weight matrix profile, Markov motif profile and polypyrimidine tract profile. Second, we consider several multi-label learning methods: partial least squares regression, canonical correlation analysis and regularized canonical correlation analysis, and use them as the basic classification engines. Third, we propose two ensemble learning schemes which integrate different features and different classifiers to build ensemble learning systems for the branchpoint prediction. One is the genetic algorithm-based weighted average ensemble method; the other is the logistic regression-based ensemble method. In the computational experiments, two ensemble learning methods outperform benchmark branchpoint prediction methods, and can produce high-accuracy results on the benchmark dataset.

  12. Hybrid robust predictive optimization method of power system dispatch

    Science.gov (United States)

    Chandra, Ramu Sharat [Niskayuna, NY; Liu, Yan [Ballston Lake, NY; Bose, Sumit [Niskayuna, NY; de Bedout, Juan Manuel [West Glenville, NY

    2011-08-02

    A method of power system dispatch control solves power system dispatch problems by integrating a larger variety of generation, load and storage assets, including without limitation, combined heat and power (CHP) units, renewable generation with forecasting, controllable loads, electric, thermal and water energy storage. The method employs a predictive algorithm to dynamically schedule different assets in order to achieve global optimization and maintain the system normal operation.

  13. Available Prediction Methods for Corrosion under Insulation (CUI: A Review

    Directory of Open Access Journals (Sweden)

    Burhani Nurul Rawaida Ain

    2014-07-01

    Full Text Available Corrosion under insulation (CUI is an increasingly important issue for the piping in industries especially petrochemical and chemical plants due to its unexpected catastrophic disaster. Therefore, attention towards the maintenance and prediction of CUI occurrence, particularly in the corrosion rates, has grown in recent years. In this study, a literature review in determining the corrosion rates by using various prediction models and method of the corrosion occurrence between the external surface piping and its insulation was carried out. The results, prediction models and methods available were presented for future research references. However, most of the prediction methods available are based on each local industrial data only which might be different based on the plant location, environment, temperature and many other factors which may contribute to the difference and reliability of the model developed. Thus, it is more reliable if those models or method supported by laboratory testing or simulation which includes the factors promoting CUI such as environment temperature, insulation types, operating temperatures, and other factors.

  14. SVM-based prediction of propeptide cleavage sites in spider toxins identifies toxin innovation in an Australian tarantula.

    Directory of Open Access Journals (Sweden)

    Emily S W Wong

    Full Text Available Spider neurotoxins are commonly used as pharmacological tools and are a popular source of novel compounds with therapeutic and agrochemical potential. Since venom peptides are inherently toxic, the host spider must employ strategies to avoid adverse effects prior to venom use. It is partly for this reason that most spider toxins encode a protective proregion that upon enzymatic cleavage is excised from the mature peptide. In order to identify the mature toxin sequence directly from toxin transcripts, without resorting to protein sequencing, the propeptide cleavage site in the toxin precursor must be predicted bioinformatically. We evaluated different machine learning strategies (support vector machines, hidden Markov model and decision tree and developed an algorithm (SpiderP for prediction of propeptide cleavage sites in spider toxins. Our strategy uses a support vector machine (SVM framework that combines both local and global sequence information. Our method is superior or comparable to current tools for prediction of propeptide sequences in spider toxins. Evaluation of the SVM method on an independent test set of known toxin sequences yielded 96% sensitivity and 100% specificity. Furthermore, we sequenced five novel peptides (not used to train the final predictor from the venom of the Australian tarantula Selenotypus plumipes to test the accuracy of the predictor and found 80% sensitivity and 99.6% 8-mer specificity. Finally, we used the predictor together with homology information to predict and characterize seven groups of novel toxins from the deeply sequenced venom gland transcriptome of S. plumipes, which revealed structural complexity and innovations in the evolution of the toxins. The precursor prediction tool (SpiderP is freely available on ArachnoServer (http://www.arachnoserver.org/spiderP.html, a web portal to a comprehensive relational database of spider toxins. All training data, test data, and scripts used are available from

  15. On splice site prediction using weight array models: a comparison of smoothing techniques

    International Nuclear Information System (INIS)

    Taher, Leila; Meinicke, Peter; Morgenstern, Burkhard

    2007-01-01

    In most eukaryotic genes, protein-coding exons are separated by non-coding introns which are removed from the primary transcript by a process called 'splicing'. The positions where introns are cut and exons are spliced together are called 'splice sites'. Thus, computational prediction of splice sites is crucial for gene finding in eukaryotes. Weight array models are a powerful probabilistic approach to splice site detection. Parameters for these models are usually derived from m-tuple frequencies in trusted training data and subsequently smoothed to avoid zero probabilities. In this study we compare three different ways of parameter estimation for m-tuple frequencies, namely (a) non-smoothed probability estimation, (b) standard pseudo counts and (c) a Gaussian smoothing procedure that we recently developed

  16. Computational prediction of cAMP receptor protein (CRP binding sites in cyanobacterial genomes

    Directory of Open Access Journals (Sweden)

    Su Zhengchang

    2009-01-01

    Full Text Available Abstract Background Cyclic AMP receptor protein (CRP, also known as catabolite gene activator protein (CAP, is an important transcriptional regulator widely distributed in many bacteria. The biological processes under the regulation of CRP are highly diverse among different groups of bacterial species. Elucidation of CRP regulons in cyanobacteria will further our understanding of the physiology and ecology of this important group of microorganisms. Previously, CRP has been experimentally studied in only two cyanobacterial strains: Synechocystis sp. PCC 6803 and Anabaena sp. PCC 7120; therefore, a systematic genome-scale study of the potential CRP target genes and binding sites in cyanobacterial genomes is urgently needed. Results We have predicted and analyzed the CRP binding sites and regulons in 12 sequenced cyanobacterial genomes using a highly effective cis-regulatory binding site scanning algorithm. Our results show that cyanobacterial CRP binding sites are very similar to those in E. coli; however, the regulons are very different from that of E. coli. Furthermore, CRP regulons in different cyanobacterial species/ecotypes are also highly diversified, ranging from photosynthesis, carbon fixation and nitrogen assimilation, to chemotaxis and signal transduction. In addition, our prediction indicates that crp genes in modern cyanobacteria are likely inherited from a common ancestral gene in their last common ancestor, and have adapted various cellular functions in different environments, while some cyanobacteria lost their crp genes as well as CRP binding sites during the course of evolution. Conclusion The CRP regulons in cyanobacteria are highly diversified, probably as a result of divergent evolution to adapt to various ecological niches. Cyanobacterial CRPs may function as lineage-specific regulators participating in various cellular processes, and are important in some lineages. However, they are dispensable in some other lineages. The

  17. G-LoSA for Prediction of Protein-Ligand Binding Sites and Structures.

    Science.gov (United States)

    Lee, Hui Sun; Im, Wonpil

    2017-01-01

    Recent advances in high-throughput structure determination and computational protein structure prediction have significantly enriched the universe of protein structure. However, there is still a large gap between the number of available protein structures and that of proteins with annotated function in high accuracy. Computational structure-based protein function prediction has emerged to reduce this knowledge gap. The identification of a ligand binding site and its structure is critical to the determination of a protein's molecular function. We present a computational methodology for predicting small molecule ligand binding site and ligand structure using G-LoSA, our protein local structure alignment and similarity measurement tool. All the computational procedures described here can be easily implemented using G-LoSA Toolkit, a package of standalone software programs and preprocessed PDB structure libraries. G-LoSA and G-LoSA Toolkit are freely available to academic users at http://compbio.lehigh.edu/GLoSA . We also illustrate a case study to show the potential of our template-based approach harnessing G-LoSA for protein function prediction.

  18. Identification of prokaryotic and eukaryotic signal peptides and prediction of their cleavage sites

    DEFF Research Database (Denmark)

    Nielsen, Henrik; Engelbrecht, Jacob; Brunak, Søren

    1997-01-01

    We have developed a new method for the identification of signal peptides and their cleavage based on neural networks trained on separate sets of prokaryotic and eukaryotic sequence. The method performs significantly better than previous prediction schemes and can easily be applied on genome...

  19. Social Networking Site Use Predicts Changes in Young Adults’ Psychological Adjustment

    Science.gov (United States)

    Szwedo, David E.; Mikami, Amori Yee; Allen, Joseph P.

    2012-01-01

    This study examined youths’ friendships and posted pictures on social networking sites as predictors of changes in their adjustment over time. Observational, self-report, and peer report data were obtained from a community sample of 89 young adults interviewed at age 21 and again at age 22. Findings were consistent with a leveling effect for online friendships, predicting decreases in internalizing symptoms for youth with lower initial levels of social acceptance, but increases in symptoms for youth with higher initial levels over the following year. Across the entire sample, deviant behavior in posted photos predicted increases in young adults’ problematic alcohol use over time. The importance of considering the interplay between online and offline social factors for predicting adjustment is discussed. PMID:23109797

  20. A design method for an intuitive web site

    Energy Technology Data Exchange (ETDEWEB)

    Quinniey, M.L.; Diegert, K.V.; Baca, B.G.; Forsythe, J.C.; Grose, E.

    1999-11-03

    The paper describes a methodology for designing a web site for human factor engineers that is applicable for designing a web site for a group of people. Many web pages on the World Wide Web are not organized in a format that allows a user to efficiently find information. Often the information and hypertext links on web pages are not organized into intuitive groups. Intuition implies that a person is able to use their knowledge of a paradigm to solve a problem. Intuitive groups are categories that allow web page users to find information by using their intuition or mental models of categories. In order to improve the human factors engineers efficiency for finding information on the World Wide Web, research was performed to develop a web site that serves as a tool for finding information effectively. The paper describes a methodology for designing a web site for a group of people who perform similar task in an organization.

  1. SVM prediction of ligand-binding sites in bacterial lipoproteins employing shape and physio-chemical descriptors.

    Science.gov (United States)

    Kadam, Kiran; Prabhakar, Prashant; Jayaraman, V K

    2012-11-01

    Bacterial lipoproteins play critical roles in various physiological processes including the maintenance of pathogenicity and numbers of them are being considered as potential candidates for generating novel vaccines. In this work, we put forth an algorithm to identify and predict ligand-binding sites in bacterial lipoproteins. The method uses three types of pocket descriptors, namely fpocket descriptors, 3D Zernike descriptors and shell descriptors, and combines them with Support Vector Machine (SVM) method for the classification. The three types of descriptors represent shape-based properties of the pocket as well as its local physio-chemical features. All three types of descriptors, along with their hybrid combinations are evaluated with SVM and to improve classification performance, WEKA-InfoGain feature selection is applied. Results obtained in the study show that the classifier successfully differentiates between ligand-binding and non-binding pockets. For the combination of three types of descriptors, 10 fold cross-validation accuracy of 86.83% is obtained for training while the selected model achieved test Matthews Correlation Coefficient (MCC) of 0.534. Individually or in combination with new and existing methods, our model can be a very useful tool for the prediction of potential ligand-binding sites in bacterial lipoproteins.

  2. Customer churn prediction using a hybrid method and censored data

    Directory of Open Access Journals (Sweden)

    Reza Tavakkoli-Moghaddam

    2013-05-01

    Full Text Available Customers are believed to be the main part of any organization’s assets and customer retention as well as customer churn management are important responsibilities of organizations. In today’s competitive environment, organization must do their best to retain their existing customers since attracting new customers cost significantly more than taking care of existing ones. In this paper, we present a hybrid method based on neural network and Cox regression analysis where neural network is used for outlier data and Cox regression method is implemented for prediction of future events. The proposed model of this paper has been implemented on some data and the results are compared based on five criteria including prediction accuracy, errors’ type I and II, root mean square error and mean absolute deviation. The preliminary results indicate that the proposed model of this paper performs better than alternative methods.

  3. Method of predicting surface deformation in the form of sinkholes

    Energy Technology Data Exchange (ETDEWEB)

    Chudek, M.; Arkuszewski, J.

    1980-06-01

    Proposes a method for predicting probability of sinkhole shaped subsidence, number of funnel-shaped subsidences and size of individual funnels. The following factors which influence the sudden subsidence of the surface in the form of funnels are analyzed: geologic structure of the strata between mining workings and the surface, mining depth, time factor, and geologic disolocations. Sudden surface subsidence is observed only in the case of workings situated up to a few dozen meters from the surface. Using the proposed method is explained with some examples. It is suggested that the method produces correct results which can be used in coal mining and in ore mining. (1 ref.) (In Polish)

  4. JASSA: a comprehensive tool for prediction of SUMOylation sites and SIMs.

    Science.gov (United States)

    Beauclair, Guillaume; Bridier-Nahmias, Antoine; Zagury, Jean-François; Saïb, Ali; Zamborlini, Alessia

    2015-11-01

    Post-translational modification by the Small Ubiquitin-like Modifier (SUMO) proteins, a process termed SUMOylation, is involved in many fundamental cellular processes. SUMO proteins are conjugated to a protein substrate, creating an interface for the recruitment of cofactors harboring SUMO-interacting motifs (SIMs). Mapping both SUMO-conjugation sites and SIMs is required to study the functional consequence of SUMOylation. To define the best candidate sites for experimental validation we designed JASSA, a Joint Analyzer of SUMOylation site and SIMs. JASSA is a predictor that uses a scoring system based on a Position Frequency Matrix derived from the alignment of experimental SUMOylation sites or SIMs. Compared with existing web-tools, JASSA displays on par or better performances. Novel features were implemented towards a better evaluation of the prediction, including identification of database hits matching the query sequence and representation of candidate sites within the secondary structural elements and/or the 3D fold of the protein of interest, retrievable from deposited PDB files. JASSA is freely accessible at http://www.jassa.fr/. Website is implemented in PHP and MySQL, with all major browsers supported. guillaume.beauclair@inserm.fr Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  5. Preparatory research to develop an operational method to calibrate airborne sensor data using a network of ground calibration sites

    International Nuclear Information System (INIS)

    Milton, E.J.; Smith, G.M.; Lawless, K.P.

    1996-01-01

    The objective of the research is to develop an operational method to convert airborne spectral radiance data to reflectance using a number of well-characterized ground calibration sites located around the UK. The study is in three phases. First, a pilot study has been conducted at a disused airfield in southern England to test the feasibility of the open-quote empirical line close-quote method of sensor calibration. The second phase is developing methods to predict temporal changes in the bidirectional reflectance of ground calibration sites. The final phase of the project will look at methods to extend such calibrations spatially. This paper presents some results from the first phase of this study. The viability of the empirical line method of correction is shown to depend upon the use of ground targets whose in-band reflectance encompasses that of the targets of interest in the spectral band(s) concerned. The experimental design for the second phase of the study, in which methods to predict temporal trends in the bidirectional reflectance of these sites will be developed, is discussed. Finally, it is planned to develop an automated method of searching through Landsat TM data for the UK to identify a number of candidate ground calibration sites for which the model can be tested. 11 refs., 5 figs., 5 tabs

  6. Methods and systems for identifying ligand-protein binding sites

    KAUST Repository

    Gao, Xin; Naveed, Hammad

    2016-01-01

    The invention provides a novel integrated structure and system-based approach for drug target prediction that enables the large-scale discovery of new targets for existing drugs Novel computer-readable storage media and computer systems are also

  7. Determination of the strain generated in InAs/InP quantum wires: prediction of nucleation sites

    International Nuclear Information System (INIS)

    Molina, S I; Ben, T; Sales, D L; Pizarro, J; Galindo, P L; Varela, M; Pennycook, S J; Fuster, D; Gonzalez, Y; Gonzalez, L

    2006-01-01

    The compositional distribution in a self-assembled InAs(P) quantum wire grown by molecular beam epitaxy on an InP(001) substrate has been determined by electron energy loss spectrum imaging. We have determined the strain and stress fields generated in and around this wire capped with a 5 nm InP layer by finite element calculations using as input the compositional map experimentally obtained. Preferential sites for nucleation of wires grown on the surface of this InP capping layer are predicted, based on chemical potential minimization, from the determined strain and stress fields on this surface. The determined preferential sites for wire nucleation agree with their experimentally measured locations. The method used in this paper, which combines electron energy loss spectroscopy, high-resolution Z contrast imaging, and elastic theory finite element calculations, is believed to be a valuable technique of wide applicability for predicting the preferential nucleation sites of epitaxial self-assembled nano-objects

  8. Determination of the strain generated in InAs/InP quantum wires: prediction of nucleation sites

    Energy Technology Data Exchange (ETDEWEB)

    Molina, S I [Departamento de Ciencia de los Materiales e I.M. y Q.I., Facultad de Ciencias, Universidad de Cadiz, Campus RIo San Pedro, s/n, 11510 Puerto Real, Cadiz (Spain); Ben, T [Departamento de Ciencia de los Materiales e I.M. y Q.I., Facultad de Ciencias, Universidad de Cadiz, Campus RIo San Pedro, s/n, 11510 Puerto Real, Cadiz (Spain); Sales, D L [Departamento de Ciencia de los Materiales e I.M. y Q.I., Facultad de Ciencias, Universidad de Cadiz, Campus RIo San Pedro, s/n, 11510 Puerto Real, Cadiz (Spain); Pizarro, J [Departamento de Lenguajes y Sistemas Informaticos, CASEM, Universidad de Cadiz, Campus RIo San Pedro, s/n, 11510 Puerto Real, Cadiz (Spain); Galindo, P L [Departamento de Lenguajes y Sistemas Informaticos, CASEM, Universidad de Cadiz, Campus RIo San Pedro, s/n, 11510 Puerto Real, Cadiz (Spain); Varela, M [Condensed Matter Sciences Division, Oak Ridge National Laboratory, Oak Ridge, TN 37831 (United States); Pennycook, S J [Condensed Matter Sciences Division, Oak Ridge National Laboratory, Oak Ridge, TN 37831 (United States); Fuster, D [Instituto de Microelectronica de Madrid (CNM, CSIC), Isaac Newton 8, 28760 Tres Cantos, Madrid (Spain); Gonzalez, Y [Instituto de Microelectronica de Madrid (CNM, CSIC), Isaac Newton 8, 28760 Tres Cantos, Madrid (Spain); Gonzalez, L [Instituto de Microelectronica de Madrid (CNM, CSIC), Isaac Newton 8, 28760 Tres Cantos, Madrid (Spain)

    2006-11-28

    The compositional distribution in a self-assembled InAs(P) quantum wire grown by molecular beam epitaxy on an InP(001) substrate has been determined by electron energy loss spectrum imaging. We have determined the strain and stress fields generated in and around this wire capped with a 5 nm InP layer by finite element calculations using as input the compositional map experimentally obtained. Preferential sites for nucleation of wires grown on the surface of this InP capping layer are predicted, based on chemical potential minimization, from the determined strain and stress fields on this surface. The determined preferential sites for wire nucleation agree with their experimentally measured locations. The method used in this paper, which combines electron energy loss spectroscopy, high-resolution Z contrast imaging, and elastic theory finite element calculations, is believed to be a valuable technique of wide applicability for predicting the preferential nucleation sites of epitaxial self-assembled nano-objects.

  9. Sequence-based prediction of protein-binding sites in DNA: comparative study of two SVM models.

    Science.gov (United States)

    Park, Byungkyu; Im, Jinyong; Tuvshinjargal, Narankhuu; Lee, Wook; Han, Kyungsook

    2014-11-01

    As many structures of protein-DNA complexes have been known in the past years, several computational methods have been developed to predict DNA-binding sites in proteins. However, its inverse problem (i.e., predicting protein-binding sites in DNA) has received much less attention. One of the reasons is that the differences between the interaction propensities of nucleotides are much smaller than those between amino acids. Another reason is that DNA exhibits less diverse sequence patterns than protein. Therefore, predicting protein-binding DNA nucleotides is much harder than predicting DNA-binding amino acids. We computed the interaction propensity (IP) of nucleotide triplets with amino acids using an extensive dataset of protein-DNA complexes, and developed two support vector machine (SVM) models that predict protein-binding nucleotides from sequence data alone. One SVM model predicts protein-binding nucleotides using DNA sequence data alone, and the other SVM model predicts protein-binding nucleotides using both DNA and protein sequences. In a 10-fold cross-validation with 1519 DNA sequences, the SVM model that uses DNA sequence data only predicted protein-binding nucleotides with an accuracy of 67.0%, an F-measure of 67.1%, and a Matthews correlation coefficient (MCC) of 0.340. With an independent dataset of 181 DNAs that were not used in training, it achieved an accuracy of 66.2%, an F-measure 66.3% and a MCC of 0.324. Another SVM model that uses both DNA and protein sequences achieved an accuracy of 69.6%, an F-measure of 69.6%, and a MCC of 0.383 in a 10-fold cross-validation with 1519 DNA sequences and 859 protein sequences. With an independent dataset of 181 DNAs and 143 proteins, it showed an accuracy of 67.3%, an F-measure of 66.5% and a MCC of 0.329. Both in cross-validation and independent testing, the second SVM model that used both DNA and protein sequence data showed better performance than the first model that used DNA sequence data. To the best of

  10. Comparison of Predictive Modeling Methods of Aircraft Landing Speed

    Science.gov (United States)

    Diallo, Ousmane H.

    2012-01-01

    Expected increases in air traffic demand have stimulated the development of air traffic control tools intended to assist the air traffic controller in accurately and precisely spacing aircraft landing at congested airports. Such tools will require an accurate landing-speed prediction to increase throughput while decreasing necessary controller interventions for avoiding separation violations. There are many practical challenges to developing an accurate landing-speed model that has acceptable prediction errors. This paper discusses the development of a near-term implementation, using readily available information, to estimate/model final approach speed from the top of the descent phase of flight to the landing runway. As a first approach, all variables found to contribute directly to the landing-speed prediction model are used to build a multi-regression technique of the response surface equation (RSE). Data obtained from operations of a major airlines for a passenger transport aircraft type to the Dallas/Fort Worth International Airport are used to predict the landing speed. The approach was promising because it decreased the standard deviation of the landing-speed error prediction by at least 18% from the standard deviation of the baseline error, depending on the gust condition at the airport. However, when the number of variables is reduced to the most likely obtainable at other major airports, the RSE model shows little improvement over the existing methods. Consequently, a neural network that relies on a nonlinear regression technique is utilized as an alternative modeling approach. For the reduced number of variables cases, the standard deviation of the neural network models errors represent over 5% reduction compared to the RSE model errors, and at least 10% reduction over the baseline predicted landing-speed error standard deviation. Overall, the constructed models predict the landing-speed more accurately and precisely than the current state-of-the-art.

  11. Rapid and accurate prediction and scoring of water molecules in protein binding sites.

    Directory of Open Access Journals (Sweden)

    Gregory A Ross

    Full Text Available Water plays a critical role in ligand-protein interactions. However, it is still challenging to predict accurately not only where water molecules prefer to bind, but also which of those water molecules might be displaceable. The latter is often seen as a route to optimizing affinity of potential drug candidates. Using a protocol we call WaterDock, we show that the freely available AutoDock Vina tool can be used to predict accurately the binding sites of water molecules. WaterDock was validated using data from X-ray crystallography, neutron diffraction and molecular dynamics simulations and correctly predicted 97% of the water molecules in the test set. In addition, we combined data-mining, heuristic and machine learning techniques to develop probabilistic water molecule classifiers. When applied to WaterDock predictions in the Astex Diverse Set of protein ligand complexes, we could identify whether a water molecule was conserved or displaced to an accuracy of 75%. A second model predicted whether water molecules were displaced by polar groups or by non-polar groups to an accuracy of 80%. These results should prove useful for anyone wishing to undertake rational design of new compounds where the displacement of water molecules is being considered as a route to improved affinity.

  12. The Dissolved Oxygen Prediction Method Based on Neural Network

    Directory of Open Access Journals (Sweden)

    Zhong Xiao

    2017-01-01

    Full Text Available The dissolved oxygen (DO is oxygen dissolved in water, which is an important factor for the aquaculture. Using BP neural network method with the combination of purelin, logsig, and tansig activation functions is proposed for the prediction of aquaculture’s dissolved oxygen. The input layer, hidden layer, and output layer are introduced in detail including the weight adjustment process. The breeding data of three ponds in actual 10 consecutive days were used for experiments; these ponds were located in Beihai, Guangxi, a traditional aquaculture base in southern China. The data of the first 7 days are used for training, and the data of the latter 3 days are used for the test. Compared with the common prediction models, curve fitting (CF, autoregression (AR, grey model (GM, and support vector machines (SVM, the experimental results show that the prediction accuracy of the neural network is the highest, and all the predicted values are less than 5% of the error limit, which can meet the needs of practical applications, followed by AR, GM, SVM, and CF. The prediction model can help to improve the water quality monitoring level of aquaculture which will prevent the deterioration of water quality and the outbreak of disease.

  13. Benchmarking pKa prediction methods for Lys115 in acetoacetate decarboxylase.

    Science.gov (United States)

    Liu, Yuli; Patel, Anand H G; Burger, Steven K; Ayers, Paul W

    2017-05-01

    Three different pK a prediction methods were used to calculate the pK a of Lys115 in acetoacetate decarboxylase (AADase): the empirical method PROPKA, the multiconformation continuum electrostatics (MCCE) method, and the molecular dynamics/thermodynamic integration (MD/TI) method with implicit solvent. As expected, accurate pK a prediction of Lys115 depends on the protonation patterns of other ionizable groups, especially the nearby Glu76. However, since the prediction methods do not explicitly sample the protonation patterns of nearby residues, this must be done manually. When Glu76 is deprotonated, all three methods give an incorrect pK a value for Lys115. If protonated, Glu76 is used in an MD/TI calculation, the pK a of Lys115 is predicted to be 5.3, which agrees well with the experimental value of 5.9. This result agrees with previous site-directed mutagenesis studies, where the mutation of Glu76 (negative charge when deprotonated) to Gln (neutral) causes no change in K m , suggesting that Glu76 has no effect on the pK a shift of Lys115. Thus, we postulate that the pK a of Glu76 is also shifted so that Glu76 is protonated (neutral) in AADase. Graphical abstract Simulated abundances of protonated species as pH is varied.

  14. A homology-based pipeline for global prediction of post-translational modification sites

    Science.gov (United States)

    Chen, Xiang; Shi, Shao-Ping; Xu, Hao-Dong; Suo, Sheng-Bao; Qiu, Jian-Ding

    2016-05-01

    The pathways of protein post-translational modifications (PTMs) have been shown to play particularly important roles for almost any biological process. Identification of PTM substrates along with information on the exact sites is fundamental for fully understanding or controlling biological processes. Alternative computational strategies would help to annotate PTMs in a high-throughput manner. Traditional algorithms are suited for identifying the common organisms and tissues that have a complete PTM atlas or extensive experimental data. While annotation of rare PTMs in most organisms is a clear challenge. In this work, to this end we have developed a novel homology-based pipeline named PTMProber that allows identification of potential modification sites for most of the proteomes lacking PTMs data. Cross-promotion E-value (CPE) as stringent benchmark has been used in our pipeline to evaluate homology to known modification sites. Independent-validation tests show that PTMProber achieves over 58.8% recall with high precision by CPE benchmark. Comparisons with other machine-learning tools show that PTMProber pipeline performs better on general predictions. In addition, we developed a web-based tool to integrate this pipeline at http://bioinfo.ncu.edu.cn/PTMProber/index.aspx. In addition to pre-constructed prediction models of PTM, the website provides an extensional functionality to allow users to customize models.

  15. Predicting surgical site infection after spine surgery: a validated model using a prospective surgical registry.

    Science.gov (United States)

    Lee, Michael J; Cizik, Amy M; Hamilton, Deven; Chapman, Jens R

    2014-09-01

    The impact of surgical site infection (SSI) is substantial. Although previous study has determined relative risk and odds ratio (OR) values to quantify risk factors, these values may be difficult to translate to the patient during counseling of surgical options. Ideally, a model that predicts absolute risk of SSI, rather than relative risk or OR values, would greatly enhance the discussion of safety of spine surgery. To date, there is no risk stratification model that specifically predicts the risk of medical complication. The purpose of this study was to create and validate a predictive model for the risk of SSI after spine surgery. This study performs a multivariate analysis of SSI after spine surgery using a large prospective surgical registry. Using the results of this analysis, this study will then create and validate a predictive model for SSI after spine surgery. The patient sample is from a high-quality surgical registry from our two institutions with prospectively collected, detailed demographic, comorbidity, and complication data. An SSI that required return to the operating room for surgical debridement. Using a prospectively collected surgical registry of more than 1,532 patients with extensive demographic, comorbidity, surgical, and complication details recorded for 2 years after the surgery, we identified several risk factors for SSI after multivariate analysis. Using the beta coefficients from those regression analyses, we created a model to predict the occurrence of SSI after spine surgery. We split our data into two subsets for internal and cross-validation of our model. We created a predictive model based on our beta coefficients from our multivariate analysis. The final predictive model for SSI had a receiver-operator curve characteristic of 0.72, considered to be a fair measure. The final model has been uploaded for use on SpineSage.com. We present a validated model for predicting SSI after spine surgery. The value in this model is that it gives

  16. A method of predicting the reliability of CDM coil insulation

    International Nuclear Information System (INIS)

    Kytasty, A.; Ogle, C.; Arrendale, H.

    1992-01-01

    This paper presents a method of predicting the reliability of the Collider Dipole Magnet (CDM) coil insulation design. The method proposes a probabilistic treatment of electrical test data, stress analysis, material properties variability and loading uncertainties to give the reliability estimate. The approach taken to predict reliability of design related failure modes of the CDM is to form analytical models of the various possible failure modes and their related mechanisms or causes, and then statistically assess the contributions of the various contributing variables. The probability of the failure mode occurring is interpreted as the number of times one would expect certain extreme situations to combine and randomly occur. One of the more complex failure modes of the CDM will be used to illustrate this methodology

  17. Drug-Target Interactions: Prediction Methods and Applications.

    Science.gov (United States)

    Anusuya, Shanmugam; Kesherwani, Manish; Priya, K Vishnu; Vimala, Antonydhason; Shanmugam, Gnanendra; Velmurugan, Devadasan; Gromiha, M Michael

    2018-01-01

    Identifying the interactions between drugs and target proteins is a key step in drug discovery. This not only aids to understand the disease mechanism, but also helps to identify unexpected therapeutic activity or adverse side effects of drugs. Hence, drug-target interaction prediction becomes an essential tool in the field of drug repurposing. The availability of heterogeneous biological data on known drug-target interactions enabled many researchers to develop various computational methods to decipher unknown drug-target interactions. This review provides an overview on these computational methods for predicting drug-target interactions along with available webservers and databases for drug-target interactions. Further, the applicability of drug-target interactions in various diseases for identifying lead compounds has been outlined. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  18. Risk prediction, safety analysis and quantitative probability methods - a caveat

    International Nuclear Information System (INIS)

    Critchley, O.H.

    1976-01-01

    Views are expressed on the use of quantitative techniques for the determination of value judgements in nuclear safety assessments, hazard evaluation, and risk prediction. Caution is urged when attempts are made to quantify value judgements in the field of nuclear safety. Criteria are given the meaningful application of reliability methods but doubts are expressed about their application to safety analysis, risk prediction and design guidances for experimental or prototype plant. Doubts are also expressed about some concomitant methods of population dose evaluation. The complexities of new designs of nuclear power plants make the problem of safety assessment more difficult but some possible approaches are suggested as alternatives to the quantitative techniques criticized. (U.K.)

  19. Water hammer prediction and control: the Green's function method

    Science.gov (United States)

    Xuan, Li-Jun; Mao, Feng; Wu, Jie-Zhi

    2012-04-01

    By Green's function method we show that the water hammer (WH) can be analytically predicted for both laminar and turbulent flows (for the latter, with an eddy viscosity depending solely on the space coordinates), and thus its hazardous effect can be rationally controlled and minimized. To this end, we generalize a laminar water hammer equation of Wang et al. (J. Hydrodynamics, B2, 51, 1995) to include arbitrary initial condition and variable viscosity, and obtain its solution by Green's function method. The predicted characteristic WH behaviors by the solutions are in excellent agreement with both direct numerical simulation of the original governing equations and, by adjusting the eddy viscosity coefficient, experimentally measured turbulent flow data. Optimal WH control principle is thereby constructed and demonstrated.

  20. River Flow Prediction Using the Nearest Neighbor Probabilistic Ensemble Method

    Directory of Open Access Journals (Sweden)

    H. Sanikhani

    2016-02-01

    Full Text Available Introduction: In the recent years, researchers interested on probabilistic forecasting of hydrologic variables such river flow.A probabilistic approach aims at quantifying the prediction reliability through a probability distribution function or a prediction interval for the unknown future value. The evaluation of the uncertainty associated to the forecast is seen as a fundamental information, not only to correctly assess the prediction, but also to compare forecasts from different methods and to evaluate actions and decisions conditionally on the expected values. Several probabilistic approaches have been proposed in the literature, including (1 methods that use resampling techniques to assess parameter and model uncertainty, such as the Metropolis algorithm or the Generalized Likelihood Uncertainty Estimation (GLUE methodology for an application to runoff prediction, (2 methods based on processing the forecast errors of past data to produce the probability distributions of future values and (3 methods that evaluate how the uncertainty propagates from the rainfall forecast to the river discharge prediction, as the Bayesian forecasting system. Materials and Methods: In this study, two different probabilistic methods are used for river flow prediction.Then the uncertainty related to the forecast is quantified. One approach is based on linear predictors and in the other, nearest neighbor was used. The nonlinear probabilistic ensemble can be used for nonlinear time series analysis using locally linear predictors, while NNPE utilize a method adapted for one step ahead nearest neighbor methods. In this regard, daily river discharge (twelve years of Dizaj and Mashin Stations on Baranduz-Chay basin in west Azerbijan and Zard-River basin in Khouzestan provinces were used, respectively. The first six years of data was applied for fitting the model. The next three years was used to calibration and the remained three yeas utilized for testing the models

  1. RADIOLYTIC HYDROGEN GENERATION INSAVANNAH RIVER SITE (SRS) HIGH LEVEL WASTETANKS COMPARISON OF SRS AND HANFORDMODELING PREDICTIONS

    Energy Technology Data Exchange (ETDEWEB)

    Crawford, C; Ned Bibler, N

    2009-04-15

    In the high level waste tanks at the Savannah River Site (SRS), hydrogen is produced continuously by interaction of the radiation in the tank with water in the waste. Consequently, the vapor spaces of the tanks are purged to prevent the accumulation of H{sub 2} and possible formation of a flammable mixture in a tank. Personnel at SRS have developed an empirical model to predict the rate of H{sub 2} formation in a tank. The basis of this model is the prediction of the G value for H{sub 2} production. This G value is the number of H{sub 2} molecules produced per 100 eV of radiolytic energy absorbed by the waste. Based on experimental studies it was found that the G value for H{sub 2} production from beta radiation and from gamma radiation were essentially equal. The G value for H{sub 2} production from alpha radiation was somewhat higher. Thus, the model has two equations, one for beta/gamma radiation and one for alpha radiation. Experimental studies have also indicated that both G values are decreased by the presence of nitrate and nitrite ions in the waste. These are the main scavengers for the precursors of H{sub 2} in the waste; thus the equations that were developed predict G values for hydrogen production as a function of the concentrations of these two ions in waste. Knowing the beta/gamma and alpha heat loads in the waste allows one to predict the total generation rate for hydrogen in a tank. With this prediction a ventilation rate can be established for each tank to ensure that a flammable mixture is not formed in the vapor space in a tank. Recently personnel at Hanford have developed a slightly different model for predicting hydrogen G values. Their model includes the same precursor for H{sub 2} as the SRS model but also includes an additional precursor not in the SRS model. Including the second precursor for H{sub 2} leads to different empirical equations for predicting the G values for H{sub 2} as a function of the nitrate and nitrite concentrations in

  2. Coastline Mapping and Cultural Review to Predict Sea Level Rise Impact on Hawaiian Archeological Sites

    Science.gov (United States)

    Clinton, J.

    2017-12-01

    Much of Hawaii's history is recorded in archeological sites. Researchers and cultural practitioners have been studying and reconstructing significant archeological sites for generations. Climate change, and more specifically, sea level rise may threaten these sites. Our research records current sea levels and then projects possible consequences to these cultural monuments due to sea level rise. In this mixed methods study, research scientists, cultural practitioners, and secondary students use plane-table mapping techniques to create maps of coastlines and historic sites. Students compare historical records to these maps, analyze current sea level rise trends, and calculate future sea levels. They also gather data through interviews with community experts and kupuna (elders). If climate change continues at projected rates, some historic sites will be in danger of negative impact due to sea level rise. Knowing projected sea levels at specific sites allows for preventative action and contributes to raised awareness of the impacts of climate change to the Hawaiian Islands. Students will share results with the community and governmental agencies in hopes of inspiring action to minimize climate change. It will take collaboration between scientists and cultural communities to inspire future action on climate change.

  3. Toward a regional power plant siting method: AEC-Maryland regional siting factors study, FY 1974 progress report

    International Nuclear Information System (INIS)

    Yaffee, S.L.; Miller, C.A.

    1974-11-01

    The ''AEC-Maryland Regional Siting Factors Study'' examines the process of siting in a regional context. It is developing an analysis method to delineate candidate areas for siting of several power plant technology packages, including both fossil-fueled and nuclear options. Tools that are being used include simulation modeling, economic and demographic forecasting, spatial analysis, and computer graphics and numerical manipulation. The approach will describe the trade-offs incurred if a power plant is located in one candidate area rather than in another. In FY 1974, a suitability analysis method was developed which uses engineering and environmental parameters to define a level of environmental cost incurred if a segment of land is used to site a specific technology package. (U.S.)

  4. Improving protein function prediction methods with integrated literature data

    Directory of Open Access Journals (Sweden)

    Gabow Aaron P

    2008-04-01

    Full Text Available Abstract Background Determining the function of uncharacterized proteins is a major challenge in the post-genomic era due to the problem's complexity and scale. Identifying a protein's function contributes to an understanding of its role in the involved pathways, its suitability as a drug target, and its potential for protein modifications. Several graph-theoretic approaches predict unidentified functions of proteins by using the functional annotations of better-characterized proteins in protein-protein interaction networks. We systematically consider the use of literature co-occurrence data, introduce a new method for quantifying the reliability of co-occurrence and test how performance differs across species. We also quantify changes in performance as the prediction algorithms annotate with increased specificity. Results We find that including information on the co-occurrence of proteins within an abstract greatly boosts performance in the Functional Flow graph-theoretic function prediction algorithm in yeast, fly and worm. This increase in performance is not simply due to the presence of additional edges since supplementing protein-protein interactions with co-occurrence data outperforms supplementing with a comparably-sized genetic interaction dataset. Through the combination of protein-protein interactions and co-occurrence data, the neighborhood around unknown proteins is quickly connected to well-characterized nodes which global prediction algorithms can exploit. Our method for quantifying co-occurrence reliability shows superior performance to the other methods, particularly at threshold values around 10% which yield the best trade off between coverage and accuracy. In contrast, the traditional way of asserting co-occurrence when at least one abstract mentions both proteins proves to be the worst method for generating co-occurrence data, introducing too many false positives. Annotating the functions with greater specificity is harder

  5. CREME96 and Related Error Rate Prediction Methods

    Science.gov (United States)

    Adams, James H., Jr.

    2012-01-01

    Predicting the rate of occurrence of single event effects (SEEs) in space requires knowledge of the radiation environment and the response of electronic devices to that environment. Several analytical models have been developed over the past 36 years to predict SEE rates. The first error rate calculations were performed by Binder, Smith and Holman. Bradford and Pickel and Blandford, in their CRIER (Cosmic-Ray-Induced-Error-Rate) analysis code introduced the basic Rectangular ParallelePiped (RPP) method for error rate calculations. For the radiation environment at the part, both made use of the Cosmic Ray LET (Linear Energy Transfer) spectra calculated by Heinrich for various absorber Depths. A more detailed model for the space radiation environment within spacecraft was developed by Adams and co-workers. This model, together with a reformulation of the RPP method published by Pickel and Blandford, was used to create the CR ME (Cosmic Ray Effects on Micro-Electronics) code. About the same time Shapiro wrote the CRUP (Cosmic Ray Upset Program) based on the RPP method published by Bradford. It was the first code to specifically take into account charge collection from outside the depletion region due to deformation of the electric field caused by the incident cosmic ray. Other early rate prediction methods and codes include the Single Event Figure of Merit, NOVICE, the Space Radiation code and the effective flux method of Binder which is the basis of the SEFA (Scott Effective Flux Approximation) model. By the early 1990s it was becoming clear that CREME and the other early models needed Revision. This revision, CREME96, was completed and released as a WWW-based tool, one of the first of its kind. The revisions in CREME96 included improved environmental models and improved models for calculating single event effects. The need for a revision of CREME also stimulated the development of the CHIME (CRRES/SPACERAD Heavy Ion Model of the Environment) and MACREE (Modeling and

  6. Predictability of bone density at posterior mandibular implant sites using cone-beam computed tomography intensity values.

    Science.gov (United States)

    Alkhader, Mustafa; Hudieb, Malik; Khader, Yousef

    2017-01-01

    The aim of this study was to investigate the predictability of bone density at posterior mandibular implant sites using cone-beam computed tomography (CBCT) intensity values. CBCT cross-sectional images for 436 posterior mandibular implant sites were selected for the study. Using Invivo software (Anatomage, San Jose, California, USA), two observers classified the bone density into three categories: low, intermediate, and high, and CBCT intensity values were generated. Based on the consensus of the two observers, 15.6% of sites were of low bone density, 47.9% were of intermediate density, and 36.5% were of high density. Receiver-operating characteristic analysis showed that CBCT intensity values had a high predictive power for predicting high density sites (area under the curve [AUC] =0.94, P < 0.005) and intermediate density sites (AUC = 0.81, P < 0.005). The best cut-off value for intensity to predict intermediate density sites was 218 (sensitivity = 0.77 and specificity = 0.76) and the best cut-off value for intensity to predict high density sites was 403 (sensitivity = 0.93 and specificity = 0.77). CBCT intensity values are considered useful for predicting bone density at posterior mandibular implant sites.

  7. GSHSite: exploiting an iteratively statistical method to identify s-glutathionylation sites with substrate specificity.

    Directory of Open Access Journals (Sweden)

    Yi-Ju Chen

    Full Text Available S-glutathionylation, the covalent attachment of a glutathione (GSH to the sulfur atom of cysteine, is a selective and reversible protein post-translational modification (PTM that regulates protein activity, localization, and stability. Despite its implication in the regulation of protein functions and cell signaling, the substrate specificity of cysteine S-glutathionylation remains unknown. Based on a total of 1783 experimentally identified S-glutathionylation sites from mouse macrophages, this work presents an informatics investigation on S-glutathionylation sites including structural factors such as the flanking amino acids composition and the accessible surface area (ASA. TwoSampleLogo presents that positively charged amino acids flanking the S-glutathionylated cysteine may influence the formation of S-glutathionylation in closed three-dimensional environment. A statistical method is further applied to iteratively detect the conserved substrate motifs with statistical significance. Support vector machine (SVM is then applied to generate predictive model considering the substrate motifs. According to five-fold cross-validation, the SVMs trained with substrate motifs could achieve an enhanced sensitivity, specificity, and accuracy, and provides a promising performance in an independent test set. The effectiveness of the proposed method is demonstrated by the correct identification of previously reported S-glutathionylation sites of mouse thioredoxin (TXN and human protein tyrosine phosphatase 1b (PTP1B. Finally, the constructed models are adopted to implement an effective web-based tool, named GSHSite (http://csb.cse.yzu.edu.tw/GSHSite/, for identifying uncharacterized GSH substrate sites on the protein sequences.

  8. Prediction of pollution into ore bearing aquifer from ISL-site

    International Nuclear Information System (INIS)

    Dolgopolov, V.; Kayukov, P.; Vyatchennikova, L.; Shishkov, I.

    2002-01-01

    Outcomes of present mineralogical investigations explain some phenomena observed in previous sorption experiences. First of all they elucidate cutback of filtration coefficients of sediments exposed under pollution process as granulometric fractions of these permeable sediments decreased in its weight ratio in according to impermeable ones. Then sorption graphics are explained by chemical and mineralogical changes in experienced substance. Explores made diagnoses and estimation about 45 authigenous and technogenous minerals playing the essential role in the process of mass transfer of uranium and polluted components in ISL-sites. They are considered in all basic fractions of studied sediments. Bright peculiarities are noticed for gravel and coarse sands, which went to pieces as a result of dilution or decomposition of its cement. Mineral composition of clayey fraction also significantly changed after pollution process. Steep lost of weight of fraction 0.005-0.001mm consisted with partial decomposition of clayey minerals basically of montmorillonite. These investigations show also that ones of the main minerals playing such the essential role are sorptive such as silica, silica gel, common opal, opal-CT, noble opal, allophane, montmorillonite and zeolite minerals. In conclusion the given work shows some filtration properties of aquifer in an ISL-site and predicts the spread of a possible pollution for ISL-site to environment. The extension of pollution could be more one-two dimension of affected by ISL-process. The prediction would be exact if the previous investigation included the desorptive characteristics and sorptive data depending on temperature and pressure. Current investigations continued pervious ones after which some inexplicable matters of migration of contaminated halo from an ISL-polygon are as follows: Identification and specification of all minerals outside ISL-polygon; Calculation of mineral balance for contaminated sites; Specification of flow

  9. Virtual Screening and Prediction of Site of Metabolism for Cytochrome P450 1A2 Ligands

    DEFF Research Database (Denmark)

    Vasanthanathan, P.; Hritz, Jozef; Taboureau, Olivier

    2009-01-01

    questions have been addressed: 1. Binding orientations and conformations were successfully predicted for various substrates. 2. A virtual screen was performed with satisfying enrichment rates. 3. A classification of individual compounds into active and inactive was performed. It was found that while docking...... can be used successfully to address the first two questions, it seems to be more difficult to perform the classification. Different scoring functions were included, and the well-characterized water molecule in the active site was included in various ways. Results are compared to experimental data...

  10. Predicting success of oligomerized pool engineering (OPEN for zinc finger target site sequences

    Directory of Open Access Journals (Sweden)

    Goodwin Mathew J

    2010-11-01

    Full Text Available Abstract Background Precise and efficient methods for gene targeting are critical for detailed functional analysis of genomes and regulatory networks and for potentially improving the efficacy and safety of gene therapies. Oligomerized Pool ENgineering (OPEN is a recently developed method for engineering C2H2 zinc finger proteins (ZFPs designed to bind specific DNA sequences with high affinity and specificity in vivo. Because generation of ZFPs using OPEN requires considerable effort, a computational method for identifying the sites in any given gene that are most likely to be successfully targeted by this method is desirable. Results Analysis of the base composition of experimentally validated ZFP target sites identified important constraints on the DNA sequence space that can be effectively targeted using OPEN. Using alternate encodings to represent ZFP target sites, we implemented Naïve Bayes and Support Vector Machine classifiers capable of distinguishing "active" targets, i.e., ZFP binding sites that can be targeted with a high rate of success, from those that are "inactive" or poor targets for ZFPs generated using current OPEN technologies. When evaluated using leave-one-out cross-validation on a dataset of 135 experimentally validated ZFP target sites, the best Naïve Bayes classifier, designated ZiFOpT, achieved overall accuracy of 87% and specificity+ of 90%, with an ROC AUC of 0.89. When challenged with a completely independent test set of 140 newly validated ZFP target sites, ZiFOpT performance was comparable in terms of overall accuracy (88% and specificity+ (92%, but with reduced ROC AUC (0.77. Users can rank potentially active ZFP target sites using a confidence score derived from the posterior probability returned by ZiFOpT. Conclusion ZiFOpT, a machine learning classifier trained to identify DNA sequences amenable for targeting by OPEN-generated zinc finger arrays, can guide users to target sites that are most likely to function

  11. Waste classification and methods applied to specific disposal sites

    International Nuclear Information System (INIS)

    Rogers, V.C.

    1979-01-01

    An adequate definition of the classes of radioactive wastes is necessary to regulating the disposal of radioactive wastes. A classification system is proposed in which wastes are classified according to characteristics relating to their disposal. Several specific sites are analyzed with the methodology in order to gain insights into the classification of radioactive wastes. Also presented is the analysis of ocean dumping as it applies to waste classification. 5 refs

  12. Predicting Nitrate Transport under Future Climate Scenarios beneath the Nebraska Management Systems Evaluation Area (MSEA) site

    Science.gov (United States)

    Li, Y.; Akbariyeh, S.; Gomez Peña, C. A.; Bartlet-Hunt, S.

    2017-12-01

    Understanding the impacts of future climate change on soil hydrological processes and solute transport is crucial to develop appropriate strategies to minimize adverse impacts of agricultural activities on groundwater quality. The goal of this work is to evaluate the direct effects of climate change on the fate and transport of nitrate beneath a center-pivot irrigated corn field in Nebraska Management Systems Evaluation Area (MSEA) site. Future groundwater recharge rate and actual evapotranspiration rate were predicted based on an inverse modeling approach using climate data generated by Weather Research and Forecasting (WRF) model under the RCP 8.5 scenario, which was downscaled from global CCSM4 model to a resolution of 24 by 24 km2. A groundwater flow model was first calibrated based on historical groundwater table measurement and was then applied to predict future groundwater table in the period 2057-2060. Finally, predicted future groundwater recharge rate, actual evapotranspiration rate, and groundwater level, together with future precipitation data from WRF, were used in a three-dimensional (3D) model, which was validated based on rich historic data set collected from 1993-1996, to predict nitrate concentration in soil and groundwater from the year 2057 to 2060. Future groundwater recharge was found to be decreasing in the study area compared to average groundwater recharge data from the literature. Correspondingly, groundwater elevation was predicted to decrease (1 to 2 ft) over the five years of simulation. Predicted higher transpiration data from climate model resulted in lower infiltration of nitrate concentration in subsurface within the root zone.

  13. Evaluation of the DayCent model to predict carbon fluxes in French crop sites

    Science.gov (United States)

    Fujisaki, Kenji; Martin, Manuel P.; Zhang, Yao; Bernoux, Martial; Chapuis-Lardy, Lydie

    2017-04-01

    Croplands in temperate regions are an important component of the carbon balance and can act as a sink or a source of carbon, depending on pedoclimatic conditions and management practices. Therefore the evaluation of carbon fluxes in croplands by modelling approach is relevant in the context of global change. This study was part of the Comete-Global project funded by the multi-Partner call FACCE JPI. Carbon fluxes, net ecosystem exchange (NEE), leaf area index (LAI), biomass, and grain production were simulated at the site level in three French crop experiments from the CarboEurope project. Several crops were studied, like winter wheat, rapeseed, barley, maize, and sunflower. Daily NEE was measured with eddy covariance and could be partitioned between gross primary production (GPP) and total ecosystem respiration (TER). Measurements were compared to DayCent simulations, a process-based model predicting plant production and soil organic matter turnover at daily time step. We compared two versions of the model: the original one with a simplified plant module and a newer version that simulates LAI. Input data for modelling were soil properties, climate, and management practices. Simulations of grain yields and biomass production were acceptable when using optimized crop parameters. Simulation of NEE was also acceptable. GPP predictions were improved with the newer version of the model, eliminating temporal shifts that could be observed with the original model. TER was underestimated by the model. Predicted NEE was more sensitive to soil tillage and nitrogen applications than measured NEE. DayCent was therefore a relevant tool to predict carbon fluxes in French crops at the site level. The introduction of LAI in the model improved its performance.

  14. IRESPred: Web Server for Prediction of Cellular and Viral Internal Ribosome Entry Site (IRES)

    Science.gov (United States)

    Kolekar, Pandurang; Pataskar, Abhijeet; Kulkarni-Kale, Urmila; Pal, Jayanta; Kulkarni, Abhijeet

    2016-01-01

    Cellular mRNAs are predominantly translated in a cap-dependent manner. However, some viral and a subset of cellular mRNAs initiate their translation in a cap-independent manner. This requires presence of a structured RNA element, known as, Internal Ribosome Entry Site (IRES) in their 5′ untranslated regions (UTRs). Experimental demonstration of IRES in UTR remains a challenging task. Computational prediction of IRES merely based on sequence and structure conservation is also difficult, particularly for cellular IRES. A web server, IRESPred is developed for prediction of both viral and cellular IRES using Support Vector Machine (SVM). The predictive model was built using 35 features that are based on sequence and structural properties of UTRs and the probabilities of interactions between UTR and small subunit ribosomal proteins (SSRPs). The model was found to have 75.51% accuracy, 75.75% sensitivity, 75.25% specificity, 75.75% precision and Matthews Correlation Coefficient (MCC) of 0.51 in blind testing. IRESPred was found to perform better than the only available viral IRES prediction server, VIPS. The IRESPred server is freely available at http://bioinfo.net.in/IRESPred/. PMID:27264539

  15. Comparison of RF spectrum prediction methods for dynamic spectrum access

    Science.gov (United States)

    Kovarskiy, Jacob A.; Martone, Anthony F.; Gallagher, Kyle A.; Sherbondy, Kelly D.; Narayanan, Ram M.

    2017-05-01

    Dynamic spectrum access (DSA) refers to the adaptive utilization of today's busy electromagnetic spectrum. Cognitive radio/radar technologies require DSA to intelligently transmit and receive information in changing environments. Predicting radio frequency (RF) activity reduces sensing time and energy consumption for identifying usable spectrum. Typical spectrum prediction methods involve modeling spectral statistics with Hidden Markov Models (HMM) or various neural network structures. HMMs describe the time-varying state probabilities of Markov processes as a dynamic Bayesian network. Neural Networks model biological brain neuron connections to perform a wide range of complex and often non-linear computations. This work compares HMM, Multilayer Perceptron (MLP), and Recurrent Neural Network (RNN) algorithms and their ability to perform RF channel state prediction. Monte Carlo simulations on both measured and simulated spectrum data evaluate the performance of these algorithms. Generalizing spectrum occupancy as an alternating renewal process allows Poisson random variables to generate simulated data while energy detection determines the occupancy state of measured RF spectrum data for testing. The results suggest that neural networks achieve better prediction accuracy and prove more adaptable to changing spectral statistics than HMMs given sufficient training data.

  16. Methods for predicting isochronous stress-strain curves

    International Nuclear Information System (INIS)

    Kiyoshige, Masanori; Shimizu, Shigeki; Satoh, Keisuke.

    1976-01-01

    Isochronous stress-strain curves show the relation between stress and total strain at a certain temperature with time as a parameter, and they are drawn up from the creep test results at various stress levels at a definite temperature. The concept regarding the isochronous stress-strain curves was proposed by McVetty in 1930s, and has been used for the design of aero-engines. Recently the high temperature characteristics of materials are shown as the isochronous stress-strain curves in the design guide for the nuclear energy equipments and structures used in high temperature creep region. It is prescribed that these curves are used as the criteria for determining design stress intensity or the data for analyzing the superposed effects of creep and fatigue. In case of the isochronous stress-strain curves used for the design of nuclear energy equipments with very long service life, it is impractical to determine the curves directly from the results of long time creep test, accordingly the method of predicting long time stress-strain curves from short time creep test results must be established. The method proposed by the authors, for which the creep constitution equations taking the first and second creep stages into account are used, and the method using Larson-Miller parameter were studied, and it was found that both methods were reliable for the prediction. (Kako, I.)

  17. A method for site-dependent planning and its application to the preselection of sites for thermal power plants

    International Nuclear Information System (INIS)

    Friedrich, R.

    1979-01-01

    In the first part of the paper a computer-aided method for dealing with the problems of site-dependent planning is described. By means of the modular program system COPLAN complex conjunction between locally varying data can be performed rapidly and accurately with respect to spatial orientation. The system consists of data input, numerous ways of processing, and graphical representation of the results. The second part shows the application of the system to preselection of sites for thermal power plants. By means of a method analyzing its usefulness, the suitability of each point in (the German Federal State of) Baden-Wuerttemberg as a power plant site is determined. Compared with the currently used methods of preliminary site selection the present method is distinguished by area-covering calculation, the possibility of balancing up advantages and disadvantages, as well as transparency and suitability for being checked up. The paper establishes and considers criteria from the fields of operational economy, safety, ecology, and district planning. The computations are performed for different orders of preference. It is shown that there are regions of sites which are acceptable with respect to a large spectrum of object systems. (orig.) [de

  18. A Lifetime Prediction Method for LEDs Considering Real Mission Profiles

    DEFF Research Database (Denmark)

    Qu, Xiaohui; Wang, Huai; Zhan, Xiaoqing

    2017-01-01

    operations due to the varying operational and environmental conditions during the entire service time (i.e., mission profiles). To overcome the challenge, this paper proposes an advanced lifetime prediction method, which takes into account the field operation mission profiles and also the statistical......The Light-Emitting Diode (LED) has become a very promising alternative lighting source with the advantages of longer lifetime and higher efficiency than traditional ones. The lifetime prediction of LEDs is important to guide the LED system designers to fulfill the design specifications...... properties of the life data available from accelerated degradation testing. The electrical and thermal characteristics of LEDs are measured by a T3Ster system, used for the electro-thermal modeling. It also identifies key variables (e.g., heat sink parameters) that can be designed to achieve a specified...

  19. Long-Term Prediction of Satellite Orbit Using Analytical Method

    Directory of Open Access Journals (Sweden)

    Jae-Cheol Yoon

    1997-12-01

    Full Text Available A long-term prediction algorithm of geostationary orbit was developed using the analytical method. The perturbation force models include geopotential upto fifth order and degree and luni-solar gravitation, and solar radiation pressure. All of the perturbation effects were analyzed by secular variations, short-period variations, and long-period variations for equinoctial elements such as the semi-major axis, eccentricity vector, inclination vector, and mean longitude of the satellite. Result of the analytical orbit propagator was compared with that of the cowell orbit propagator for the KOREASAT. The comparison indicated that the analytical solution could predict the semi-major axis with an accuarcy of better than ~35meters over a period of 3 month.

  20. Nuclear site selection and environmental protection. The decision making methods

    International Nuclear Information System (INIS)

    Bresson, G.; Lacourly, G.; Fitoussi, L.

    1975-01-01

    The selection of the site of a nuclear plant most often comes to seek out and compound between two trends: that of the operator who will try and reduce the cost price of his product to the lowest and that of the protectionist who will try and reduce to the minimum the hazards resulting from the plant operation. Such a compromise is the result of a more or less empirical choice, which enters within the frame of a cost-benefit analysis, in which theoretically, the choice between several possible solutions is made of the selection giving the higher advantage [fr

  1. Review of geophysical characterization methods used at the Hanford Site

    Energy Technology Data Exchange (ETDEWEB)

    GV Last; DG Horton

    2000-03-23

    This paper presents a review of geophysical methods used at Hanford in two parts: (1) shallow surface-based geophysical methods and (2) borehole geophysical methods. This review was not intended to be ``all encompassing'' but should represent the vast majority (>90% complete) of geophysical work conducted onsite and aimed at hazardous waste investigations in the vadose zone and/or uppermost groundwater aquifers. This review did not cover geophysical methods aimed at large-scale geologic structures or seismicity and, in particular, did not include those efforts conducted in support of the Basalt Waste Isolation Program. This review focused primarily on the more recent efforts.

  2. Review of geophysical characterization methods used at the Hanford Site

    International Nuclear Information System (INIS)

    GV Last; DG Horton

    2000-01-01

    This paper presents a review of geophysical methods used at Hanford in two parts: (1) shallow surface-based geophysical methods and (2) borehole geophysical methods. This review was not intended to be ''all encompassing'' but should represent the vast majority (>90% complete) of geophysical work conducted onsite and aimed at hazardous waste investigations in the vadose zone and/or uppermost groundwater aquifers. This review did not cover geophysical methods aimed at large-scale geologic structures or seismicity and, in particular, did not include those efforts conducted in support of the Basalt Waste Isolation Program. This review focused primarily on the more recent efforts

  3. Prediction of Chloride Diffusion in Concrete Structure Using Meshless Methods

    Directory of Open Access Journals (Sweden)

    Ling Yao

    2016-01-01

    Full Text Available Degradation of RC structures due to chloride penetration followed by reinforcement corrosion is a serious problem in civil engineering. The numerical simulation methods at present mainly involve finite element methods (FEM, which are based on mesh generation. In this study, element-free Galerkin (EFG and meshless weighted least squares (MWLS methods are used to solve the problem of simulation of chloride diffusion in concrete. The range of a scaling parameter is presented using numerical examples based on meshless methods. One- and two-dimensional numerical examples validated the effectiveness and accuracy of the two meshless methods by comparing results obtained by MWLS with results computed by EFG and FEM and results calculated by an analytical method. A good agreement is obtained among MWLS and EFG numerical simulations and the experimental data obtained from an existing marine concrete structure. These results indicate that MWLS and EFG are reliable meshless methods that can be used for the prediction of chloride ingress in concrete structures.

  4. Fingerprint image reconstruction for swipe sensor using Predictive Overlap Method

    Directory of Open Access Journals (Sweden)

    Mardiansyah Ahmad Zafrullah

    2018-01-01

    Full Text Available Swipe sensor is one of many biometric authentication sensor types that widely applied to embedded devices. The sensor produces an overlap on every pixel block of the image, so the picture requires a reconstruction process before heading to the feature extraction process. Conventional reconstruction methods require extensive computation, causing difficult to apply to embedded devices that have limited computing process. In this paper, image reconstruction is proposed using predictive overlap method, which determines the image block shift from the previous set of change data. The experiments were performed using 36 images generated by a swipe sensor with 128 x 8 pixels size of the area, where each image has an overlap in each block. The results reveal computation can increase up to 86.44% compared with conventional methods, with accuracy decreasing to 0.008% in average.

  5. Pyramiding tumuli waste disposal site and method of construction thereof

    Science.gov (United States)

    Golden, Martin P.

    1989-01-01

    An improved waste disposal site for the above-ground disposal of low-level nuclear waste as disclosed herein. The disposal site is formed from at least three individual waste-containing tumuli, wherein each tumuli includes a central raised portion bordered by a sloping side portion. Two of the tumuli are constructed at ground level with adjoining side portions, and a third above-ground tumulus is constructed over the mutually adjoining side portions of the ground-level tumuli. Both the floor and the roof of each tumulus includes a layer of water-shedding material such as compacted clay, and the clay layer in the roofs of the two ground-level tumuli form the compacted clay layer of the floor of the third above-ground tumulus. Each tumulus further includes a shield wall, preferably formed from a solid array of low-level handleable nuclear wate packages. The provision of such a shield wall protects workers from potentially harmful radiation when higher-level, non-handleable packages of nuclear waste are stacked in the center of the tumulus.

  6. Explorability and predictability of the paleozoic sedimentary sequence beneath the Bruce nuclear site

    International Nuclear Information System (INIS)

    Parmenter, A.; Jensen, M.; Crowe, R.; Raven, K.

    2011-01-01

    Ontario Power Generation (OPG) is proposing to develop a Deep Geologic Repository (DGR) for the long-term management of its Low and Intermediate Level Waste (L&ILW) at the Bruce nuclear site located in the Municipality of Kincardine, Ontario. A 4-year program of geoscientific studies to assess the suitability of the 850 m thick Palaeozoic age sedimentary sequence beneath the site to host the DGR was completed in 2010. The studies provide evidence of a geologic setting in which the DGR concept would be safely implemented at a nominal depth of 680 m within the argillaceous limestone of the Cobourg Formation. This paper describes the geologic framework of the Bruce nuclear site with a focus on illustrating the high degree of stratigraphic continuity and traceability at site-specific and regional scales within the Ordovician sediments proposed to host and enclose the DGR. As part of the site-specific studies, a program of deep drilling/coring (6 boreholes) and in-situ testing through the sedimentary sequence was completed from 4 drill sites situated beyond the DGR footprint, approximately 1 km apart. Core logging reveals that the stratigraphic sequence comprises 34 distinct bedrock formations/members/units consistent with the known regional stratigraphic framework. These layered sedimentary formations dip 0.6 o (~10 m/km) to the southwest with highly uniform thicknesses both at the site- and regional-scale, particularly, the Ordovician sediments, which vary on the order of metres. The occurrence of steeply-dipping faults within the sedimentary sequence is not revealed through surface outcrop fracture mapping, micro-seismic (M ≥ 1) monitoring, inclined borehole coring or intersection of hydrothermal type dolomitized reservoir systems. Potential fault structures, interpreted from a 2-D seismic survey, were targeted by angled boreholes which found no evidence for their existence. Formation specific continuity is also evidence by the lateral traceability of physical rock

  7. Bicycle Frame Prediction Techniques with Fuzzy Logic Method

    Directory of Open Access Journals (Sweden)

    Rafiuddin Syam

    2015-03-01

    Full Text Available In general, an appropriate size bike frame would get comfort to the rider while biking. This study aims to predict the simulation system on the bike frame sizes with fuzzy logic. Testing method used is the simulation test. In this study, fuzzy logic will be simulated using Matlab language to test their performance. Mamdani fuzzy logic using 3 variables and 1 output variable intake. Triangle function for the input and output. The controller is designed in the type mamdani with max-min composition and the method deffuzification using center of gravity method. The results showed that height, inseam and Crank Size generating appropriate frame size for the rider associated with comfort. Has a height range between 142 cm and 201 cm. Inseam has a range between 64 cm and 97 cm. Crank has a size range between 175 mm and 180 mm. The simulation results have a range of frame sizes between 13 inches and 22 inches. By using the fuzzy logic can be predicted the size frame of bicycle suitable for the biker.

  8. Bicycle Frame Prediction Techniques with Fuzzy Logic Method

    Directory of Open Access Journals (Sweden)

    Rafiuddin Syam

    2017-03-01

    Full Text Available In general, an appropriate size bike frame would get comfort to the rider while biking. This study aims to predict the simulation system on the bike frame sizes with fuzzy logic. Testing method used is the simulation test. In this study, fuzzy logic will be simulated using Matlab language to test their performance. Mamdani fuzzy logic using 3 variables and 1 output variable intake. Triangle function for the input and output. The controller is designed in the type mamdani with max-min composition and the method deffuzification using center of gravity method. The results showed that height, inseam and Crank Size generating appropriate frame size for the rider associated with comfort. Has a height range between 142 cm and 201 cm. Inseam has a range between 64 cm and 97 cm. Crank has a size range between 175 mm and 180 mm. The simulation results have a range of frame sizes between 13 inches and 22 inches. By using the fuzzy logic can be predicted the size frame of bicycle suitable for the biker.

  9. A neural network based methodology to predict site-specific spectral acceleration values

    Science.gov (United States)

    Kamatchi, P.; Rajasankar, J.; Ramana, G. V.; Nagpal, A. K.

    2010-12-01

    A general neural network based methodology that has the potential to replace the computationally-intensive site-specific seismic analysis of structures is proposed in this paper. The basic framework of the methodology consists of a feed forward back propagation neural network algorithm with one hidden layer to represent the seismic potential of a region and soil amplification effects. The methodology is implemented and verified with parameters corresponding to Delhi city in India. For this purpose, strong ground motions are generated at bedrock level for a chosen site in Delhi due to earthquakes considered to originate from the central seismic gap of the Himalayan belt using necessary geological as well as geotechnical data. Surface level ground motions and corresponding site-specific response spectra are obtained by using a one-dimensional equivalent linear wave propagation model. Spectral acceleration values are considered as a target parameter to verify the performance of the methodology. Numerical studies carried out to validate the proposed methodology show that the errors in predicted spectral acceleration values are within acceptable limits for design purposes. The methodology is general in the sense that it can be applied to other seismically vulnerable regions and also can be updated by including more parameters depending on the state-of-the-art in the subject.

  10. ALTERNATIVE METHOD FOR ON SITE EVALUATION OF THERMAL TRANSMITTANCE

    Directory of Open Access Journals (Sweden)

    Aleksandar Janković

    2017-08-01

    Full Text Available Thermal transmittance or U-value is an indicator of the building envelope thermal properties and a key parameter for evaluation of heat losses through the building elements due to heat transmission. It can be determined by calculation based on thermal characteristics of the building element layers. However, this value does not take into account the effects of irregularities and degradation of certain elements of the envelope caused by aging, which may lead to errors in calculation of the heat losses. An effective and simple method for determination of thermal transmittance is in situ measurement, which is governed by the ISO 9869-1:2014 that defines heat flow meter method. This relatively expensive method leaves marks and damages surface of the building element. Furthermore, the final result is not always reliable, in particular when the building element is light or when the weather conditions are not suitable. In order to avoid the above mentioned problems and to estimate the real thermal transmittance value an alternative experimental method, here referred as the natural convection and radiation method, is proposed in this paper. For determination of thermal transmittance, this method requires only temperatures of inside and outside air, as well as the inner wall surface temperature. A detailed statistical analysis, performed by the software package SPSS ver. 20, shows several more advantages of this method comparing to the standard heat flow meter one, besides economic and non-destructive benefits.

  11. Alternative Testing Methods for Predicting Health Risk from Environmental Exposures

    Directory of Open Access Journals (Sweden)

    Annamaria Colacci

    2014-08-01

    Full Text Available Alternative methods to animal testing are considered as promising tools to support the prediction of toxicological risks from environmental exposure. Among the alternative testing methods, the cell transformation assay (CTA appears to be one of the most appropriate approaches to predict the carcinogenic properties of single chemicals, complex mixtures and environmental pollutants. The BALB/c 3T3 CTA shows a good degree of concordance with the in vivo rodent carcinogenesis tests. Whole-genome transcriptomic profiling is performed to identify genes that are transcriptionally regulated by different kinds of exposures. Its use in cell models representative of target organs may help in understanding the mode of action and predicting the risk for human health. Aiming at associating the environmental exposure to health-adverse outcomes, we used an integrated approach including the 3T3 CTA and transcriptomics on target cells, in order to evaluate the effects of airborne particulate matter (PM on toxicological complex endpoints. Organic extracts obtained from PM2.5 and PM1 samples were evaluated in the 3T3 CTA in order to identify effects possibly associated with different aerodynamic diameters or airborne chemical components. The effects of the PM2.5 extracts on human health were assessed by using whole-genome 44 K oligo-microarray slides. Statistical analysis by GeneSpring GX identified genes whose expression was modulated in response to the cell treatment. Then, modulated genes were associated with pathways, biological processes and diseases through an extensive biological analysis. Data derived from in vitro methods and omics techniques could be valuable for monitoring the exposure to toxicants, understanding the modes of action via exposure-associated gene expression patterns and to highlight the role of genes in key events related to adversity.

  12. Method for predicting peptide detection in mass spectrometry

    Science.gov (United States)

    Kangas, Lars [West Richland, WA; Smith, Richard D [Richland, WA; Petritis, Konstantinos [Richland, WA

    2010-07-13

    A method of predicting whether a peptide present in a biological sample will be detected by analysis with a mass spectrometer. The method uses at least one mass spectrometer to perform repeated analysis of a sample containing peptides from proteins with known amino acids. The method then generates a data set of peptides identified as contained within the sample by the repeated analysis. The method then calculates the probability that a specific peptide in the data set was detected in the repeated analysis. The method then creates a plurality of vectors, where each vector has a plurality of dimensions, and each dimension represents a property of one or more of the amino acids present in each peptide and adjacent peptides in the data set. Using these vectors, the method then generates an algorithm from the plurality of vectors and the calculated probabilities that specific peptides in the data set were detected in the repeated analysis. The algorithm is thus capable of calculating the probability that a hypothetical peptide represented as a vector will be detected by a mass spectrometry based proteomic platform, given that the peptide is present in a sample introduced into a mass spectrometer.

  13. Assessment of Soil Erosion Methods for Sludge Recovery, Savannah River Site

    National Research Council Canada - National Science Library

    Smith, Lawson

    1997-01-01

    ...) from selected storage tanks at the Savannah River Site (SRS) was assessed conceptually. Soil erosion methods are defined as the processes of soil detachment, entrainment, transport, and deposition...

  14. InterProSurf: a web server for predicting interacting sites on protein surfaces

    Science.gov (United States)

    Negi, Surendra S.; Schein, Catherine H.; Oezguen, Numan; Power, Trevor D.; Braun, Werner

    2009-01-01

    Summary A new web server, InterProSurf, predicts interacting amino acid residues in proteins that are most likely to interact with other proteins, given the 3D structures of subunits of a protein complex. The prediction method is based on solvent accessible surface area of residues in the isolated subunits, a propensity scale for interface residues and a clustering algorithm to identify surface regions with residues of high interface propensities. Here we illustrate the application of InterProSurf to determine which areas of Bacillus anthracis toxins and measles virus hemagglutinin protein interact with their respective cell surface receptors. The computationally predicted regions overlap with those regions previously identified as interface regions by sequence analysis and mutagenesis experiments. PMID:17933856

  15. A lifetime prediction method for LEDs considering mission profiles

    DEFF Research Database (Denmark)

    Qu, Xiaohui; Wang, Huai; Zhan, Xiaoqing

    2016-01-01

    and to benchmark the cost-competitiveness of different lighting technologies. The existing lifetime data released by LED manufacturers or standard organizations are usually applicable only for specific temperature and current levels. Significant lifetime discrepancies may be observed in field operations due...... to the varying operational and environmental conditions during the entire service time (i.e., mission profiles). To overcome the challenge, this paper proposes an advanced lifetime prediction method, which takes into account the field operation mission profiles and the statistical properties of the life data...

  16. Case history update: RCRA waste site remediation by telerobotic methods

    International Nuclear Information System (INIS)

    Yemington, C.R.; Stone, J.

    1992-01-01

    This paper presents a summary of the first 18 months of closure work at the Kerr Hollow Quarry site on the DOE reservation at Oak Ridge, Tennessee. Closure work includes recovery and processing of explosive, toxic and radioactive waste. As of January 1992, more than 10,000 items had been processed and removed from the quarry, exclusively by remotely operated equipment. Drums, buckets, tubing assemblies and other containers are being shredded to react any explosive contents. Concussion and projectiles are controlled by operating the shredder under 30 feet of water. The performance of the shredder, the effectiveness of the approach, production rates and maintenance requirements are addressed in the paper. To avoid exposing personnel to hazards, all work in the restricted area is done remotely. Two remotely operated vehicles were used to clear a pad, set a stand and install the 200-hp shredder. Some materials exposed by shredding are stable in water but react when exposed to air. In addition, radioactive items are mixed in with the other wastes. Safety considerations have therefore led to use of remote techniques for handling and examining materials after recovery. Deteriorated gas cylinders, which may contain pressurized toxic materials, are recovered and handled exclusively by remotely operated equipment. Waste retrieval work at the Kerr Hollow Quarry has proven the capability and cost-effectiveness of remotely operated equipment to deal with a wide variety of hazardous materials in an unstructured waste site environment. A mixture of radioactive materials, toxic chemicals, explosives and asbestos has been found and processed. Remotely operated vehicles have retrieved, sorted and processed more than 10,000 items including drums, buckets, pipe manifolds, gas cylinders and other containers

  17. Work characteristics predict the development of multi-site musculoskeletal pain.

    Science.gov (United States)

    Oakman, Jodi; de Wind, Astrid; van den Heuvel, Swenne G; van der Beek, Allard J

    2017-10-01

    Musculoskeletal pain in more than one body region is common and a barrier to sustaining employment. We aimed to examine whether work characteristics predict the development of multi-site pain (MSP), and to determine differences in work-related predictors between age groups. This study is based on 5136 employees from the Study on Transitions in Employment, Ability and Motivation (STREAM) who reported no MSP at baseline. Measures included physical, emotional, mental, and psychological job demands, social support and autonomy. Predictors of MSP were studied by logistic regression analyses. Univariate and multivariate analyses with age stratification (45-49, 50-54, 55-59, and 60-64 years) were done to explore differences between age groups. All work characteristics with the exception of autonomy were predictive of the development of MSP, with odds ratios varying from 1.21 (95% CI 1.04-1.40) for mental job demands to 1.63 (95% CI 1.43-1.86) for physical job demands. No clear pattern of age-related differences in the predictors of MSP emerged, with the exception of social support, which was predictive of MSP developing in all age groups except for the age group 60-64 years. Adverse physical and psychosocial work characteristics are associated with MSP. Organisations need to comprehensively assess work environments to ensure that all relevant workplace hazards, physical and psychosocial, are identified and then controlled for across all age groups.

  18. Streamflow prediction using multi-site rainfall obtained from hydroclimatic teleconnection

    Science.gov (United States)

    Kashid, S. S.; Ghosh, Subimal; Maity, Rajib

    2010-12-01

    SummarySimultaneous variations in weather and climate over widely separated regions are commonly known as "hydroclimatic teleconnections". Rainfall and runoff patterns, over continents, are found to be significantly teleconnected, with large-scale circulation patterns, through such hydroclimatic teleconnections. Though such teleconnections exist in nature, it is very difficult to model them, due to their inherent complexity. Statistical techniques and Artificial Intelligence (AI) tools gain popularity in modeling hydroclimatic teleconnection, based on their ability, in capturing the complicated relationship between the predictors (e.g. sea surface temperatures) and predictand (e.g., rainfall). Genetic Programming is such an AI tool, which is capable of capturing nonlinear relationship, between predictor and predictand, due to its flexible functional structure. In the present study, gridded multi-site weekly rainfall is predicted from El Niño Southern Oscillation (ENSO) indices, Equatorial Indian Ocean Oscillation (EQUINOO) indices, Outgoing Longwave Radiation (OLR) and lag rainfall at grid points, over the catchment, using Genetic Programming. The predicted rainfall is further used in a Genetic Programming model to predict streamflows. The model is applied for weekly forecasting of streamflow in Mahanadi River, India, and satisfactory performance is observed.

  19. Prediction of multiphase equilibria in associating fluids by a contact-site quasichemical equation of state

    International Nuclear Information System (INIS)

    Prikhodko, I.V.; Victorov, A.I.; Loos, Th.W.de

    1995-01-01

    A contract-site quasichemical equation of state has been used for the modeling of different kinds of fluid phase equilibria (between a gas phase and one or more liquids) over a wide range of conditions. Among the systems of interest are the ternary mixtures water + alkanols + hydrocarbons (alkanes or alkynes), water + alkanols (or acetone) + CO 2 , water + polyoxyethyleneglycol ethers + heavy alkanes. The model has been applied to describing the thermodynamic properties of the binary subsystems and to predict the phase behavior of the ternary systems. For longer-chain alkanols and hydrocarbons a group-contribution approach is implemented, which allows the modeling when no experimental data are available. The model gives reasonable predictions of phase behavior and the correct trends in the calculated phase diagrams in most cases. The concentrations of associates in liquid and gas phases are estimated by the model and compared with some experimental and computer simulation data. The predictive abilities of the model, its limitations, and possible ways of its improvement are discussed

  20. A strategy for interaction site prediction between phospho-binding modules and their partners identified from proteomic data.

    Science.gov (United States)

    Aucher, Willy; Becker, Emmanuelle; Ma, Emilie; Miron, Simona; Martel, Arnaud; Ochsenbein, Françoise; Marsolier-Kergoat, Marie-Claude; Guerois, Raphaël

    2010-12-01

    Small and large scale proteomic technologies are providing a wealth of potential interactions between proteins bearing phospho-recognition modules and their substrates. Resulting interaction maps reveal such a dense network of interactions that the functional dissection and understanding of these networks often require to break specific interactions while keeping the rest intact. Here, we developed a computational strategy, called STRIP, to predict the precise interaction site involved in an interaction with a phospho-recognition module. The method was validated by a two-hybrid screen carried out using the ForkHead Associated (FHA)1 domain of Rad53, a key protein of Saccharomyces cerevisiae DNA checkpoint, as a bait. In this screen we detected 11 partners, including Cdc7 and Cdc45, essential components of the DNA replication machinery. FHA domains are phospho-threonine binding modules and the threonines involved in both interactions could be predicted using the STRIP strategy. The threonines T484 and T189 in Cdc7 and Cdc45, respectively, were mutated and loss of binding could be monitored experimentally with the full-length proteins. The method was further tested for the analysis of 63 known Rad53 binding partners and provided several key insights regarding the threonines likely involved in these interactions. The STRIP method relies on a combination of conservation, phosphorylation likelihood, and binding specificity criteria and can be accessed via a web interface at http://biodev.extra.cea.fr/strip/.

  1. A Strategy for Interaction Site Prediction between Phospho-binding Modules and their Partners Identified from Proteomic Data*

    Science.gov (United States)

    Aucher, Willy; Becker, Emmanuelle; Ma, Emilie; Miron, Simona; Martel, Arnaud; Ochsenbein, Françoise; Marsolier-Kergoat, Marie-Claude; Guerois, Raphaël

    2010-01-01

    Small and large scale proteomic technologies are providing a wealth of potential interactions between proteins bearing phospho-recognition modules and their substrates. Resulting interaction maps reveal such a dense network of interactions that the functional dissection and understanding of these networks often require to break specific interactions while keeping the rest intact. Here, we developed a computational strategy, called STRIP, to predict the precise interaction site involved in an interaction with a phospho-recognition module. The method was validated by a two-hybrid screen carried out using the ForkHead Associated (FHA)1 domain of Rad53, a key protein of Saccharomyces cerevisiae DNA checkpoint, as a bait. In this screen we detected 11 partners, including Cdc7 and Cdc45, essential components of the DNA replication machinery. FHA domains are phospho-threonine binding modules and the threonines involved in both interactions could be predicted using the STRIP strategy. The threonines T484 and T189 in Cdc7 and Cdc45, respectively, were mutated and loss of binding could be monitored experimentally with the full-length proteins. The method was further tested for the analysis of 63 known Rad53 binding partners and provided several key insights regarding the threonines likely involved in these interactions. The STRIP method relies on a combination of conservation, phosphorylation likelihood, and binding specificity criteria and can be accessed via a web interface at http://biodev.extra.cea.fr/strip/. PMID:20733106

  2. Methods to develop site specific spectra and a review of the important parameters that influence the spectra

    International Nuclear Information System (INIS)

    Bernreuter, D.L.

    1979-05-01

    Problems with using risk analysis methodologies to estimate the seismic hazard at a site are discussed in the context of the U.S. Nuclear Regulatory Commission's Systematic Evaluation Program (SEP). Various methodologies that may reasonably define seismic hazard are outlined. The major assumptions that can lead to significant variations in the predicted hazard are identified. Guidance is provided to appropriate choices of parameters, and possible corrections that can extend the meager earthquake data base for sites located in the eastern United States are presented. A method that incorporates various interpretations of the same data is recommended

  3. Prediction strategies in a TV recommender system - Method and experiments

    NARCIS (Netherlands)

    van Setten, M.J.; Veenstra, M.; van Dijk, Elisabeth M.A.G.; Nijholt, Antinus; Isaísas, P.; Karmakar, N.

    2003-01-01

    Predicting the interests of a user in information is an important process in personalized information systems. In this paper, we present a way to create prediction engines that allow prediction techniques to be easily combined into prediction strategies. Prediction strategies choose one or a

  4. Method of draining water through a solid waste site without leaching

    Science.gov (United States)

    Treat, Russell L.; Gee, Glendon W.; Whyatt, Greg A.

    1993-01-01

    The present invention is a method of preventing water from leaching solid waste sites by preventing atmospheric precipitation from contacting waste as the water flows through a solid waste site. The method comprises placing at least one drain hole through the solid waste site. The drain hole is seated to prevent waste material from entering the drain hole, and the solid waste site cover material is layered and graded to direct water to flow toward the drain hole and to soil beneath the waste site.

  5. Data Based Prediction of Blood Glucose Concentrations Using Evolutionary Methods.

    Science.gov (United States)

    Hidalgo, J Ignacio; Colmenar, J Manuel; Kronberger, Gabriel; Winkler, Stephan M; Garnica, Oscar; Lanchares, Juan

    2017-08-08

    Predicting glucose values on the basis of insulin and food intakes is a difficult task that people with diabetes need to do daily. This is necessary as it is important to maintain glucose levels at appropriate values to avoid not only short-term, but also long-term complications of the illness. Artificial intelligence in general and machine learning techniques in particular have already lead to promising results in modeling and predicting glucose concentrations. In this work, several machine learning techniques are used for the modeling and prediction of glucose concentrations using as inputs the values measured by a continuous monitoring glucose system as well as also previous and estimated future carbohydrate intakes and insulin injections. In particular, we use the following four techniques: genetic programming, random forests, k-nearest neighbors, and grammatical evolution. We propose two new enhanced modeling algorithms for glucose prediction, namely (i) a variant of grammatical evolution which uses an optimized grammar, and (ii) a variant of tree-based genetic programming which uses a three-compartment model for carbohydrate and insulin dynamics. The predictors were trained and tested using data of ten patients from a public hospital in Spain. We analyze our experimental results using the Clarke error grid metric and see that 90% of the forecasts are correct (i.e., Clarke error categories A and B), but still even the best methods produce 5 to 10% of serious errors (category D) and approximately 0.5% of very serious errors (category E). We also propose an enhanced genetic programming algorithm that incorporates a three-compartment model into symbolic regression models to create smoothed time series of the original carbohydrate and insulin time series.

  6. Decision tree methods: applications for classification and prediction.

    Science.gov (United States)

    Song, Yan-Yan; Lu, Ying

    2015-04-25

    Decision tree methodology is a commonly used data mining method for establishing classification systems based on multiple covariates or for developing prediction algorithms for a target variable. This method classifies a population into branch-like segments that construct an inverted tree with a root node, internal nodes, and leaf nodes. The algorithm is non-parametric and can efficiently deal with large, complicated datasets without imposing a complicated parametric structure. When the sample size is large enough, study data can be divided into training and validation datasets. Using the training dataset to build a decision tree model and a validation dataset to decide on the appropriate tree size needed to achieve the optimal final model. This paper introduces frequently used algorithms used to develop decision trees (including CART, C4.5, CHAID, and QUEST) and describes the SPSS and SAS programs that can be used to visualize tree structure.

  7. Development of Tsunami PSA method for Korean NPP site

    International Nuclear Information System (INIS)

    Kim, Min Kyu; Choi, In Kil; Park, Jin Hee

    2010-01-01

    A methodology of tsunami PSA was developed in this study. A tsunami PSA consists of tsunami hazard analysis, tsunami fragility analysis and system analysis. In the case of tsunami hazard analysis, evaluation of tsunami return period is major task. For the evaluation of tsunami return period, numerical analysis and empirical method can be applied. The application of this method was applied to a nuclear power plant, Ulchin 56 NPP, which is located in the east coast of Korean peninsula. Through this study, whole tsunami PSA working procedure was established and example calculation was performed for one of real nuclear power plant in Korea

  8. Use of simplified methods for predicting natural resource damages

    International Nuclear Information System (INIS)

    Loreti, C.P.; Boehm, P.D.; Gundlach, E.R.; Healy, E.A.; Rosenstein, A.B.; Tsomides, H.J.; Turton, D.J.; Webber, H.M.

    1995-01-01

    To reduce transaction costs and save time, the US Department of the Interior (DOI) and the National Oceanic and Atmospheric Administration (NOAA) have developed simplified methods for assessing natural resource damages from oil and chemical spills. DOI has proposed the use of two computer models, the Natural Resource Damage Assessment Model for Great Lakes Environments (NRDAM/GLE) and a revised Natural Resource Damage Assessment Model for Coastal and Marine Environments (NRDAM/CME) for predicting monetary damages for spills of oils and chemicals into the Great Lakes and coastal and marine environments. NOAA has used versions of these models to create Compensation Formulas, which it has proposed for calculating natural resource damages for oil spills of up to 50,000 gallons anywhere in the US. Based on a review of the documentation supporting the methods, the results of hundreds of sample runs of DOI's models, and the outputs of the thousands of model runs used to create NOAA's Compensation Formulas, this presentation discusses the ability of these simplified assessment procedures to make realistic damage estimates. The limitations of these procedures are described, and the need for validating the assumptions used in predicting natural resource injuries is discussed

  9. VAN method of short-term earthquake prediction shows promise

    Science.gov (United States)

    Uyeda, Seiya

    Although optimism prevailed in the 1970s, the present consensus on earthquake prediction appears to be quite pessimistic. However, short-term prediction based on geoelectric potential monitoring has stood the test of time in Greece for more than a decade [VarotsosandKulhanek, 1993] Lighthill, 1996]. The method used is called the VAN method.The geoelectric potential changes constantly due to causes such as magnetotelluric effects, lightning, rainfall, leakage from manmade sources, and electrochemical instabilities of electrodes. All of this noise must be eliminated before preseismic signals are identified, if they exist at all. The VAN group apparently accomplished this task for the first time. They installed multiple short (100-200m) dipoles with different lengths in both north-south and east-west directions and long (1-10 km) dipoles in appropriate orientations at their stations (one of their mega-stations, Ioannina, for example, now has 137 dipoles in operation) and found that practically all of the noise could be eliminated by applying a set of criteria to the data.

  10. Predictive ability of machine learning methods for massive crop yield prediction

    Directory of Open Access Journals (Sweden)

    Alberto Gonzalez-Sanchez

    2014-04-01

    Full Text Available An important issue for agricultural planning purposes is the accurate yield estimation for the numerous crops involved in the planning. Machine learning (ML is an essential approach for achieving practical and effective solutions for this problem. Many comparisons of ML methods for yield prediction have been made, seeking for the most accurate technique. Generally, the number of evaluated crops and techniques is too low and does not provide enough information for agricultural planning purposes. This paper compares the predictive accuracy of ML and linear regression techniques for crop yield prediction in ten crop datasets. Multiple linear regression, M5-Prime regression trees, perceptron multilayer neural networks, support vector regression and k-nearest neighbor methods were ranked. Four accuracy metrics were used to validate the models: the root mean square error (RMS, root relative square error (RRSE, normalized mean absolute error (MAE, and correlation factor (R. Real data of an irrigation zone of Mexico were used for building the models. Models were tested with samples of two consecutive years. The results show that M5-Prime and k-nearest neighbor techniques obtain the lowest average RMSE errors (5.14 and 4.91, the lowest RRSE errors (79.46% and 79.78%, the lowest average MAE errors (18.12% and 19.42%, and the highest average correlation factors (0.41 and 0.42. Since M5-Prime achieves the largest number of crop yield models with the lowest errors, it is a very suitable tool for massive crop yield prediction in agricultural planning.

  11. Predicted effectiveness of in-situ activated carbon amendment for field sediment sites with variable site- and compound-specific characteristics

    Energy Technology Data Exchange (ETDEWEB)

    Choi, Yongju, E-mail: ychoi81@snu.ac.kr [Department of Civil and Environmental Engineering, Seoul National University, Seoul 151-744 (Korea, Republic of); Cho, Yeo-Myoung; Luthy, Richard G. [Department of Civil and Environmental Engineering, Stanford University, Stanford, CA 94305-4020 (United States); Werner, David [School of Civil Engineering and Geosciences, Newcastle University, Newcastle upon Tyne NE1 7RU (United Kingdom)

    2016-01-15

    Highlights: • The model accounts for the heterogeneity of AC distribution in field applications. • AC amendment effectiveness is predicted for ten sediment sites. • An HOC mass transfer model and calibrated parameters provide reliable predictions. • AC amendment is predicted to be effective for most sites. • K{sub ow}, K{sub d}, and equilibrium-based calculations are useful indicators. - Abstract: A growing body of evidence shows that the effectiveness of in-situ activated carbon (AC) amendment to treat hydrophobic organic contaminants (HOCs) in sediments can be reliably predicted using a mass transfer modeling approach. This study analyzes available field data for characterizing AC-sediment distribution after mechanical mixing of AC into sediment. Those distributions are used to develop an HOC mass transfer model that accounts for plausible heterogeneities resulting from mixing of AC into sediment. The model is applied to ten field sites in the U.S. and Europe with 2–3 representative HOCs from each site using site- and HOC-specific model parameters collected from the literature. The model predicts that the AC amendment reduces the pore-water HOC concentrations by more than 95% fifteen years after AC deployment for 18 of the 25 total simulated cases when the AC is applied at doses of 1.5 times sediment total organic carbon content with an upper limit of 5 dry wt%. The predicted effectiveness shows negative correlation with the HOC octanol–water partitioning coefficients and the sediment-water distribution coefficients, and positive correlation with the effectiveness calculated based on equilibrium coefficients of sediment and AC, suggesting the possibility for use of the values for screening-level assessments.

  12. Predicted effectiveness of in-situ activated carbon amendment for field sediment sites with variable site- and compound-specific characteristics

    International Nuclear Information System (INIS)

    Choi, Yongju; Cho, Yeo-Myoung; Luthy, Richard G.; Werner, David

    2016-01-01

    Highlights: • The model accounts for the heterogeneity of AC distribution in field applications. • AC amendment effectiveness is predicted for ten sediment sites. • An HOC mass transfer model and calibrated parameters provide reliable predictions. • AC amendment is predicted to be effective for most sites. • K ow , K d , and equilibrium-based calculations are useful indicators. - Abstract: A growing body of evidence shows that the effectiveness of in-situ activated carbon (AC) amendment to treat hydrophobic organic contaminants (HOCs) in sediments can be reliably predicted using a mass transfer modeling approach. This study analyzes available field data for characterizing AC-sediment distribution after mechanical mixing of AC into sediment. Those distributions are used to develop an HOC mass transfer model that accounts for plausible heterogeneities resulting from mixing of AC into sediment. The model is applied to ten field sites in the U.S. and Europe with 2–3 representative HOCs from each site using site- and HOC-specific model parameters collected from the literature. The model predicts that the AC amendment reduces the pore-water HOC concentrations by more than 95% fifteen years after AC deployment for 18 of the 25 total simulated cases when the AC is applied at doses of 1.5 times sediment total organic carbon content with an upper limit of 5 dry wt%. The predicted effectiveness shows negative correlation with the HOC octanol–water partitioning coefficients and the sediment-water distribution coefficients, and positive correlation with the effectiveness calculated based on equilibrium coefficients of sediment and AC, suggesting the possibility for use of the values for screening-level assessments.

  13. A hybrid measure-correlate-predict method for long-term wind condition assessment

    International Nuclear Information System (INIS)

    Zhang, Jie; Chowdhury, Souma; Messac, Achille; Hodge, Bri-Mathias

    2014-01-01

    Highlights: • A hybrid measure-correlate-predict (MCP) methodology with greater accuracy is developed. • Three sets of performance metrics are proposed to evaluate the hybrid MCP method. • Both wind speed and direction are considered in the hybrid MCP method. • The best combination of MCP algorithms is determined. • The developed hybrid MCP method is uniquely helpful for long-term wind resource assessment. - Abstract: This paper develops a hybrid measure-correlate-predict (MCP) strategy to assess long-term wind resource variations at a farm site. The hybrid MCP method uses recorded data from multiple reference stations to estimate long-term wind conditions at a target wind plant site with greater accuracy than is possible with data from a single reference station. The weight of each reference station in the hybrid strategy is determined by the (i) distance and (ii) elevation differences between the target farm site and each reference station. In this case, the wind data is divided into sectors according to the wind direction, and the MCP strategy is implemented for each wind direction sector separately. The applicability of the proposed hybrid strategy is investigated using five MCP methods: (i) the linear regression; (ii) the variance ratio; (iii) the Weibull scale; (iv) the artificial neural networks; and (v) the support vector regression. To implement the hybrid MCP methodology, we use hourly averaged wind data recorded at five stations in the state of Minnesota between 07-01-1996 and 06-30-2004. Three sets of performance metrics are used to evaluate the hybrid MCP method. The first set of metrics analyze the statistical performance, including the mean wind speed, wind speed variance, root mean square error, and mean absolute error. The second set of metrics evaluate the distribution of long-term wind speed; to this end, the Weibull distribution and the Multivariate and Multimodal Wind Distribution models are adopted. The third set of metrics analyze

  14. A Generic Multi-Compartmental CNS Distribution Model Structure for 9 Drugs Allows Prediction of Human Brain Target Site Concentrations

    NARCIS (Netherlands)

    Yamamoto, Yumi; Valitalo, Pyry A.; van den Berg, Dirk-Jan; Hartman, Robin; van den Brink, Willem; Wong, Yin Cheong; Huntjens, Dymphy R.; Proost, Johannes H.; Vermeulen, An; Krauwinkel, Walter; Bakshi, Suruchi; Aranzana-Climent, Vincent; Marchand, Sandrine; Dahyot-Fizelier, Claire; Couet, William; Danhof, Meindert; van Hasselt, Johan G. C.; de lange, Elizabeth C. M.

    Purpose Predicting target site drug concentration in the brain is of key importance for the successful development of drugs acting on the central nervous system. We propose a generic mathematical model to describe the pharmacokinetics in brain compartments, and apply this model to predict human

  15. Prediction of the flooding process at the Ronneburg site - results of an integrated approach

    International Nuclear Information System (INIS)

    Paul, M.; Saenger, H.-J.; Snagowski, S.; Maerten, H.; Eckart, M.

    1998-01-01

    The flooding process of the Ronneburg uranium mine (WISMUT) was initiated at the turn of the year 1997 to 1998. In order to prepare the flooding process and to derive and optimize technological measures an integrated modelling approach was chosen which includes several coupled modules. The most important issues to be answered are: (1) prediction of the flooding time (2) prediction of the groundwater level at the post-flooding stage, assessment of amount, location and quality of flooding waters entering the receiving streams at the final stage (3) water quality prediction within the mine during the flooding process (4) definition of technological measures and assessment of their efficiency A box model which includes the three-dimensional distribution of the cavity volume in the mine represents the model core. The model considers the various types of dewatered cavity volumes for each mine level / mining field and the degree of vertical and horizontal connection between the mining fields. Different types of open mine space as well as the dewatered geological pore and joint volume are considered taking into account the contour of the depression cone prior to flooding and the characteristics of the different rock types. Based on the mine water balance and the flooding technology the model predicts the rise of the water table over time during the flooding process for each mine field separately. In order to predict the mine water quality and the efficiency of in-situ water treatment the box model was linked to a geochemical model (PHREEQC). A three-dimensional flow model is used to evaluate the post-flooding situation at the Ronneburg site. This model is coupled to the box model. The modelling results of various flooding scenarios show that a prediction of the post-flooding geohydraulic situation is possible despite of uncertainties concerning the input parameters which still exist. The post-flooding water table in the central part of the Ronneburg mine will be 270 m

  16. A highly accurate predictive-adaptive method for lithium-ion battery remaining discharge energy prediction in electric vehicle applications

    International Nuclear Information System (INIS)

    Liu, Guangming; Ouyang, Minggao; Lu, Languang; Li, Jianqiu; Hua, Jianfeng

    2015-01-01

    Highlights: • An energy prediction (EP) method is introduced for battery E RDE determination. • EP determines E RDE through coupled prediction of future states, parameters, and output. • The PAEP combines parameter adaptation and prediction to update model parameters. • The PAEP provides improved E RDE accuracy compared with DC and other EP methods. - Abstract: In order to estimate the remaining driving range (RDR) in electric vehicles, the remaining discharge energy (E RDE ) of the applied battery system needs to be precisely predicted. Strongly affected by the load profiles, the available E RDE varies largely in real-world applications and requires specific determination. However, the commonly-used direct calculation (DC) method might result in certain energy prediction errors by relating the E RDE directly to the current state of charge (SOC). To enhance the E RDE accuracy, this paper presents a battery energy prediction (EP) method based on the predictive control theory, in which a coupled prediction of future battery state variation, battery model parameter change, and voltage response, is implemented on the E RDE prediction horizon, and the E RDE is subsequently accumulated and real-timely optimized. Three EP approaches with different model parameter updating routes are introduced, and the predictive-adaptive energy prediction (PAEP) method combining the real-time parameter identification and the future parameter prediction offers the best potential. Based on a large-format lithium-ion battery, the performance of different E RDE calculation methods is compared under various dynamic profiles. Results imply that the EP methods provide much better accuracy than the traditional DC method, and the PAEP could reduce the E RDE error by more than 90% and guarantee the relative energy prediction error under 2%, proving as a proper choice in online E RDE prediction. The correlation of SOC estimation and E RDE calculation is then discussed to illustrate the

  17. Development and Validation of a Preprocedural Risk Score to Predict Access Site Complications After Peripheral Vascular Interventions Based on the Vascular Quality Initiative Database

    Directory of Open Access Journals (Sweden)

    Daniel Ortiz

    2016-01-01

    Full Text Available Purpose: Access site complications following peripheral vascular intervention (PVI are associated with prolonged hospitalization and increased mortality. Prediction of access site complication risk may optimize PVI care; however, there is no tool designed for this. We aimed to create a clinical scoring tool to stratify patients according to their risk of developing access site complications after PVI. Methods: The Society for Vascular Surgery’s Vascular Quality Initiative database yielded 27,997 patients who had undergone PVI at 131 North American centers. Clinically and statistically significant preprocedural risk factors associated with in-hospital, post-PVI access site complications were included in a multivariate logistic regression model, with access site complications as the outcome variable. A predictive model was developed with a random sample of 19,683 (70% PVI procedures and validated in 8,314 (30%. Results: Access site complications occurred in 939 (3.4% patients. The risk tool predictors are female gender, age > 70 years, white race, bedridden ambulatory status, insulin-treated diabetes mellitus, prior minor amputation, procedural indication of claudication, and nonfemoral arterial access site (model c-statistic = 0.638. Of these predictors, insulin-treated diabetes mellitus and prior minor amputation were protective of access site complications. The discriminatory power of the risk model was confirmed by the validation dataset (c-statistic = 0.6139. Higher risk scores correlated with increased frequency of access site complications: 1.9% for low risk, 3.4% for moderate risk and 5.1% for high risk. Conclusions: The proposed clinical risk score based on eight preprocedural characteristics is a tool to stratify patients at risk for post-PVI access site complications. The risk score may assist physicians in identifying patients at risk for access site complications and selection of patients who may benefit from bleeding avoidance

  18. METHOD FOR THE MEASUREMENT OF SITE-SPECIFIC TAUTOMERIC AND ZWITTERIONIC MICROSPECIES EQUILIBRIUM CONSTANTS

    Science.gov (United States)

    We describe a method for the individual measurement of simultaneously occurring, unimolecular, site-specific “microequilibrium” constants as in, for example, prototropic tautomerism and zwitterionic equilibria. Our method represents an elaboration of that of Nygren et al. (Anal. ...

  19. Predicting human height by Victorian and genomic methods.

    Science.gov (United States)

    Aulchenko, Yurii S; Struchalin, Maksim V; Belonogova, Nadezhda M; Axenovich, Tatiana I; Weedon, Michael N; Hofman, Albert; Uitterlinden, Andre G; Kayser, Manfred; Oostra, Ben A; van Duijn, Cornelia M; Janssens, A Cecile J W; Borodin, Pavel M

    2009-08-01

    In the Victorian era, Sir Francis Galton showed that 'when dealing with the transmission of stature from parents to children, the average height of the two parents, ... is all we need care to know about them' (1886). One hundred and twenty-two years after Galton's work was published, 54 loci showing strong statistical evidence for association to human height were described, providing us with potential genomic means of human height prediction. In a population-based study of 5748 people, we find that a 54-loci genomic profile explained 4-6% of the sex- and age-adjusted height variance, and had limited ability to discriminate tall/short people, as characterized by the area under the receiver-operating characteristic curve (AUC). In a family-based study of 550 people, with both parents having height measurements, we find that the Galtonian mid-parental prediction method explained 40% of the sex- and age-adjusted height variance, and showed high discriminative accuracy. We have also explored how much variance a genomic profile should explain to reach certain AUC values. For highly heritable traits such as height, we conclude that in applications in which parental phenotypic information is available (eg, medicine), the Victorian Galton's method will long stay unsurpassed, in terms of both discriminative accuracy and costs. For less heritable traits, and in situations in which parental information is not available (eg, forensics), genomic methods may provide an alternative, given that the variants determining an essential proportion of the trait's variation can be identified.

  20. Methods and approaches to prediction in the meat industry

    Directory of Open Access Journals (Sweden)

    A. B. Lisitsyn

    2016-01-01

    Full Text Available The modern stage of the agro-industrial complex is characterized by an increasing complexity, intensification of technological processes of complex processing of materials of animal origin also the need for a systematic analysis of the variety of determining factors and relationships between them, complexity of the objective function of product quality and severe restrictions on technological regimes. One of the main tasks that face the employees of the enterprises of the agro-industrial complex, which are engaged in processing biotechnological raw materials, is the further organizational improvement of work at all stages of the food chain, besides an increase in the production volume. The meat industry as a part of the agro-industrial complex has to use the biological raw materials with maximum efficiency, while reducing and even eliminating losses at all stages of processing; rationally use raw material when selecting a type of processing products; steadily increase quality, biological and food value of products; broaden the assortment of manufactured products in order to satisfy increasing consumer requirements and extend the market for their realization in the conditions of uncertainty of external environment, due to the uneven receipt of raw materials, variations in its properties and parameters, limited time sales and fluctuations in demand for products. The challenges facing the meat industry cannot be solved without changes to the strategy for scientific and technological development of the industry. To achieve these tasks, it is necessary to use the prediction as a method of constant improvement of all technological processes and their performance under the rational and optimal regimes, while constantly controlling quality of raw material, semi-prepared products and finished products at all stages of the technological processing by the physico-chemical, physico-mechanical (rheological, microbiological and organoleptic methods. The paper

  1. Prediction of DtxR regulon: Identification of binding sites and operons controlled by Diphtheria toxin repressor in Corynebacterium diphtheriae

    Directory of Open Access Journals (Sweden)

    Hasnain Seyed

    2004-09-01

    Full Text Available Abstract Background The diphtheria toxin repressor, DtxR, of Corynebacterium diphtheriae has been shown to be an iron-activated transcription regulator that controls not only the expression of diphtheria toxin but also of iron uptake genes. This study aims to identify putative binding sites and operons controlled by DtxR to understand the role of DtxR in patho-physiology of Corynebacterium diphtheriae. Result Positional Shannon relative entropy method was used to build the DtxR-binding site recognition profile and the later was used to identify putative regulatory sites of DtxR within C. diphtheriae genome. In addition, DtxR-regulated operons were also identified taking into account the predicted DtxR regulatory sites and genome annotation. Few of the predicted motifs were experimentally validated by electrophoretic mobility shift assay. The analysis identifies motifs upstream to the novel iron-regulated genes that code for Formamidopyrimidine-DNA glycosylase (FpG, an enzyme involved in DNA-repair and starvation inducible DNA-binding protein (Dps which is involved in iron storage and oxidative stress defense. In addition, we have found the DtxR motifs upstream to the genes that code for sortase which catalyzes anchoring of host-interacting proteins to the cell wall of pathogenic bacteria and the proteins of secretory system which could be involved in translocation of various iron-regulated virulence factors including diphtheria toxin. Conclusions We have used an in silico approach to identify the putative binding sites and genes controlled by DtxR in Corynebacterium diphtheriae. Our analysis shows that DtxR could provide a molecular link between Fe+2-induced Fenton's reaction and protection of DNA from oxidative damage. DtxR-regulated Dps prevents lethal combination of Fe+2 and H2O2 and also protects DNA by nonspecific DNA-binding. In addition DtxR could play an important role in host interaction and virulence by regulating the levels of sortase

  2. Cell-type specificity of ChIP-predicted transcription factor binding sites

    Directory of Open Access Journals (Sweden)

    Håndstad Tony

    2012-08-01

    Full Text Available Abstract Background Context-dependent transcription factor (TF binding is one reason for differences in gene expression patterns between different cellular states. Chromatin immunoprecipitation followed by high-throughput sequencing (ChIP-seq identifies genome-wide TF binding sites for one particular context—the cells used in the experiment. But can such ChIP-seq data predict TF binding in other cellular contexts and is it possible to distinguish context-dependent from ubiquitous TF binding? Results We compared ChIP-seq data on TF binding for multiple TFs in two different cell types and found that on average only a third of ChIP-seq peak regions are common to both cell types. Expectedly, common peaks occur more frequently in certain genomic contexts, such as CpG-rich promoters, whereas chromatin differences characterize cell-type specific TF binding. We also find, however, that genotype differences between the cell types can explain differences in binding. Moreover, ChIP-seq signal intensity and peak clustering are the strongest predictors of common peaks. Compared with strong peaks located in regions containing peaks for multiple transcription factors, weak and isolated peaks are less common between the cell types and are less associated with data that indicate regulatory activity. Conclusions Together, the results suggest that experimental noise is prevalent among weak peaks, whereas strong and clustered peaks represent high-confidence binding events that often occur in other cellular contexts. Nevertheless, 30-40% of the strongest and most clustered peaks show context-dependent regulation. We show that by combining signal intensity with additional data—ranging from context independent information such as binding site conservation and position weight matrix scores to context dependent chromatin structure—we can predict whether a ChIP-seq peak is likely to be present in other cellular contexts.

  3. FREEZING AND THAWING TIME PREDICTION METHODS OF FOODS II: NUMARICAL METHODS

    Directory of Open Access Journals (Sweden)

    Yahya TÜLEK

    1999-03-01

    Full Text Available Freezing is one of the excellent methods for the preservation of foods. If freezing and thawing processes and frozen storage method are carried out correctly, the original characteristics of the foods can remain almost unchanged over an extended periods of time. It is very important to determine the freezing and thawing time period of the foods, as they strongly influence the both quality of food material and process productivity and the economy. For developing a simple and effectively usable mathematical model, less amount of process parameters and physical properties should be enrolled in calculations. But it is a difficult to have all of these in one prediction method. For this reason, various freezing and thawing time prediction methods were proposed in literature and research studies have been going on.

  4. Climate change research methods and its significance in the study of choosing candidate site in nuclear waste disposal sites

    International Nuclear Information System (INIS)

    Zhao Yong; Zhang Zhanshi

    2008-01-01

    A high-level nuclear waste is the inevitable product accompanies the development of the nuclear power station. How to dispose it properly has become focused by all over the world. Some of the important progresses have been achieved in the fields of site setting, performance assessment and underground laboratory recently. Palaeoclimate patter and the tendency of climate change are very important aspects for the site setting This paper discussed some of the important progresses on the disposal of unclear wastes, the influence of the climate change on the site setting and main methods such as lake sediments, marine sediment, loess, ancient soil and ice core deal with palaeoclimate and palaeo environment study. (authors)

  5. Prediction of evacuation time for emergency planning zone of Uljin nuclear site

    International Nuclear Information System (INIS)

    Jeon, In Young; Lee, Jai Ki

    2002-01-01

    The time for evacuation of residents in Emergency Planning Zone (EPZ) of Uljin nuclear site in case of a radiological emergency was estimated with traffic analysis. Evacuees were classified into 4 groups by considering population density, local jurisdictions, and whether they are residents or transients. The survey to investigate the behavioral characteristics of the residents was made for 200 households and included a hypothetical scenario explaining the accident situation and questions such as dwelling place, time demand for evacuation preparation, transportation means for evacuation, sheltering place, and evacuation direction. The microscopic traffic simulation model, CORSIM, was used to simulate the behavior of evacuating vehicles on networks. The results showed that the evacuation time required for total vehicles to move out from EPZ took longer in the daytime than at night in spite that the delay times at intersections were longer at night than in the daytime. This was analyzed due to the differences of the trip generation time distribution. To validate whether the CORSIM model can appropriately simulate the congested traffic phenomena assumable in case of emergency, a benchmark study was conducted at an intersection without an actuated traffic signal near Uljin site during the traffic peak-time in the morning. This study indicated that the predicted output by the CORSIM model was in good agreement with the observed data, satisfying the purpose of this study

  6. Combined geophysical methods for mapping infiltration pathways at the Aurora Water Aquifer recharge and recovery site

    Science.gov (United States)

    Jasper, Cameron A.

    Although aquifer recharge and recovery systems are a sustainable, decentralized, low cost, and low energy approach for the reclamation, treatment, and storage of post- treatment wastewater, they can suffer from poor infiltration rates and the development of a near-surface clogging layer within infiltration ponds. One such aquifer recharge and recovery system, the Aurora Water site in Colorado, U.S.A, functions at about 25% of its predicted capacity to recharge floodplain deposits by flooding infiltration ponds with post-treatment wastewater extracted from river bank aquifers along the South Platte River. The underwater self-potential method was developed to survey self-potential signals at the ground surface in a flooded infiltration pond for mapping infiltration pathways. A method for using heat as a groundwater tracer within the infiltration pond used an array of in situ high-resolution temperature sensing probes. Both relatively positive and negative underwater self-potential anomalies are consistent with observed recovery well pumping rates and specific discharge estimates from temperature data. Results from electrical resistivity tomography and electromagnetics surveys provide consistent electrical conductivity distributions associated with sediment textures. A lab method was developed for resistivity tests of near-surface sediment samples. Forward numerical modeling synthesizes the geophysical information to best match observed self- potential anomalies and provide permeability distributions, which is important for effective aquifer recharge and recovery system design, and optimization strategy development.

  7. Evaluation of mathematical methods for predicting optimum dose of gamma radiation in sugarcane (Saccharum sp.)

    International Nuclear Information System (INIS)

    Wu, K.K.; Siddiqui, S.H.; Heinz, D.J.; Ladd, S.L.

    1978-01-01

    Two mathematical methods - the reversed logarithmic method and the regression method - were used to compare the predicted and the observed optimum gamma radiation dose (OD 50 ) in vegetative propagules of sugarcane. The reversed logarithmic method, usually used in sexually propagated crops, showed the largest difference between the predicted and observed optimum dose. The regression method resulted in a better prediction of the observed values and is suggested as a better method for the prediction of optimum dose for vegetatively propagated crops. (author)

  8. Use of the Hazard Prediction and Assessment Capability (HPAC) at the Savannah River Site

    International Nuclear Information System (INIS)

    BUCKLEY, ROBERT

    2004-01-01

    This report provides background information on the implementation of the Hazard Prediction and Assessment Capability (HPAC) software package at the Savannah River Site for atmospheric modeling. Developed by the Defense Threat Reduction Agency (DTRA), HPAC is actually a suite of models that allows for various modes of release of radiological, chemical and biological agents, generates interpolated meteorological data fields based on inputted meteorology, and transports the material using a tested transport and diffusion model. A discussion of meteorological data availability for direct use by HPAC is given, and is important for applications to emergency response. An evaluation of the data input methodology for a release from SRS is examined, as well as an application to the World Trade Center attacks in New York City in September 2001. Benefits of the newer versions of HPAC now in use are then discussed. These include access to more meteorological data sources and improved graphical cap abilities. The HPAC software package is another tool to help the Atmospheric Technologies Group (ATG) provide atmospheric transport and dispersion predictions in the event of hazardous atmospheric releases

  9. Integrating intrusive and nonintrusive characterization methods to achieve a conceptual site model for the SLDA FUSRAP site

    International Nuclear Information System (INIS)

    Durham, L.A.; Peterson, J.M.; Frothingham, D.G.; Frederick, W.T.; Lenart, W.

    2008-01-01

    The US Army Corps of Engineers (USACE) is addressing radiological contamination following Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA) requirements at the Shallow Land Disposal Area (SLDA) site, which is a radiologically contaminated property that is part of the Formerly utilized Sites Remedial Action Program (FUSRAP). The SLDA is an 18-hectare (44-acre) site in Parks township, Armstrong County, Pennsylvania, about 37 kilometers (23 miles) east-northeast of Pittsburgh. According to historical record, radioactive wastes were disposed of at the SLDA in a series of trenches by the Nuclear Materials and Equipment Company (NUMEC) in the 1960s. The wastes originated from the nearby Apollo nuclear fuel fabrication facility, which began operations under NUMEC in the late 1950s and fabricated enriched uranium into naval reactor fuel elements. It is believed that the waste materials were buried in a series of pits constructed adjacent to one another in accordance with an Atomic Energy Commission (AEC) regulation that has since been rescinded. A CERCLA remedial investigation/feasibility study (RI/FS) process was completed for the SLDA site, and the results of the human health risk assessment indicated that the radiologically contaminated wastes could pose a risk to human health in the future. There are no historical records that provide the exact location of these pits. However, based on geophysical survey results conducted in the 1980s, these pits were defined by geophysical anomalies and were depicted on historical site drawings as trenches. At the SLDA site, a combination of investigative methods and tools was used in the RI/FS and site characterization activities. The SLDA site provides an excellent example of how historical documents and data, historical aerial photo analysis, physical sampling, and nonintrusive geophysical and gamma walkover surveys were used in combination to reduce the uncertainty in the location of the trenches. The

  10. PREDICTION OF MEAT PRODUCT QUALITY BY THE MATHEMATICAL PROGRAMMING METHODS

    Directory of Open Access Journals (Sweden)

    A. B. Lisitsyn

    2016-01-01

    Full Text Available Abstract Use of the prediction technologies is one of the directions of the research work carried out both in Russia and abroad. Meat processing is accompanied by the complex physico-chemical, biochemical and mechanical processes. To predict the behavior of meat raw material during the technological processing, a complex of physico-technological and structural-mechanical indicators, which objectively reflects its quality, is used. Among these indicators are pH value, water binding and fat holding capacities, water activity, adhesiveness, viscosity, plasticity and so on. The paper demonstrates the influence of animal proteins (beef and pork on the physico-chemical and functional properties before and after thermal treatment of minced meat made from meat raw material with different content of the connective and fat tissues. On the basis of the experimental data, the model (stochastic dependence parameters linking the quantitative resultant and factor variables were obtained using the regression analysis, and the degree of the correlation with the experimental data was assessed. The maximum allowable levels of meat raw material replacement with animal proteins (beef and pork were established by the methods of mathematical programming. Use of the information technologies will significantly reduce the costs of the experimental search and substantiation of the optimal level of replacement of meat raw material with animal proteins (beef, pork, and will also allow establishing a relationship of product quality indicators with quantity and quality of minced meat ingredients.

  11. A comparison of different methods for predicting coal devolatilisation kinetics

    Energy Technology Data Exchange (ETDEWEB)

    Arenillas, A.; Rubiera, F.; Pevida, C.; Pis, J.J. [Instituto Nacional del Carbon, CSIC, Apartado 73, 33080 Oviedo (Spain)

    2001-04-01

    Knowledge of the coal devolatilisation rate is of great importance because it exerts a marked effect on the overall combustion behaviour. Different approaches can be used to obtain the kinetics of the complex devolatilisation process. The simplest are empirical and employ global kinetics, where the Arrhenius expression is used to correlate rates of mass loss with temperature. In this study a high volatile bituminous coal was devolatilised at four different heating rates in a thermogravimetric analyser (TG) linked to a mass spectrometer (MS). As a first approach, the Arrhenius kinetic parameters (k and A) were calculated from the experimental results, assuming a single step process. Another approach is the distributed-activation energy model, which is more complex due to the assumption that devolatilisation occurs through several first-order reactions, which occur simultaneously. Recent advances in the understanding of coal structure have led to more fundamental approaches for modelling devolatilisation behaviour, such as network models. These are based on a physico-chemical description of coal structure. In the present study the FG-DVC (Functional Group-Depolymerisation, Vaporisation and Crosslinking) computer code was used as the network model and the FG-DVC predicted evolution of volatile compounds was compared with the experimental results. In addition, the predicted rate of mass loss from the FG-DVC model was used to obtain a third devolatilisation kinetic approach. The three methods were compared and discussed, with the experimental results as a reference.

  12. Predicting lattice thermal conductivity with help from ab initio methods

    Science.gov (United States)

    Broido, David

    2015-03-01

    The lattice thermal conductivity is a fundamental transport parameter that determines the utility a material for specific thermal management applications. Materials with low thermal conductivity find applicability in thermoelectric cooling and energy harvesting. High thermal conductivity materials are urgently needed to help address the ever-growing heat dissipation problem in microelectronic devices. Predictive computational approaches can provide critical guidance in the search and development of new materials for such applications. Ab initio methods for calculating lattice thermal conductivity have demonstrated predictive capability, but while they are becoming increasingly efficient, they are still computationally expensive particularly for complex crystals with large unit cells . In this talk, I will review our work on first principles phonon transport for which the intrinsic lattice thermal conductivity is limited only by phonon-phonon scattering arising from anharmonicity. I will examine use of the phase space for anharmonic phonon scattering and the Grüneisen parameters as measures of the thermal conductivities for a range of materials and compare these to the widely used guidelines stemming from the theory of Liebfried and Schölmann. This research was supported primarily by the NSF under Grant CBET-1402949, and by the S3TEC, an Energy Frontier Research Center funded by the US DOE, office of Basic Energy Sciences under Award No. DE-SC0001299.

  13. Development of nondestructive method for prediction of crack instability

    International Nuclear Information System (INIS)

    Schroeder, J.L.; Eylon, D.; Shell, E.B.; Matikas, T.E.

    2000-01-01

    A method to characterize the deformation zone at a crack tip and predict upcoming fracture under load using white light interference microscopy was developed and studied. Cracks were initiated in notched Ti-6Al-4V specimens through fatigue loading. Following crack initiation, specimens were subjected to static loading during in-situ observation of the deformation area ahead of the crack. Nondestructive in-situ observations were performed using white light interference microscopy. Profilometer measurements quantified the area, volume, and shape of the deformation ahead of the crack front. Results showed an exponential relationship between the area and volume of deformation and the stress intensity factor of the cracked alloy. These findings also indicate that it is possible to determine a critical rate of change in deformation versus the stress intensity factor that can predict oncoming catastrophic failure. In addition, crack front deformation zones were measured as a function of time under sustained load, and crack tip deformation zone enlargement over time was observed

  14. Extremely Randomized Machine Learning Methods for Compound Activity Prediction

    Directory of Open Access Journals (Sweden)

    Wojciech M. Czarnecki

    2015-11-01

    Full Text Available Speed, a relatively low requirement for computational resources and high effectiveness of the evaluation of the bioactivity of compounds have caused a rapid growth of interest in the application of machine learning methods to virtual screening tasks. However, due to the growth of the amount of data also in cheminformatics and related fields, the aim of research has shifted not only towards the development of algorithms of high predictive power but also towards the simplification of previously existing methods to obtain results more quickly. In the study, we tested two approaches belonging to the group of so-called ‘extremely randomized methods’—Extreme Entropy Machine and Extremely Randomized Trees—for their ability to properly identify compounds that have activity towards particular protein targets. These methods were compared with their ‘non-extreme’ competitors, i.e., Support Vector Machine and Random Forest. The extreme approaches were not only found out to improve the efficiency of the classification of bioactive compounds, but they were also proved to be less computationally complex, requiring fewer steps to perform an optimization procedure.

  15. Predicting Greater Prairie-Chicken Lek Site Suitability to Inform Conservation Actions.

    Directory of Open Access Journals (Sweden)

    Torre J Hovick

    Full Text Available The demands of a growing human population dictates that expansion of energy infrastructure, roads, and other development frequently takes place in native rangelands. Particularly, transmission lines and roads commonly divide rural landscapes and increase fragmentation. This has direct and indirect consequences on native wildlife that can be mitigated through thoughtful planning and proactive approaches to identifying areas of high conservation priority. We used nine years (2003-2011 of Greater Prairie-Chicken (Tympanuchus cupido lek locations totaling 870 unique leks sites in Kansas and seven geographic information system (GIS layers describing land cover, topography, and anthropogenic structures to model habitat suitability across the state. The models obtained had low omission rates (0.81, indicating high model performance and reliability of predicted habitat suitability for Greater Prairie-Chickens. We found that elevation was the most influential in predicting lek locations, contributing three times more predictive power than any other variable. However, models were improved by the addition of land cover and anthropogenic features (transmission lines, roads, and oil and gas structures. Overall, our analysis provides a hierarchal understanding of Greater Prairie-Chicken habitat suitability that is broadly based on geomorphological features followed by land cover suitability. We found that when land features and vegetation cover are suitable for Greater Prairie-Chickens, fragmentation by anthropogenic sources such as roadways and transmission lines are a concern. Therefore, it is our recommendation that future human development in Kansas avoid areas that our models identified as highly suitable for Greater Prairie-Chickens and focus development on land cover types that are of lower conservation concern.

  16. RANDOM FUNCTIONS AND INTERVAL METHOD FOR PREDICTING THE RESIDUAL RESOURCE OF BUILDING STRUCTURES

    Directory of Open Access Journals (Sweden)

    Shmelev Gennadiy Dmitrievich

    2017-11-01

    Full Text Available Subject: possibility of using random functions and interval prediction method for estimating the residual life of building structures in the currently used buildings. Research objectives: coordination of ranges of values to develop predictions and random functions that characterize the processes being predicted. Materials and methods: when performing this research, the method of random functions and the method of interval prediction were used. Results: in the course of this work, the basic properties of random functions, including the properties of families of random functions, are studied. The coordination of time-varying impacts and loads on building structures is considered from the viewpoint of their influence on structures and representation of the structures’ behavior in the form of random functions. Several models of random functions are proposed for predicting individual parameters of structures. For each of the proposed models, its scope of application is defined. The article notes that the considered approach of forecasting has been used many times at various sites. In addition, the available results allowed the authors to develop a methodology for assessing the technical condition and residual life of building structures for the currently used facilities. Conclusions: we studied the possibility of using random functions and processes for the purposes of forecasting the residual service lives of structures in buildings and engineering constructions. We considered the possibility of using an interval forecasting approach to estimate changes in defining parameters of building structures and their technical condition. A comprehensive technique for forecasting the residual life of building structures using the interval approach is proposed.

  17. Prediction of Nepsilon-acetylation on internal lysines implemented in Bayesian Discriminant Method.

    Science.gov (United States)

    Li, Ao; Xue, Yu; Jin, Changjiang; Wang, Minghui; Yao, Xuebiao

    2006-12-01

    Protein acetylation is an important and reversible post-translational modification (PTM), and it governs a variety of cellular dynamics and plasticity. Experimental identification of acetylation sites is labor-intensive and often limited by the availability of reagents such as acetyl-specific antibodies and optimization of enzymatic reactions. Computational analyses may facilitate the identification of potential acetylation sites and provide insights into further experimentation. In this manuscript, we present a novel protein acetylation prediction program named PAIL, prediction of acetylation on internal lysines, implemented in a BDM (Bayesian Discriminant Method) algorithm. The accuracies of PAIL are 85.13%, 87.97%, and 89.21% at low, medium, and high thresholds, respectively. Both Jack-Knife validation and n-fold cross-validation have been performed to show that PAIL is accurate and robust. Taken together, we propose that PAIL is a novel predictor for identification of protein acetylation sites and may serve as an important tool to study the function of protein acetylation. PAIL has been implemented in PHP and is freely available on a web server at: http://bioinformatics.lcd-ustc.org/pail.

  18. Prediction of Nε-acetylation on internal lysines implemented in Bayesian Discriminant Method

    Science.gov (United States)

    Li, Ao; Xue, Yu; Jin, Changjiang; Wang, Minghui; Yao, Xuebiao

    2007-01-01

    Protein acetylation is an important and reversible post-translational modification (PTM), and it governs a variety of cellular dynamics and plasticity. Experimental identification of acetylation sites is labor-intensive and often limited by the availability reagents such as acetyl-specific antibodies and optimization of enzymatic reactions. Computational analyses may facilitate the identification of potential acetylation sites and provide insights into further experimentation. In this manuscript, we present a novel protein acetylation prediction program named PAIL, prediction of acetylation on internal lysines, implemented in a BDM (Bayesian Discriminant Method) algorithm. The accuracies of PAIL are 85.13%, 87.97% and 89.21% at low, medium and high thresholds, respectively. Both Jack-Knife validation and n-fold cross validation have been performed to show that PAIL is accurate and robust. Taken together, we propose that PAIL is a novel predictor for identification of protein acetylation sites and may serve as an important tool to study the function of protein acetylation. PAIL has been implemented in PHP and is freely available on a web server at: http://bioinformatics.lcd-ustc.org/pail. PMID:17045240

  19. Predicting long-term moisture contents of earthen covers at uranium mill tailings sites

    International Nuclear Information System (INIS)

    Gee, G.W.; Nielson, K.K.; Rogers, V.C.

    1984-09-01

    The three methods for long-term moisture prediction covered in this report are: estimates from water retention (permanent wilting point) data, correlation with climate and soil type, and detailed model simulation. The test results have shown: soils vary greatly in residual moisture. Expected long-term moisture saturation ratios (based on generalized soil characteristics) range from 0.2 to 0.8 for soils ranging in texture from sand to clay, respectively. These values hold for noncompacted field soils. Measured radon diffusion coefficients for soils at 15-bar water contents ranged from 5.0E-2 cm 2 /s to 5.0E-3 cm 2 /s for sands and clays, respectively, at typical field densities. In contrast, fine-textured pit-run earthen materials, subjected to optimum compaction (>85% Proctor density) and dried to the 15-bar water content, ranged from 0.7 to 0.9 moisture saturation. Compacted pit-run soils at these moisture contents exhibited radon diffusion coefficients as low as 3.0E-4 cm 2 /s. The residual moisture saturation for cover soils is not known since no engineered barrier has been in place for more than a few years. A comparison of methods for predicting moisture saturation indicates that model simulations are useful for predicting effects of climatic changes on residual soil moisture, but that long-term moisture also can be predicted with some degree of confidence using generalized soil properties or empirical correlations based both on soils and climatic information. The optimal soil cover design will likely include more than one layer of soil. A two-layer system using a thick (1-m minimum) plant root zone of uncompacted soil placed over a moistened, tightly compacted fine-textured soil is recommended. This design concept has been tested successfully at the Grand Junction, Colorado, tailings piles

  20. A predictive estimation method for carbon dioxide transport by data-driven modeling with a physically-based data model

    Science.gov (United States)

    Jeong, Jina; Park, Eungyu; Han, Weon Shik; Kim, Kue-Young; Jun, Seong-Chun; Choung, Sungwook; Yun, Seong-Taek; Oh, Junho; Kim, Hyun-Jun

    2017-11-01

    In this study, a data-driven method for predicting CO2 leaks and associated concentrations from geological CO2 sequestration is developed. Several candidate models are compared based on their reproducibility and predictive capability for CO2 concentration measurements from the Environment Impact Evaluation Test (EIT) site in Korea. Based on the data mining results, a one-dimensional solution of the advective-dispersive equation for steady flow (i.e., Ogata-Banks solution) is found to be most representative for the test data, and this model is adopted as the data model for the developed method. In the validation step, the method is applied to estimate future CO2 concentrations with the reference estimation by the Ogata-Banks solution, where a part of earlier data is used as the training dataset. From the analysis, it is found that the ensemble mean of multiple estimations based on the developed method shows high prediction accuracy relative to the reference estimation. In addition, the majority of the data to be predicted are included in the proposed quantile interval, which suggests adequate representation of the uncertainty by the developed method. Therefore, the incorporation of a reasonable physically-based data model enhances the prediction capability of the data-driven model. The proposed method is not confined to estimations of CO2 concentration and may be applied to various real-time monitoring data from subsurface sites to develop automated control, management or decision-making systems.

  1. Site-specific strong ground motion prediction using 2.5-D modelling

    Science.gov (United States)

    Narayan, J. P.

    2001-08-01

    An algorithm was developed using the 2.5-D elastodynamic wave equation, based on the displacement-stress relation. One of the most significant advantages of the 2.5-D simulation is that the 3-D radiation pattern can be generated using double-couple point shear-dislocation sources in the 2-D numerical grid. A parsimonious staggered grid scheme was adopted instead of the standard staggered grid scheme, since this is the only scheme suitable for computing the dislocation. This new 2.5-D numerical modelling avoids the extensive computational cost of 3-D modelling. The significance of this exercise is that it makes it possible to simulate the strong ground motion (SGM), taking into account the energy released, 3-D radiation pattern, path effects and local site conditions at any location around the epicentre. The slowness vector (py) was used in the supersonic region for each layer, so that all the components of the inertia coefficient are positive. The double-couple point shear-dislocation source was implemented in the numerical grid using the moment tensor components as the body-force couples. The moment per unit volume was used in both the 3-D and 2.5-D modelling. A good agreement in the 3-D and 2.5-D responses for different grid sizes was obtained when the moment per unit volume was further reduced by a factor equal to the finite-difference grid size in the case of the 2.5-D modelling. The components of the radiation pattern were computed in the xz-plane using 3-D and 2.5-D algorithms for various focal mechanisms, and the results were in good agreement. A comparative study of the amplitude behaviour of the 3-D and 2.5-D wavefronts in a layered medium reveals the spatial and temporal damped nature of the 2.5-D elastodynamic wave equation. 3-D and 2.5-D simulated responses at a site using a different strike direction reveal that strong ground motion (SGM) can be predicted just by rotating the strike of the fault counter-clockwise by the same amount as the azimuth of

  2. Prediction of protein modification sites of pyrrolidone carboxylic acid using mRMR feature selection and analysis.

    Directory of Open Access Journals (Sweden)

    Lu-Lu Zheng

    Full Text Available Pyrrolidone carboxylic acid (PCA is formed during a common post-translational modification (PTM of extracellular and multi-pass membrane proteins. In this study, we developed a new predictor to predict the modification sites of PCA based on maximum relevance minimum redundancy (mRMR and incremental feature selection (IFS. We incorporated 727 features that belonged to 7 kinds of protein properties to predict the modification sites, including sequence conservation, residual disorder, amino acid factor, secondary structure and solvent accessibility, gain/loss of amino acid during evolution, propensity of amino acid to be conserved at protein-protein interface and protein surface, and deviation of side chain carbon atom number. Among these 727 features, 244 features were selected by mRMR and IFS as the optimized features for the prediction, with which the prediction model achieved a maximum of MCC of 0.7812. Feature analysis showed that all feature types contributed to the modification process. Further site-specific feature analysis showed that the features derived from PCA's surrounding sites contributed more to the determination of PCA sites than other sites. The detailed feature analysis in this paper might provide important clues for understanding the mechanism of the PCA formation and guide relevant experimental validations.

  3. An integrative computational framework based on a two-step random forest algorithm improves prediction of zinc-binding sites in proteins.

    Directory of Open Access Journals (Sweden)

    Cheng Zheng

    Full Text Available Zinc-binding proteins are the most abundant metalloproteins in the Protein Data Bank where the zinc ions usually have catalytic, regulatory or structural roles critical for the function of the protein. Accurate prediction of zinc-binding sites is not only useful for the inference of protein function but also important for the prediction of 3D structure. Here, we present a new integrative framework that combines multiple sequence and structural properties and graph-theoretic network features, followed by an efficient feature selection to improve prediction of zinc-binding sites. We investigate what information can be retrieved from the sequence, structure and network levels that is relevant to zinc-binding site prediction. We perform a two-step feature selection using random forest to remove redundant features and quantify the relative importance of the retrieved features. Benchmarking on a high-quality structural dataset containing 1,103 protein chains and 484 zinc-binding residues, our method achieved >80% recall at a precision of 75% for the zinc-binding residues Cys, His, Glu and Asp on 5-fold cross-validation tests, which is a 10%-28% higher recall at the 75% equal precision compared to SitePredict and zincfinder at residue level using the same dataset. The independent test also indicates that our method has achieved recall of 0.790 and 0.759 at residue and protein levels, respectively, which is a performance better than the other two methods. Moreover, AUC (the Area Under the Curve and AURPC (the Area Under the Recall-Precision Curve by our method are also respectively better than those of the other two methods. Our method can not only be applied to large-scale identification of zinc-binding sites when structural information of the target is available, but also give valuable insights into important features arising from different levels that collectively characterize the zinc-binding sites. The scripts and datasets are available at http://protein.cau.edu.cn/zincidentifier/.

  4. Assessment method to predict the rate of unresolved false alarms

    International Nuclear Information System (INIS)

    Reardon, P.T.; Eggers, R.F.; Heaberlin, S.W.

    1982-06-01

    A method has been developed to predict the rate of unresolved false alarms of material loss in a nuclear facility. The computer program DETRES-1 was developed. The program first assigns the true values of control unit components receipts, shipments, beginning and ending inventories. A normal random number generator is used to generate measured values of each component. A loss estimator is calculated from the control unit's measured values. If the loss estimator triggers a detection alarm, a response is simulated. The response simulation is divided into two phases. The first phase is to simulate remeasurement of the components of the detection loss estimator using the same or better measurement methods or inferences from surrounding control units. If this phase of response continues to indicate a material loss, phase of response simulating a production shutdown and comprehensive cleanout is initiated. A new loss estimator is found, and tested against the alarm thresholds. If the estimator value is below the threshold, the original detection alarm is considered resolved; if above the threshold, an unresolved alarm has occurred. A tally is kept of valid alarms, unresolved false alarms, and failure to alarm upon a true loss

  5. Predicted impacts of future water level decline on monitoring wells using a ground-water model of the Hanford Site

    International Nuclear Information System (INIS)

    Wurstner, S.K.; Freshley, M.D.

    1994-12-01

    A ground-water flow model was used to predict water level decline in selected wells in the operating areas (100, 200, 300, and 400 Areas) and the 600 Area. To predict future water levels, the unconfined aquifer system was stimulated with the two-dimensional version of a ground-water model of the Hanford Site, which is based on the Coupled Fluid, Energy, and Solute Transport (CFEST) Code in conjunction with the Geographic Information Systems (GIS) software package. The model was developed using the assumption that artificial recharge to the unconfined aquifer system from Site operations was much greater than any natural recharge from precipitation or from the basalt aquifers below. However, artificial recharge is presently decreasing and projected to decrease even more in the future. Wells currently used for monitoring at the Hanford Site are beginning to go dry or are difficult to sample, and as the water table declines over the next 5 to 10 years, a larger number of wells is expected to be impacted. The water levels predicted by the ground-water model were compared with monitoring well completion intervals to determine which wells will become dry in the future. Predictions of wells that will go dry within the next 5 years have less uncertainty than predictions for wells that will become dry within 5 to 10 years. Each prediction is an estimate based on assumed future Hanford Site operating conditions and model assumptions

  6. A novel time series link prediction method: Learning automata approach

    Science.gov (United States)

    Moradabadi, Behnaz; Meybodi, Mohammad Reza

    2017-09-01

    Link prediction is a main social network challenge that uses the network structure to predict future links. The common link prediction approaches to predict hidden links use a static graph representation where a snapshot of the network is analyzed to find hidden or future links. For example, similarity metric based link predictions are a common traditional approach that calculates the similarity metric for each non-connected link and sort the links based on their similarity metrics and label the links with higher similarity scores as the future links. Because people activities in social networks are dynamic and uncertainty, and the structure of the networks changes over time, using deterministic graphs for modeling and analysis of the social network may not be appropriate. In the time-series link prediction problem, the time series link occurrences are used to predict the future links In this paper, we propose a new time series link prediction based on learning automata. In the proposed algorithm for each link that must be predicted there is one learning automaton and each learning automaton tries to predict the existence or non-existence of the corresponding link. To predict the link occurrence in time T, there is a chain consists of stages 1 through T - 1 and the learning automaton passes from these stages to learn the existence or non-existence of the corresponding link. Our preliminary link prediction experiments with co-authorship and email networks have provided satisfactory results when time series link occurrences are considered.

  7. Protein-protein interaction site predictions with three-dimensional probability distributions of interacting atoms on protein surfaces.

    Directory of Open Access Journals (Sweden)

    Ching-Tai Chen

    Full Text Available Protein-protein interactions are key to many biological processes. Computational methodologies devised to predict protein-protein interaction (PPI sites on protein surfaces are important tools in providing insights into the biological functions of proteins and in developing therapeutics targeting the protein-protein interaction sites. One of the general features of PPI sites is that the core regions from the two interacting protein surfaces are complementary to each other, similar to the interior of proteins in packing density and in the physicochemical nature of the amino acid composition. In this work, we simulated the physicochemical complementarities by constructing three-dimensional probability density maps of non-covalent interacting atoms on the protein surfaces. The interacting probabilities were derived from the interior of known structures. Machine learning algorithms were applied to learn the characteristic patterns of the probability density maps specific to the PPI sites. The trained predictors for PPI sites were cross-validated with the training cases (consisting of 432 proteins and were tested on an independent dataset (consisting of 142 proteins. The residue-based Matthews correlation coefficient for the independent test set was 0.423; the accuracy, precision, sensitivity, specificity were 0.753, 0.519, 0.677, and 0.779 respectively. The benchmark results indicate that the optimized machine learning models are among the best predictors in identifying PPI sites on protein surfaces. In particular, the PPI site prediction accuracy increases with increasing size of the PPI site and with increasing hydrophobicity in amino acid composition of the PPI interface; the core interface regions are more likely to be recognized with high prediction confidence. The results indicate that the physicochemical complementarity patterns on protein surfaces are important determinants in PPIs, and a substantial portion of the PPI sites can be predicted

  8. Protein-Protein Interaction Site Predictions with Three-Dimensional Probability Distributions of Interacting Atoms on Protein Surfaces

    Science.gov (United States)

    Chen, Ching-Tai; Peng, Hung-Pin; Jian, Jhih-Wei; Tsai, Keng-Chang; Chang, Jeng-Yih; Yang, Ei-Wen; Chen, Jun-Bo; Ho, Shinn-Ying; Hsu, Wen-Lian; Yang, An-Suei

    2012-01-01

    Protein-protein interactions are key to many biological processes. Computational methodologies devised to predict protein-protein interaction (PPI) sites on protein surfaces are important tools in providing insights into the biological functions of proteins and in developing therapeutics targeting the protein-protein interaction sites. One of the general features of PPI sites is that the core regions from the two interacting protein surfaces are complementary to each other, similar to the interior of proteins in packing density and in the physicochemical nature of the amino acid composition. In this work, we simulated the physicochemical complementarities by constructing three-dimensional probability density maps of non-covalent interacting atoms on the protein surfaces. The interacting probabilities were derived from the interior of known structures. Machine learning algorithms were applied to learn the characteristic patterns of the probability density maps specific to the PPI sites. The trained predictors for PPI sites were cross-validated with the training cases (consisting of 432 proteins) and were tested on an independent dataset (consisting of 142 proteins). The residue-based Matthews correlation coefficient for the independent test set was 0.423; the accuracy, precision, sensitivity, specificity were 0.753, 0.519, 0.677, and 0.779 respectively. The benchmark results indicate that the optimized machine learning models are among the best predictors in identifying PPI sites on protein surfaces. In particular, the PPI site prediction accuracy increases with increasing size of the PPI site and with increasing hydrophobicity in amino acid composition of the PPI interface; the core interface regions are more likely to be recognized with high prediction confidence. The results indicate that the physicochemical complementarity patterns on protein surfaces are important determinants in PPIs, and a substantial portion of the PPI sites can be predicted correctly with

  9. Genomic prediction based on data from three layer lines: a comparison between linear methods

    NARCIS (Netherlands)

    Calus, M.P.L.; Huang, H.; Vereijken, J.; Visscher, J.; Napel, ten J.; Windig, J.J.

    2014-01-01

    Background The prediction accuracy of several linear genomic prediction models, which have previously been used for within-line genomic prediction, was evaluated for multi-line genomic prediction. Methods Compared to a conventional BLUP (best linear unbiased prediction) model using pedigree data, we

  10. Prediction of residual stress using explicit finite element method

    Directory of Open Access Journals (Sweden)

    W.A. Siswanto

    2015-12-01

    Full Text Available This paper presents the residual stress behaviour under various values of friction coefficients and scratching displacement amplitudes. The investigation is based on numerical solution using explicit finite element method in quasi-static condition. Two different aeroengine materials, i.e. Super CMV (Cr-Mo-V and Titanium alloys (Ti-6Al-4V, are examined. The usage of FEM analysis in plate under normal contact is validated with Hertzian theoretical solution in terms of contact pressure distributions. The residual stress distributions along with normal and shear stresses on elastic and plastic regimes of the materials are studied for a simple cylinder-on-flat contact configuration model subjected to normal loading, scratching and followed by unloading. The investigated friction coefficients are 0.3, 0.6 and 0.9, while scratching displacement amplitudes are 0.05 mm, 0.10 mm and 0.20 mm respectively. It is found that friction coefficient of 0.6 results in higher residual stress for both materials. Meanwhile, the predicted residual stress is proportional to the scratching displacement amplitude, higher displacement amplitude, resulting in higher residual stress. It is found that less residual stress is predicted on Super CMV material compared to Ti-6Al-4V material because of its high yield stress and ultimate strength. Super CMV material with friction coefficient of 0.3 and scratching displacement amplitude of 0.10 mm is recommended to be used in contact engineering applications due to its minimum possibility of fatigue.

  11. Success: evolutionary and structural properties of amino acids prove effective for succinylation site prediction.

    Science.gov (United States)

    López, Yosvany; Sharma, Alok; Dehzangi, Abdollah; Lal, Sunil Pranit; Taherzadeh, Ghazaleh; Sattar, Abdul; Tsunoda, Tatsuhiko

    2018-01-19

    Post-translational modification is considered an important biological mechanism with critical impact on the diversification of the proteome. Although a long list of such modifications has been studied, succinylation of lysine residues has recently attracted the interest of the scientific community. The experimental detection of succinylation sites is an expensive process, which consumes a lot of time and resources. Therefore, computational predictors of this covalent modification have emerged as a last resort to tackling lysine succinylation. In this paper, we propose a novel computational predictor called 'Success', which efficiently uses the structural and evolutionary information of amino acids for predicting succinylation sites. To do this, each lysine was described as a vector that combined the above information of surrounding amino acids. We then designed a support vector machine with a radial basis function kernel for discriminating between succinylated and non-succinylated residues. We finally compared the Success predictor with three state-of-the-art predictors in the literature. As a result, our proposed predictor showed a significant improvement over the compared predictors in statistical metrics, such as sensitivity (0.866), accuracy (0.838) and Matthews correlation coefficient (0.677) on a benchmark dataset. The proposed predictor effectively uses the structural and evolutionary information of the amino acids surrounding a lysine. The bigram feature extraction approach, while retaining the same number of features, facilitates a better description of lysines. A support vector machine with a radial basis function kernel was used to discriminate between modified and unmodified lysines. The aforementioned aspects make the Success predictor outperform three state-of-the-art predictors in succinylation detection.

  12. Measured and predicted aerosol light scattering enhancement factors at the high alpine site Jungfraujoch

    Directory of Open Access Journals (Sweden)

    R. Fierz-Schmidhauser

    2010-03-01

    Full Text Available Ambient relative humidity (RH determines the water content of atmospheric aerosol particles and thus has an important influence on the amount of visible light scattered by particles. The RH dependence of the particle light scattering coefficient (σsp is therefore an important variable for climate forcing calculations. We used a humidification system for a nephelometer which allows for the measurement of σsp at a defined RH in the range of 20–95%. In this paper we present measurements of light scattering enhancement factors f(RH=σsp(RH/σsp(dry from a 1-month campaign (May 2008 at the high alpine site Jungfraujoch (3580 m a.s.l., Switzerland. Measurements at the Jungfraujoch are representative for the lower free troposphere above Central Europe. For this aerosol type hardly any information about the f(RH is available so far. At this site, f(RH=85% varied between 1.2 and 3.3. Measured f(RH agreed well with f(RH calculated with Mie theory using measurements of the size distribution, chemical composition and hygroscopic diameter growth factors as input. Good f(RH predictions at RH<85% were also obtained with a simplified model, which uses the Ångström exponent of σsp(dry as input. RH influences further intensive optical aerosol properties. The backscatter fraction decreased by about 30% from 0.128 to 0.089, and the single scattering albedo increased on average by 0.05 at 85% RH compared to dry conditions. These changes in σsp, backscatter fraction and single scattering albedo have a distinct impact on the radiative forcing of the Jungfraujoch aerosol.

  13. Geophysical methods for fracture characterization in and around potential sites for nuclear waste disposal

    International Nuclear Information System (INIS)

    Majer, E.L.; Lee, K.H.; Morrison, H.F.

    1992-08-01

    Historically, geophysical methods have been used extensively to successfully explore the subsurface for petroleum, gas, mineral, and geothermal resources. Their application, however, for site characterization, and monitoring the performance of near surface waste sites or repositories has been somewhat limited. Presented here is an overview of the geophysical methods that could contribute to defining the subsurface heterogeneity and extrapolating point measurements at the surface and in boreholes to volumetric descriptions in a fractured rock. In addition to site characterization a significant application of geophysical methods may be in performance assessment and in monitoring the repository to determine if the performance is as expected

  14. Evaluation of a morphing based method to estimate muscle attachment sites of the lower extremity.

    Science.gov (United States)

    Pellikaan, P; van der Krogt, M M; Carbone, V; Fluit, R; Vigneron, L M; Van Deun, J; Verdonschot, N; Koopman, H F J M

    2014-03-21

    To generate subject-specific musculoskeletal models for clinical use, the location of muscle attachment sites needs to be estimated with accurate, fast and preferably automated tools. For this purpose, an automatic method was used to estimate the muscle attachment sites of the lower extremity, based on the assumption of a relation between the bone geometry and the location of muscle attachment sites. The aim of this study was to evaluate the accuracy of this morphing based method. Two cadaver dissections were performed to measure the contours of 72 muscle attachment sites on the pelvis, femur, tibia and calcaneus. The geometry of the bones including the muscle attachment sites was morphed from one cadaver to the other and vice versa. For 69% of the muscle attachment sites, the mean distance between the measured and morphed muscle attachment sites was smaller than 15 mm. Furthermore, the muscle attachment sites that had relatively large distances had shown low sensitivity to these deviations. Therefore, this morphing based method is a promising tool for estimating subject-specific muscle attachment sites in the lower extremity in a fast and automated manner. Copyright © 2013 Elsevier Ltd. All rights reserved.

  15. Analysis of the uranium price predicted to 24 months, implementing neural networks and the Monte Carlo method like predictive tools

    International Nuclear Information System (INIS)

    Esquivel E, J.; Ramirez S, J. R.; Palacios H, J. C.

    2011-11-01

    The present work shows predicted prices of the uranium, using a neural network. The importance of predicting financial indexes of an energy resource, in this case, allows establishing budgetary measures, as well as the costs of the resource to medium period. The uranium is part of the main energy generating fuels and as such, its price rebounds in the financial analyses, due to this is appealed to predictive methods to obtain an outline referent to the financial behaviour that will have in a certain time. In this study, two methodologies are used for the prediction of the uranium price: the Monte Carlo method and the neural networks. These methods allow predicting the indexes of monthly costs, for a two years period, starting from the second bimonthly of 2011. For the prediction the uranium costs are used, registered from the year 2005. (Author)

  16. ProBiS-ligands: a web server for prediction of ligands by examination of protein binding sites.

    Science.gov (United States)

    Konc, Janez; Janežič, Dušanka

    2014-07-01

    The ProBiS-ligands web server predicts binding of ligands to a protein structure. Starting with a protein structure or binding site, ProBiS-ligands first identifies template proteins in the Protein Data Bank that share similar binding sites. Based on the superimpositions of the query protein and the similar binding sites found, the server then transposes the ligand structures from those sites to the query protein. Such ligand prediction supports many activities, e.g. drug repurposing. The ProBiS-ligands web server, an extension of the ProBiS web server, is open and free to all users at http://probis.cmm.ki.si/ligands. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.

  17. Benefits of Multiple Methods for Evaluating HIV Counseling and Testing Sites in Pennsylvania.

    Science.gov (United States)

    Encandela, John A.; Gehl, Mary Beth; Silvestre, Anthony; Schelzel, George

    1999-01-01

    Examines results from two methods used to evaluate publicly funded human immunodeficiency virus (HIV) counseling and testing in Pennsylvania. Results of written mail surveys of all sites and interviews from a random sample of 30 sites were similar in terms of questions posed and complementary in other ways. (SLD)

  18. Evaluation of a morphing based method to estimate muscle attachment sites of the lower extremity

    NARCIS (Netherlands)

    Pellikaan, P.; van der Krogt, Marjolein; Carbone, Vincenzo; Fluit, René; Vigneron, L.M.; van Deun, J.; Verdonschot, Nicolaas Jacobus Joseph; Koopman, Hubertus F.J.M.

    2014-01-01

    To generate subject-specific musculoskeletal models for clinical use, the location of muscle attachment sites needs to be estimated with accurate, fast and preferably automated tools. For this purpose, an automatic method was used to estimate the muscle attachment sites of the lower extremity, based

  19. Using a Web Site in an Elementary Science Methods Class: Are We Opening a Pandora's Box?

    Science.gov (United States)

    Lewis, Scott P.; O'Brien, George E.

    This paper describes the introduction and use of the World Wide Web (WWW) in an elementary science methods course at Florida International University (FIU). The goals of creating a web site include engaging conversations among educators, providing access to local resources for students, and examining student use of web sites and the Internet. The…

  20. Prediction method of seismic residual deformation of caisson quay wall in liquefied foundation

    Science.gov (United States)

    Wang, Li-Yan; Liu, Han-Long; Jiang, Peng-Ming; Chen, Xiang-Xiang

    2011-03-01

    The multi-spring shear mechanism plastic model in this paper is defined in strain space to simulate pore pressure generation and development in sands under cyclic loading and undrained conditions, and the rotation of principal stresses can also be simulated by the model with cyclic behavior of anisotropic consolidated sands. Seismic residual deformations of typical caisson quay walls under different engineering situations are analyzed in detail by the plastic model, and then an index of liquefaction extent is applied to describe the regularity of seismic residual deformation of caisson quay wall top under different engineering situations. Some correlated prediction formulas are derived from the results of regression analysis between seismic residual deformation of quay wall top and extent of liquefaction in the relative safety backfill sand site. Finally, the rationality and the reliability of the prediction methods are validated by test results of a 120 g-centrifuge shaking table, and the comparisons show that some reliable seismic residual deformation of caisson quay can be predicted by appropriate prediction formulas and appropriate index of liquefaction extent.

  1. Topography and geology site effects from the intensity prediction model (ShakeMap) for Austria

    Science.gov (United States)

    del Puy Papí Isaba, María; Jia, Yan; Weginger, Stefan

    2017-04-01

    The seismicity in Austria can be categorized as moderated. Despite the fact that the hazard seems to be rather low, earthquakes can cause great damage and losses, specially in densely populated and industrialized areas. It is well known, that equations which predict intensity as a function of magnitude and distance, among other parameters, are useful tool for hazard and risk assessment. Therefore, this study aims to determine an empirical model of the ground shaking intensities (ShakeMap) of a series of earthquakes occurred in Austria between 1000 and 2014. Furthermore, the obtained empirical model will lead to further interpretation of both, contemporary and historical earthquakes. A total of 285 events, which epicenters were located in Austria, and a sum of 22.739 reported macreoseismic data points from Austria and adjoining countries, were used. These events are enclosed in the period 1000-2014 and characterized by having a local magnitude greater than 3. In the first state of the model development, the data was careful selected, e.g. solely intensities equal or greater than III were used. In a second state the data was adjusted to the selected empirical model. Finally, geology and topography corrections were obtained by means of the model residuals in order to derive intensity-based site amplification effects.

  2. Stimulation site within the MRI-defined STN predicts postoperative motor outcome.

    Science.gov (United States)

    Wodarg, Fritz; Herzog, Jan; Reese, René; Falk, Daniela; Pinsker, Markus O; Steigerwald, Frank; Jansen, Olav; Deuschl, Günther; Mehdorn, H Maximillian; Volkmann, Jens

    2012-06-01

    High-frequency stimulation of the subthalamic nucleus (STN-HFS) is highly effective in treating motor symptoms in Parkinson's disease (PD) and medication side effects as well as in improving quality of life. Despite preoperative screening for patients as eligible candidates for this treatment, electrode position may furthermore influence treatment quality. Here, we investigated the relationship between the anatomical site of stimulation within the MRI-defined STN and the outcome of PD patients after STN-HFS. In 30 PD patients with bilateral STN stimulation, we retrospectively defined the boundaries of the STN within the axial target plane of the stereotactic T2-weighted MRI and determined the position of the active electrode contact in relation to the border of the STN. The position of the active contact within the STN was the only variable to predict the outcome of STN stimulation. In contrast, covariates such as age, disease duration, symptom severity, and response to levodopa had no effect. The lateral position of the stimulation contact within the STN led to significantly better clinical improvement, lower stimulation parameters, and less need for postoperative dopaminergic medication. The outcome of patients with stimulation contacts within the medial region of the STN was significantly worse. Precise targeting of the lateral region of the STN is essential for achieving sufficient stimulation efficacy. Preoperative T2-weighted MRI might be a useful component of the targeting procedure to improve the outcome of PD patients. Copyright © 2012 Movement Disorder Society.

  3. COMPARATIVE MODELLING AND LIGAND BINDING SITE PREDICTION OF A FAMILY 43 GLYCOSIDE HYDROLASE FROM Clostridium thermocellum

    Directory of Open Access Journals (Sweden)

    Shadab Ahmed

    2012-06-01

    Full Text Available The phylogenetic analysis of Clostridium thermocellum family 43 glycoside hydrolase (CtGH43 showed close evolutionary relation with carbohydrate binding family 6 proteins from C. cellulolyticum, C. papyrosolvens, C. cellulyticum, and A. cellulyticum. Comparative modeling of CtGH43 was performed based on crystal structures with PDB IDs 3C7F, 1YIF, 1YRZ, 2EXH and 1WL7. The structure having lowest MODELLER objective function was selected. The three-dimensional structure revealed typical 5-fold beta–propeller architecture. Energy minimization and validation of predicted model with VERIFY 3D indicated acceptability of the proposed atomic structure. The Ramachandran plot analysis by RAMPAGE confirmed that family 43 glycoside hydrolase (CtGH43 contains little or negligible segments of helices. It also showed that out of 301 residues, 267 (89.3% were in most favoured region, 23 (7.7% were in allowed region and 9 (3.0% were in outlier region. IUPred analysis of CtGH43 showed no disordered region. Active site analysis showed presence of two Asp and one Glu, assumed to form a catalytic triad. This study gives us information about three-dimensional structure and reaffirms the fact that it has the similar core 5-fold beta–propeller architecture and so probably has the same inverting mechanism of action with the formation of above mentioned catalytic triad for catalysis of polysaccharides.

  4. Experimental method to predict avalanches based on neural networks

    Directory of Open Access Journals (Sweden)

    V. V. Zhdanov

    2016-01-01

    Full Text Available The article presents results of experimental use of currently available statistical methods to classify the avalanche‑dangerous precipitations and snowfalls in the Kishi Almaty river basin. The avalanche service of Kazakhstan uses graphical methods for prediction of avalanches developed by I.V. Kondrashov and E.I. Kolesnikov. The main objective of this work was to develop a modern model that could be used directly at the avalanche stations. Classification of winter precipitations into dangerous snowfalls and non‑dangerous ones was performed by two following ways: the linear discriminant function (canonical analysis and artificial neural networks. Observational data on weather and avalanches in the gorge Kishi Almaty in the gorge Kishi Almaty were used as a training sample. Coefficients for the canonical variables were calculated by the software «Statistica» (Russian version 6.0, and then the necessary formula had been constructed. The accuracy of the above classification was 96%. Simulator by the authors L.N. Yasnitsky and F.М. Cherepanov was used to learn the neural networks. The trained neural network demonstrated 98% accuracy of the classification. Prepared statistical models are recommended to be tested at the snow‑avalanche stations. Results of the tests will be used for estimation of the model quality and its readiness for the operational work. In future, we plan to apply these models for classification of the avalanche danger by the five‑point international scale.

  5. A METHOD OF PREDICTING BREAST CANCER USING QUESTIONNAIRES

    Directory of Open Access Journals (Sweden)

    V. N. Malashenko

    2017-01-01

    Full Text Available Purpose. Simplify and increase the accuracy of the questionnaire method of predicting breast cancer (BC for subsequent computer processing and Automated dispensary at risk without the doctor.Materials and methods. The work was based on statistical data obtained by surveying 305 women. The questionnaire included 63 items: 17 open-ended questions, 46 — with a choice of response. It was established multifactor model, the development of which, in addition to the survey data were used materials from the medical histories of patients and respondents data immuno-histochemical studies. Data analysis was performed using Statistica 10.0 and MedCalc 12.7.0 programs.Results. The ROC analysis was performas and the questionnaire data revealed 8 significant predictors of breast cancer. On their basis we created the formula for calculating the prognostic factor of risk of development of breast cancer with a sensitivity 83,12% and a specificity of 91,43%.Conclusions. The completed developments allow to create a computer program for automated processing of profiles on the formation of groups at risk of breast cancer and clinical supervision. The introduction of a screening questionnaire over the Internet with subsequent computer processing of the results, without the direct involvement of doctors, will increase the coverage of the female population of the Russian Federation activities related to the prevention of breast cancer. It can free up time for physicians to receive primary patients, as well as improve oncological vigilance of the female population of the Russian Federation.

  6. Site Classification using Multichannel Channel Analysis of Surface Wave (MASW) method on Soft and Hard Ground

    Science.gov (United States)

    Ashraf, M. A. M.; Kumar, N. S.; Yusoh, R.; Hazreek, Z. A. M.; Aziman, M.

    2018-04-01

    Site classification utilizing average shear wave velocity (Vs(30) up to 30 meters depth is a typical parameter. Numerous geophysical methods have been proposed for estimation of shear wave velocity by utilizing assortment of testing configuration, processing method, and inversion algorithm. Multichannel Analysis of Surface Wave (MASW) method is been rehearsed by numerous specialist and professional to geotechnical engineering for local site characterization and classification. This study aims to determine the site classification on soft and hard ground using MASW method. The subsurface classification was made utilizing National Earthquake Hazards Reduction Program (NERHP) and international Building Code (IBC) classification. Two sites are chosen to acquire the shear wave velocity which is in the state of Pulau Pinang for soft soil and Perlis for hard rock. Results recommend that MASW technique can be utilized to spatially calculate the distribution of shear wave velocity (Vs(30)) in soil and rock to characterize areas.

  7. Alternative methods of salt disposal at the seven salt sites for a nuclear waste repository

    International Nuclear Information System (INIS)

    1987-02-01

    This study discusses the various alternative salt management techniques for the disposal of excess mined salt at seven potentially acceptable nuclear waste repository sites: Deaf Smith and Swisher Counties, Texas; Richton and Cypress Creek Domes, Mississippi; Vacherie Dome, Louisiana; and Davis and Lavender Canyons, Utah. Because the repository development involves the underground excavation of corridors and waste emplacement rooms, in either bedded or domed salt formations, excess salt will be mined and must be disposed of offsite. The salt disposal alternatives examined for all the sites include commercial use, ocean disposal, deep well injection, landfill disposal, and underground mine disposal. These alternatives (and other site-specific disposal methods) are reviewed, using estimated amounts of excavated, backfilled, and excess salt. Methods of transporting the excess salt are discussed, along with possible impacts of each disposal method and potential regulatory requirements. A preferred method of disposal is recommended for each potentially acceptable repository site. 14 refs., 5 tabs

  8. Numerical Modeling Tools for the Prediction of Solution Migration Applicable to Mining Site

    International Nuclear Information System (INIS)

    Martell, M.; Vaughn, P.

    1999-01-01

    Mining has always had an important influence on cultures and traditions of communities around the globe and throughout history. Today, because mining legislation places heavy emphasis on environmental protection, there is great interest in having a comprehensive understanding of ancient mining and mining sites. Multi-disciplinary approaches (i.e., Pb isotopes as tracers) are being used to explore the distribution of metals in natural environments. Another successful approach is to model solution migration numerically. A proven method to simulate solution migration in natural rock salt has been applied to project through time for 10,000 years the system performance and solution concentrations surrounding a proposed nuclear waste repository. This capability is readily adaptable to simulate solution migration around mining

  9. Application of a simple parameter estimation method to predict effluent transport in the Savannah River

    International Nuclear Information System (INIS)

    Hensel, S.J.; Hayes, D.W.

    1993-01-01

    A simple parameter estimation method has been developed to determine the dispersion and velocity parameters associated with stream/river transport. The unsteady one dimensional Burgers' equation was chosen as the model equation, and the method has been applied to recent Savannah River dye tracer studies. The computed Savannah River transport coefficients compare favorably with documented values, and the time/concentration curves calculated from these coefficients compare well with the actual tracer data. The coefficients were used as a predictive capability and applied to Savannah River tritium concentration data obtained during the December 1991 accidental tritium discharge from the Savannah River Site. The peak tritium concentration at the intersection of Highway 301 and the Savannah River was underpredicted by only 5% using the coefficients computed from the dye data

  10. The Moulded Site Data (MSD) wind correlation method: description and assessment

    Energy Technology Data Exchange (ETDEWEB)

    King, C.; Hurley, B.

    2004-12-01

    The long-term wind resource at a potential windfarm site may be estimated by correlating short-term on-site wind measurements with data from a regional meteorological station. A correlation method developed at Airtricity is described in sufficient detail to be reproduced. An assessment of its performance is also described; the results may serve as a guide to expected accuracy when using the method as part of an annual electricity production estimate for a proposed windfarm. (Author)

  11. The Comparison Study of Short-Term Prediction Methods to Enhance the Model Predictive Controller Applied to Microgrid Energy Management

    Directory of Open Access Journals (Sweden)

    César Hernández-Hernández

    2017-06-01

    Full Text Available Electricity load forecasting, optimal power system operation and energy management play key roles that can bring significant operational advantages to microgrids. This paper studies how methods based on time series and neural networks can be used to predict energy demand and production, allowing them to be combined with model predictive control. Comparisons of different prediction methods and different optimum energy distribution scenarios are provided, permitting us to determine when short-term energy prediction models should be used. The proposed prediction models in addition to the model predictive control strategy appear as a promising solution to energy management in microgrids. The controller has the task of performing the management of electricity purchase and sale to the power grid, maximizing the use of renewable energy sources and managing the use of the energy storage system. Simulations were performed with different weather conditions of solar irradiation. The obtained results are encouraging for future practical implementation.

  12. Predicting risk of trace element pollution from municipal roads using site-specific soil samples and remotely sensed data.

    Science.gov (United States)

    Reeves, Mari Kathryn; Perdue, Margaret; Munk, Lee Ann; Hagedorn, Birgit

    2018-07-15

    Studies of environmental processes exhibit spatial variation within data sets. The ability to derive predictions of risk from field data is a critical path forward in understanding the data and applying the information to land and resource management. Thanks to recent advances in predictive modeling, open source software, and computing, the power to do this is within grasp. This article provides an example of how we predicted relative trace element pollution risk from roads across a region by combining site specific trace element data in soils with regional land cover and planning information in a predictive model framework. In the Kenai Peninsula of Alaska, we sampled 36 sites (191 soil samples) adjacent to roads for trace elements. We then combined this site specific data with freely-available land cover and urban planning data to derive a predictive model of landscape scale environmental risk. We used six different model algorithms to analyze the dataset, comparing these in terms of their predictive abilities and the variables identified as important. Based on comparable predictive abilities (mean R 2 from 30 to 35% and mean root mean square error from 65 to 68%), we averaged all six model outputs to predict relative levels of trace element deposition in soils-given the road surface, traffic volume, sample distance from the road, land cover category, and impervious surface percentage. Mapped predictions of environmental risk from toxic trace element pollution can show land managers and transportation planners where to prioritize road renewal or maintenance by each road segment's relative environmental and human health risk. Published by Elsevier B.V.

  13. Methods and codes for assessing the off-site Consequences of nuclear accidents. Volume 2

    International Nuclear Information System (INIS)

    Kelly, G.N.; Luykx, F.

    1991-01-01

    The Commission of the European Communities, within the framework of its 1980-84 radiation protection research programme, initiated a two-year project in 1983 entitled methods for assessing the radiological impact of accidents (Maria). This project was continued in a substantially enlarged form within the 1985-89 research programme. The main objectives of the project were, firstly, to develop a new probabilistic accident consequence code that was modular, incorporated the best features of those codes already in use, could be readily modified to take account of new data and model developments and would be broadly applicable within the EC; secondly, to acquire a better understanding of the limitations of current models and to develop more rigorous approaches where necessary; and, thirdly, to quantify the uncertainties associated with the model predictions. This research led to the development of the accident consequence code Cosyma (COde System from MAria), which will be made generally available later in 1990. The numerous and diverse studies that have been undertaken in support of this development are summarized in this paper, together with indications of where further effort might be most profitably directed. Consideration is also given to related research directed towards the development of real-time decision support systems for use in off-site emergency management

  14. Bayesian Poisson hierarchical models for crash data analysis: Investigating the impact of model choice on site-specific predictions.

    Science.gov (United States)

    Khazraee, S Hadi; Johnson, Valen; Lord, Dominique

    2018-08-01

    The Poisson-gamma (PG) and Poisson-lognormal (PLN) regression models are among the most popular means for motor vehicle crash data analysis. Both models belong to the Poisson-hierarchical family of models. While numerous studies have compared the overall performance of alternative Bayesian Poisson-hierarchical models, little research has addressed the impact of model choice on the expected crash frequency prediction at individual sites. This paper sought to examine whether there are any trends among candidate models predictions e.g., that an alternative model's prediction for sites with certain conditions tends to be higher (or lower) than that from another model. In addition to the PG and PLN models, this research formulated a new member of the Poisson-hierarchical family of models: the Poisson-inverse gamma (PIGam). Three field datasets (from Texas, Michigan and Indiana) covering a wide range of over-dispersion characteristics were selected for analysis. This study demonstrated that the model choice can be critical when the calibrated models are used for prediction at new sites, especially when the data are highly over-dispersed. For all three datasets, the PIGam model would predict higher expected crash frequencies than would the PLN and PG models, in order, indicating a clear link between the models predictions and the shape of their mixing distributions (i.e., gamma, lognormal, and inverse gamma, respectively). The thicker tail of the PIGam and PLN models (in order) may provide an advantage when the data are highly over-dispersed. The analysis results also illustrated a major deficiency of the Deviance Information Criterion (DIC) in comparing the goodness-of-fit of hierarchical models; models with drastically different set of coefficients (and thus predictions for new sites) may yield similar DIC values, because the DIC only accounts for the parameters in the lowest (observation) level of the hierarchy and ignores the higher levels (regression coefficients

  15. NetPhosYeast: prediction of protein phosphorylation sites in yeast

    DEFF Research Database (Denmark)

    Ingrell, C.R.; Miller, Martin Lee; Jensen, O.N.

    2007-01-01

    sites compared to those in humans, suggesting the need for an yeast-specific phosphorylation site predictor. NetPhosYeast achieves a correlation coefficient close to 0.75 with a sensitivity of 0.84 and specificity of 0.90 and outperforms existing predictors in the identification of phosphorylation sites...

  16. Neural network model for the prediction of PM10 daily concentrations in two sites in the Western Mediterranean.

    Science.gov (United States)

    de Gennaro, Gianluigi; Trizio, Livia; Di Gilio, Alessia; Pey, Jorge; Pérez, Noemi; Cusack, Michael; Alastuey, Andrés; Querol, Xavier

    2013-10-01

    An artificial neural network (ANN) was developed and tested to forecast PM10 daily concentration in two contrasted environments in NE Spain, a regional background site (Montseny), and an urban background site (Barcelona-CSIC), which was highly influenced by vehicular emissions. In order to predict 24-h average PM10 concentrations, the artificial neural network previously developed by Caselli et al. (2009) was improved by using hourly PM concentrations and deterministic factors such as a Saharan dust alert. In particular, the model input data for prediction were the hourly PM10 concentrations 1-day in advance, local meteorological data and information about air masses origin. The forecasted performance indexes for both sites were calculated and they showed better results for the regional background site in Montseny (R(2)=0.86, SI=0.75) than for urban site in Barcelona (R(2)=0.73, SI=0.58), influenced by local and sometimes unexpected sources. Moreover, a sensitivity analysis conducted to understand the importance of the different variables included among the input data, showed that local meteorology and air masses origin are key factors in the model forecasts. This result explains the reason for the improvement of ANN's forecasting performance at the Montseny site with respect to the Barcelona site. Moreover, the artificial neural network developed in this work could prove useful to predict PM10 concentrations, especially, at regional background sites such as those on the Mediterranean Basin which are primarily affected by long-range transports. Hence, the artificial neural network presented here could be a powerful tool for obtaining real time information on air quality status and could aid stakeholders in their development of cost-effective control strategies. © 2013 Elsevier B.V. All rights reserved.

  17. Skill forecasting from different wind power ensemble prediction methods

    International Nuclear Information System (INIS)

    Pinson, Pierre; Nielsen, Henrik A; Madsen, Henrik; Kariniotakis, George

    2007-01-01

    This paper presents an investigation on alternative approaches to the providing of uncertainty estimates associated to point predictions of wind generation. Focus is given to skill forecasts in the form of prediction risk indices, aiming at giving a comprehensive signal on the expected level of forecast uncertainty. Ensemble predictions of wind generation are used as input. A proposal for the definition of prediction risk indices is given. Such skill forecasts are based on the dispersion of ensemble members for a single prediction horizon, or over a set of successive look-ahead times. It is shown on the test case of a Danish offshore wind farm how prediction risk indices may be related to several levels of forecast uncertainty (and energy imbalances). Wind power ensemble predictions are derived from the transformation of ECMWF and NCEP ensembles of meteorological variables to power, as well as by a lagged average approach alternative. The ability of risk indices calculated from the various types of ensembles forecasts to resolve among situations with different levels of uncertainty is discussed

  18. Determination of antenna factors using a three-antenna method at open-field test site

    Science.gov (United States)

    Masuzawa, Hiroshi; Tejima, Teruo; Harima, Katsushige; Morikawa, Takao

    1992-09-01

    Recently NIST has used the three-antenna method for calibration of the antenna factor of an antenna used for EMI measurements. This method does not require the specially designed standard antennas which are necessary in the standard field method or the standard antenna method, and can be used at an open-field test site. This paper theoretically and experimentally examines the measurement errors of this method and evaluates the precision of the antenna-factor calibration. It is found that the main source of the error is the non-ideal propagation characteristics of the test site, which should therefore be measured before the calibration. The precision of the antenna-factor calibration at the test site used in these experiments, is estimated to be 0.5 dB.

  19. Prediction and characterization of novel epitopes of serotype A foot-and-mouth disease viruses circulating in East Africa using site-directed mutagenesis

    Science.gov (United States)

    Bari, Fufa Dawo; Parida, Satya; Asfor, Amin S.; Haydon, Daniel T.; Reeve, Richard; Paton, David J.

    2015-01-01

    Epitopes on the surface of the foot-and-mouth disease virus (FMDV) capsid have been identified by monoclonal antibody (mAb) escape mutant studies leading to the designation of four antigenic sites in serotype A FMDV. Previous work focused on viruses isolated mainly from Asia, Europe and Latin America. In this study we report on the prediction of epitopes in African serotype A FMDVs and testing of selected epitopes using reverse genetics. Twenty-four capsid amino acid residues were predicted to be of antigenic significance by analysing the capsid sequences (n = 56) using in silico methods, and six residues by correlating capsid sequence with serum–virus neutralization data. The predicted residues were distributed on the surface-exposed capsid regions, VP1–VP3. The significance of residue changes at eight of the predicted epitopes was tested by site-directed mutagenesis using a cDNA clone resulting in the generation of 12 mutant viruses involving seven sites. The effect of the amino acid substitutions on the antigenic nature of the virus was assessed by virus neutralization (VN) test. Mutations at four different positions, namely VP1-43, VP1-45, VP2-191 and VP3-132, led to significant reduction in VN titre (P value = 0.05, 0.05, 0.001 and 0.05, respectively). This is the first time, to our knowledge, that the antigenic regions encompassing amino acids VP1-43 to -45 (equivalent to antigenic site 3 in serotype O), VP2-191 and VP3-132 have been predicted as epitopes and evaluated serologically for serotype A FMDVs. This identifies novel capsid epitopes of recently circulating serotype A FMDVs in East Africa. PMID:25614587

  20. Predictive mutagenesis of ligation-independent cloning (LIC) vectors for protein expression and site-specific chemical conjugation

    DEFF Research Database (Denmark)

    Vernet, Erik; Sauer, Jørgen; Andersen, Peter Andreas

    2011-01-01

    Ligation-independent cloning (LIC) allows for cloning of DNA constructs independent of insert restriction sites and ligases. However, any required mutations are typically introduced by additional, time-consuming steps. We present a rapid, inexpensive method for mutagenesis in the 5' LIC site...

  1. A prediction method based on wavelet transform and multiple models fusion for chaotic time series

    International Nuclear Information System (INIS)

    Zhongda, Tian; Shujiang, Li; Yanhong, Wang; Yi, Sha

    2017-01-01

    In order to improve the prediction accuracy of chaotic time series, a prediction method based on wavelet transform and multiple models fusion is proposed. The chaotic time series is decomposed and reconstructed by wavelet transform, and approximate components and detail components are obtained. According to different characteristics of each component, least squares support vector machine (LSSVM) is used as predictive model for approximation components. At the same time, an improved free search algorithm is utilized for predictive model parameters optimization. Auto regressive integrated moving average model (ARIMA) is used as predictive model for detail components. The multiple prediction model predictive values are fusion by Gauss–Markov algorithm, the error variance of predicted results after fusion is less than the single model, the prediction accuracy is improved. The simulation results are compared through two typical chaotic time series include Lorenz time series and Mackey–Glass time series. The simulation results show that the prediction method in this paper has a better prediction.

  2. V S30, slope, H 800 and f 0: performance of various site-condition proxies in reducing ground-motion aleatory variability and predicting nonlinear site response

    Science.gov (United States)

    Derras, Boumédiène; Bard, Pierre-Yves; Cotton, Fabrice

    2017-09-01

    The aim of this paper is to investigate the ability of various site-condition proxies (SCPs) to reduce ground-motion aleatory variability and evaluate how SCPs capture nonlinearity site effects. The SCPs used here are time-averaged shear-wave velocity in the top 30 m ( V S30), the topographical slope (slope), the fundamental resonance frequency ( f 0) and the depth beyond which V s exceeds 800 m/s ( H 800). We considered first the performance of each SCP taken alone and then the combined performance of the 6 SCP pairs [ V S30- f 0], [ V S30- H 800], [ f 0-slope], [ H 800-slope], [ V S30-slope] and [ f 0- H 800]. This analysis is performed using a neural network approach including a random effect applied on a KiK-net subset for derivation of ground-motion prediction equations setting the relationship between various ground-motion parameters such as peak ground acceleration, peak ground velocity and pseudo-spectral acceleration PSA ( T), and M w, R JB, focal depth and SCPs. While the choice of SCP is found to have almost no impact on the median ground-motion prediction, it does impact the level of aleatory uncertainty. V S30 is found to perform the best of single proxies at short periods ( T < 0.6 s), while f 0 and H 800 perform better at longer periods; considering SCP pairs leads to significant improvements, with particular emphasis on [ V S30- H 800] and [ f 0-slope] pairs. The results also indicate significant nonlinearity on the site terms for soft sites and that the most relevant loading parameter for characterising nonlinear site response is the "stiff" spectral ordinate at the considered period.[Figure not available: see fulltext.

  3. The Surgical Site Infection Risk Score (SSIRS: A Model to Predict the Risk of Surgical Site Infections.

    Directory of Open Access Journals (Sweden)

    Carl van Walraven

    Full Text Available Surgical site infections (SSI are an important cause of peri-surgical morbidity with risks that vary extensively between patients and surgeries. Quantifying SSI risk would help identify candidates most likely to benefit from interventions to decrease the risk of SSI.We randomly divided all surgeries recorded in the National Surgical Quality Improvement Program from 2010 into a derivation and validation population. We used multivariate logistic regression to determine the independent association of patient and surgical covariates with the risk of any SSI (including superficial, deep, and organ space SSI within 30 days of surgery. To capture factors particular to specific surgeries, we developed a surgical risk score specific to all surgeries having a common first 3 numbers of their CPT code.Derivation (n = 181 894 and validation (n = 181 146 patients were similar for all demographics, past medical history, and surgical factors. Overall SSI risk was 3.9%. The SSI Risk Score (SSIRS found that risk increased with patient factors (smoking, increased body mass index, certain comorbidities (peripheral vascular disease, metastatic cancer, chronic steroid use, recent sepsis, and operative characteristics (surgical urgency; increased ASA class; longer operation duration; infected wounds; general anaesthesia; performance of more than one procedure; and CPT score. In the validation population, the SSIRS had good discrimination (c-statistic 0.800, 95% CI 0.795-0.805 and calibration.SSIRS can be calculated using patient and surgery information to estimate individual risk of SSI for a broad range of surgery types.

  4. GPS 2.1: enhanced prediction of kinase-specific phosphorylation sites with an algorithm of motif length selection.

    Science.gov (United States)

    Xue, Yu; Liu, Zexian; Cao, Jun; Ma, Qian; Gao, Xinjiao; Wang, Qingqi; Jin, Changjiang; Zhou, Yanhong; Wen, Longping; Ren, Jian

    2011-03-01

    As the most important post-translational modification of proteins, phosphorylation plays essential roles in all aspects of biological processes. Besides experimental approaches, computational prediction of phosphorylated proteins with their kinase-specific phosphorylation sites has also emerged as a popular strategy, for its low-cost, fast-speed and convenience. In this work, we developed a kinase-specific phosphorylation sites predictor of GPS 2.1 (Group-based Prediction System), with a novel but simple approach of motif length selection (MLS). By this approach, the robustness of the prediction system was greatly improved. All algorithms in GPS old versions were also reserved and integrated in GPS 2.1. The online service and local packages of GPS 2.1 were implemented in JAVA 1.5 (J2SE 5.0) and freely available for academic researches at: http://gps.biocuckoo.org.

  5. Implications of next generation attenuation ground motion prediction equations for site coefficients used in earthquake resistant design

    Science.gov (United States)

    Borcherdt, Roger D.

    2014-01-01

    Proposals are developed to update Tables 11.4-1 and 11.4-2 of Minimum Design Loads for Buildings and Other Structures published as American Society of Civil Engineers Structural Engineering Institute standard 7-10 (ASCE/SEI 7–10). The updates are mean next generation attenuation (NGA) site coefficients inferred directly from the four NGA ground motion prediction equations used to derive the maximum considered earthquake response maps adopted in ASCE/SEI 7–10. Proposals include the recommendation to use straight-line interpolation to infer site coefficients at intermediate values of (average shear velocity to 30-m depth). The NGA coefficients are shown to agree well with adopted site coefficients at low levels of input motion (0.1 g) and those observed from the Loma Prieta earthquake. For higher levels of input motion, the majority of the adopted values are within the 95% epistemic-uncertainty limits implied by the NGA estimates with the exceptions being the mid-period site coefficient, Fv, for site class D and the short-period coefficient, Fa, for site class C, both of which are slightly less than the corresponding 95% limit. The NGA data base shows that the median value  of 913 m/s for site class B is more typical than 760 m/s as a value to characterize firm to hard rock sites as the uniform ground condition for future maximum considered earthquake response ground motion estimates. Future updates of NGA ground motion prediction equations can be incorporated easily into future adjustments of adopted site coefficients using procedures presented herein. 

  6. Seismic analysis of APR1400 RCS for site envelope using big mass method

    International Nuclear Information System (INIS)

    Kim, J. Y.; Jeon, J. H.; Lee, D. H.; Park, S. H.

    2002-01-01

    One of design concepts of APR1400 is the site envelope considering various soil sites as well as rock site. The KSNP's are constructed on the rock site where only the translational excitations are directly transferred to the plant. On the other hand, the rotational motions affect the responses of the structures in the soil cases. In this study, a Big Mass Method is used to consider rotational motions as excitations at the foundation in addition to translational ones to obtain seismic responses of the APR1400 RCS main components. The seismic analyses for the APR1400 excited simultaneously by translation and rotational motions were performed. The results show that the effect of soil sites is not significant for the design of main components and supports of the RCS, but it may be considerable for the design of reactor vessel internals, piping, and nozzles which have lower natural frequencies

  7. Multiple Site-Directed and Saturation Mutagenesis by the Patch Cloning Method.

    Science.gov (United States)

    Taniguchi, Naohiro; Murakami, Hiroshi

    2017-01-01

    Constructing protein-coding genes with desired mutations is a basic step for protein engineering. Herein, we describe a multiple site-directed and saturation mutagenesis method, termed MUPAC. This method has been used to introduce multiple site-directed mutations in the green fluorescent protein gene and in the moloney murine leukemia virus reverse transcriptase gene. Moreover, this method was also successfully used to introduce randomized codons at five desired positions in the green fluorescent protein gene, and for simple DNA assembly for cloning.

  8. Different protein-protein interface patterns predicted by different machine learning methods.

    Science.gov (United States)

    Wang, Wei; Yang, Yongxiao; Yin, Jianxin; Gong, Xinqi

    2017-11-22

    Different types of protein-protein interactions make different protein-protein interface patterns. Different machine learning methods are suitable to deal with different types of data. Then, is it the same situation that different interface patterns are preferred for prediction by different machine learning methods? Here, four different machine learning methods were employed to predict protein-protein interface residue pairs on different interface patterns. The performances of the methods for different types of proteins are different, which suggest that different machine learning methods tend to predict different protein-protein interface patterns. We made use of ANOVA and variable selection to prove our result. Our proposed methods taking advantages of different single methods also got a good prediction result compared to single methods. In addition to the prediction of protein-protein interactions, this idea can be extended to other research areas such as protein structure prediction and design.

  9. Predicting Solar Activity Using Machine-Learning Methods

    Science.gov (United States)

    Bobra, M.

    2017-12-01

    Of all the activity observed on the Sun, two of the most energetic events are flares and coronal mass ejections. However, we do not, as of yet, fully understand the physical mechanism that triggers solar eruptions. A machine-learning algorithm, which is favorable in cases where the amount of data is large, is one way to [1] empirically determine the signatures of this mechanism in solar image data and [2] use them to predict solar activity. In this talk, we discuss the application of various machine learning algorithms - specifically, a Support Vector Machine, a sparse linear regression (Lasso), and Convolutional Neural Network - to image data from the photosphere, chromosphere, transition region, and corona taken by instruments aboard the Solar Dynamics Observatory in order to predict solar activity on a variety of time scales. Such an approach may be useful since, at the present time, there are no physical models of flares available for real-time prediction. We discuss our results (Bobra and Couvidat, 2015; Bobra and Ilonidis, 2016; Jonas et al., 2017) as well as other attempts to predict flares using machine-learning (e.g. Ahmed et al., 2013; Nishizuka et al. 2017) and compare these results with the more traditional techniques used by the NOAA Space Weather Prediction Center (Crown, 2012). We also discuss some of the challenges in using machine-learning algorithms for space science applications.

  10. The size, morphology, site, and access score predicts critical outcomes of endoscopic mucosal resection in the colon.

    Science.gov (United States)

    Sidhu, Mayenaaz; Tate, David J; Desomer, Lobke; Brown, Gregor; Hourigan, Luke F; Lee, Eric Y T; Moss, Alan; Raftopoulos, Spiro; Singh, Rajvinder; Williams, Stephen J; Zanati, Simon; Burgess, Nicholas; Bourke, Michael J

    2018-01-25

    The SMSA (size, morphology, site, access) polyp scoring system is a method of stratifying the difficulty of polypectomy through assessment of four domains. The aim of this study was to evaluate the ability of SMSA to predict critical outcomes of endoscopic mucosal resection (EMR). We retrospectively applied SMSA to a prospectively collected multicenter database of large colonic laterally spreading lesions (LSLs) ≥ 20 mm referred for EMR. Standard inject-and-resect EMR procedures were performed. The primary end points were correlation of SMSA level with technical success, adverse events, and endoscopic recurrence. 2675 lesions in 2675 patients (52.6 % male) underwent EMR. Failed single-session EMR occurred in 124 LSLs (4.6 %) and was predicted by the SMSA score ( P  < 0.001). Intraprocedural and clinically significant postendoscopic bleeding was significantly less common for SMSA 2 LSLs (odds ratio [OR] 0.36, P  < 0.001 and OR 0.23, P  < 0.01) and SMSA 3 LSLs (OR 0.41, P  < 0.001 and OR 0.60, P  = 0.05) compared with SMSA 4 lesions. Similarly, endoscopic recurrence at first surveillance was less likely among SMSA 2 (OR 0.19, P  < 0.001) and SMSA 3 (OR 0.33, P  < 0.001) lesions compared with SMSA 4 lesions. This also extended to second surveillance among SMSA 4 LSLs. SMSA is a simple, readily applicable, clinical score that identifies a subgroup of patients who are at increased risk of failed EMR, adverse events, and adenoma recurrence at surveillance colonoscopy. This information may be useful for improving informed consent, planning endoscopy lists, and developing quality control measures for practitioners of EMR, with potential implications for EMR benchmarking and training. © Georg Thieme Verlag KG Stuttgart · New York.

  11. Predicting summer site occupancy for an invasive species, the common brushtail possum (Trichosurus vulpecula, in an urban environment.

    Directory of Open Access Journals (Sweden)

    Amy L Adams

    Full Text Available Invasive species are often favoured in fragmented, highly-modified, human-dominated landscapes such as urban areas. Because successful invasive urban adapters can occupy habitat that is quite different from that in their original range, effective management programmes for invasive species in urban areas require an understanding of distribution, habitat and resource requirements at a local scale that is tailored to the fine-scale heterogeneity typical of urban landscapes. The common brushtail possum (Trichosurus vulpecula is one of New Zealand's most destructive invasive pest species. As brushtail possums traditionally occupy forest habitat, control in New Zealand has focussed on rural and forest habitats, and forest fragments in cities. However, as successful urban adapters, possums may be occupying a wider range of habitats. Here we use site occupancy methods to determine the distribution of brushtail possums across five distinguishable urban habitat types during summer, which is when possums have the greatest impacts on breeding birds. We collected data on possum presence/absence and habitat characteristics, including possible sources of supplementary food (fruit trees, vegetable gardens, compost heaps, and the availability of forest fragments from 150 survey locations. Predictive distribution models constructed using the programme PRESENCE revealed that while occupancy rates were highest in forest fragments, possums were still present across a large proportion of residential habitat with occupancy decreasing as housing density increased and green cover decreased. The presence of supplementary food sources was important in predicting possum occupancy, which may reflect the high nutritional value of these food types. Additionally, occupancy decreased as the proportion of forest fragment decreased, indicating the importance of forest fragments in determining possum distribution. Control operations to protect native birds from possum predation in

  12. Biological variables for the site survey of surface ecosystems - existing data and survey methods

    International Nuclear Information System (INIS)

    Kylaekorpi, Lasse; Berggren, Jens; Larsson, Mats; Liberg, Maria; Rydgren, Bernt

    2000-06-01

    In the process of selecting a safe and environmentally acceptable location for the deep level repository of nuclear waste, site surveys will be carried out. These site surveys will also include studies of the biota at the site, in order to assure that the chosen site will not conflict with important ecological interests, and to establish a thorough baseline for future impact assessments and monitoring programmes. As a preparation to the site survey programme, a review of the variables that need to be surveyed is conducted. This report contains the review for some of those variables. For each variable, existing data sources and their characteristics are listed. For those variables for which existing data sources are inadequate, suggestions are made for appropriate methods that will enable the establishment of an acceptable baseline. In this report the following variables are reviewed: Fishery, Landscape, Vegetation types, Key biotopes, Species (flora and fauna), Red-listed species (flora and fauna), Biomass (flora and fauna), Water level, water retention time (incl. water body and flow), Nutrients/toxins, Oxygen concentration, Layering, stratification, Light conditions/transparency, Temperature, Sediment transport, (Marine environments are excluded from this review). For a major part of the variables, the existing data coverage is most likely insufficient. Both the temporal and/or the geographical resolution is often limited, which means that complementary surveys must be performed during (or before) the site surveys. It is, however, in general difficult to make exact judgements on the extent of existing data, and also to give suggestions for relevant methods to use in the site surveys. This can be finally decided only when the locations for the sites are decided upon. The relevance of the different variables also depends on the environmental characteristics of the sites. Therefore, we suggest that when the survey sites are selected, an additional review is

  13. Biological variables for the site survey of surface ecosystems - existing data and survey methods

    Energy Technology Data Exchange (ETDEWEB)

    Kylaekorpi, Lasse; Berggren, Jens; Larsson, Mats; Liberg, Maria; Rydgren, Bernt [SwedPower AB, Stockholm (Sweden)

    2000-06-01

    In the process of selecting a safe and environmentally acceptable location for the deep level repository of nuclear waste, site surveys will be carried out. These site surveys will also include studies of the biota at the site, in order to assure that the chosen site will not conflict with important ecological interests, and to establish a thorough baseline for future impact assessments and monitoring programmes. As a preparation to the site survey programme, a review of the variables that need to be surveyed is conducted. This report contains the review for some of those variables. For each variable, existing data sources and their characteristics are listed. For those variables for which existing data sources are inadequate, suggestions are made for appropriate methods that will enable the establishment of an acceptable baseline. In this report the following variables are reviewed: Fishery, Landscape, Vegetation types, Key biotopes, Species (flora and fauna), Red-listed species (flora and fauna), Biomass (flora and fauna), Water level, water retention time (incl. water body and flow), Nutrients/toxins, Oxygen concentration, Layering, stratification, Light conditions/transparency, Temperature, Sediment transport, (Marine environments are excluded from this review). For a major part of the variables, the existing data coverage is most likely insufficient. Both the temporal and/or the geographical resolution is often limited, which means that complementary surveys must be performed during (or before) the site surveys. It is, however, in general difficult to make exact judgements on the extent of existing data, and also to give suggestions for relevant methods to use in the site surveys. This can be finally decided only when the locations for the sites are decided upon. The relevance of the different variables also depends on the environmental characteristics of the sites. Therefore, we suggest that when the survey sites are selected, an additional review is

  14. Slope stability and bearing capacity of landfills and simple on-site test methods.

    Science.gov (United States)

    Yamawaki, Atsushi; Doi, Yoichi; Omine, Kiyoshi

    2017-07-01

    This study discusses strength characteristics (slope stability, bearing capacity, etc.) of waste landfills through on-site tests that were carried out at 29 locations in 19 sites in Japan and three other countries, and proposes simple methods to test and assess the mechanical strength of landfills on site. Also, the possibility of using a landfill site was investigated by a full-scale eccentric loading test. As a result of this, landfills containing more than about 10 cm long plastics or other fibrous materials were found to be resilient and hard to yield. An on-site full scale test proved that no differential settlement occurs. The repose angle test proposed as a simple on-site test method has been confirmed to be a good indicator for slope stability assessment. The repose angle test suggested that landfills which have high, near-saturation water content have considerably poorer slope stability. The results of our repose angle test and the impact acceleration test were related to the internal friction angle and the cohesion, respectively. In addition to this, it was found that the air pore volume ratio measured by an on-site air pore volume ratio test is likely to be related to various strength parameters.

  15. Wind erosion in semiarid landscapes: Predictive models and remote sensing methods for the influence of vegetation

    Science.gov (United States)

    Musick, H. Brad

    1993-01-01

    The objectives of this research are: to develop and test predictive relations for the quantitative influence of vegetation canopy structure on wind erosion of semiarid rangeland soils, and to develop remote sensing methods for measuring the canopy structural parameters that determine sheltering against wind erosion. The influence of canopy structure on wind erosion will be investigated by means of wind-tunnel and field experiments using structural variables identified by the wind-tunnel and field experiments using model roughness elements to simulate plant canopies. The canopy structural variables identified by the wind-tunnel and field experiments as important in determining vegetative sheltering against wind erosion will then be measured at a number of naturally vegetated field sites and compared with estimates of these variables derived from analysis of remotely sensed data.

  16. A method for estimating the local area economic damages of Superfund waste sites

    International Nuclear Information System (INIS)

    Walker, D.R.

    1992-01-01

    National Priority List (NPL) sites, or more commonly called Superfund sites, are hazardous waste sites (HWS) deemed by the Environmental Protection Agency (EPA) to impose the greatest risks to human health or welfare or to the environment. HWS are placed and ranked for cleanup on the NPL based on a score derived from the Hazard Ranking System (HRS), which is a scientific assessment of the health and environmental risks posed by HWS. A concern of the HRS is that the rank of sites is not based on benefit-cost analysis. The main objective of this dissertation is to develop a method for estimating the local area economic damages associated with Superfund waste sites. Secondarily, the model is used to derive county-level damage estimates for use in ranking the county level damages from Superfund sites. The conceptual model used to describe the damages associated with Superfund sites is a household-firm location decision model. In this model assumes that households and firms make their location choice based on the local level of wages, rents and amenities. The model was empirically implemented using 1980 census microdata on households and workers in 253 counties across the US. The household sample includes data on the value and structural characteristics of homes. The worker sample includes the annual earnings of workers and a vector worker attributes. The microdata was combined with county level amenity data, including the number of Superfund sites. The hedonic pricing technique was used to estimate the effect of Superfund sites on average annual wages per household and on monthly expenditures on housing. The results show that Superfund sites impose statistically significant damages on households. The annual county damages from Superfund sites for a sample of 151 counties was over 14 billion dollars. The ranking of counties using the damage estimates is correlated with the rank of counties using the HRS

  17. Predicting Plasma Glucose From Interstitial Glucose Observations Using Bayesian Methods

    DEFF Research Database (Denmark)

    Hansen, Alexander Hildenbrand; Duun-Henriksen, Anne Katrine; Juhl, Rune

    2014-01-01

    One way of constructing a control algorithm for an artificial pancreas is to identify a model capable of predicting plasma glucose (PG) from interstitial glucose (IG) observations. Stochastic differential equations (SDEs) make it possible to account both for the unknown influence of the continuous...... glucose monitor (CGM) and for unknown physiological influences. Combined with prior knowledge about the measurement devices, this approach can be used to obtain a robust predictive model. A stochastic-differential-equation-based gray box (SDE-GB) model is formulated on the basis of an identifiable...

  18. A comparison of methods of predicting maximum oxygen uptake.

    OpenAIRE

    Grant, S; Corbett, K; Amjad, A M; Wilson, J; Aitchison, T

    1995-01-01

    The aim of this study was to compare the results from a Cooper walk run test, a multistage shuttle run test, and a submaximal cycle test with the direct measurement of maximum oxygen uptake on a treadmill. Three predictive tests of maximum oxygen uptake--linear extrapolation of heart rate of VO2 collected from a submaximal cycle ergometer test (predicted L/E), the Cooper 12 min walk, run test, and a multi-stage progressive shuttle run test (MST)--were performed by 22 young healthy males (mean...

  19. Structure prediction and binding sites analysis of curcin protein of Jatropha curcas using computational approaches.

    Science.gov (United States)

    Srivastava, Mugdha; Gupta, Shishir K; Abhilash, P C; Singh, Nandita

    2012-07-01

    Ribosome inactivating proteins (RIPs) are defense proteins in a number of higher-plant species that are directly targeted toward herbivores. Jatropha curcas is one of the biodiesel plants having RIPs. The Jatropha seed meal, after extraction of oil, is rich in curcin, a highly toxic RIP similar to ricin, which makes it unsuitable for animal feed. Although the toxicity of curcin is well documented in the literature, the detailed toxic properties and the 3D structure of curcin has not been determined by X-ray crystallography, NMR spectroscopy or any in silico techniques to date. In this pursuit, the structure of curcin was modeled by a composite approach of 3D structure prediction using threading and ab initio modeling. Assessment of model quality was assessed by methods which include Ramachandran plot analysis and Qmean score estimation. Further, we applied the protein-ligand docking approach to identify the r-RNA binding residue of curcin. The present work provides the first structural insight into the binding mode of r-RNA adenine to the curcin protein and forms the basis for designing future inhibitors of curcin. Cloning of a future peptide inhibitor within J. curcas can produce non-toxic varieties of J. curcas, which would make the seed-cake suitable as animal feed without curcin detoxification.

  20. What Predicts Method Effects in Child Behavior Ratings

    Science.gov (United States)

    Low, Justin A.; Keith, Timothy Z.; Jensen, Megan

    2015-01-01

    The purpose of this research was to determine whether child, parent, and teacher characteristics such as sex, socioeconomic status (SES), parental depressive symptoms, the number of years of teaching experience, number of children in the classroom, and teachers' disciplinary self-efficacy predict deviations from maternal ratings in a…

  1. A method for predicting the probability of business network profitability

    NARCIS (Netherlands)

    Johnson, P.; Iacob, Maria Eugenia; Välja, M.; van Sinderen, Marten J.; Magnusson, C; Ladhe, T.

    2014-01-01

    In the design phase of business collaboration, it is desirable to be able to predict the profitability of the business-to-be. Therefore, techniques to assess qualities such as costs, revenues, risks, and profitability have been previously proposed. However, they do not allow the modeler to properly

  2. Statistical tests for equal predictive ability across multiple forecasting methods

    DEFF Research Database (Denmark)

    Borup, Daniel; Thyrsgaard, Martin

    We develop a multivariate generalization of the Giacomini-White tests for equal conditional predictive ability. The tests are applicable to a mixture of nested and non-nested models, incorporate estimation uncertainty explicitly, and allow for misspecification of the forecasting model as well as ...

  3. Genomic breeding value prediction:methods and procedures

    NARCIS (Netherlands)

    Calus, M.P.L.

    2010-01-01

    Animal breeding faces one of the most significant changes of the past decades – the implementation of genomic selection. Genomic selection uses dense marker maps to predict the breeding value of animals with reported accuracies that are up to 0.31 higher than those of pedigree indexes, without the

  4. Link Prediction Methods and Their Accuracy for Different Social Networks and Network Metrics

    Directory of Open Access Journals (Sweden)

    Fei Gao

    2015-01-01

    Full Text Available Currently, we are experiencing a rapid growth of the number of social-based online systems. The availability of the vast amounts of data gathered in those systems brings new challenges that we face when trying to analyse it. One of the intensively researched topics is the prediction of social connections between users. Although a lot of effort has been made to develop new prediction approaches, the existing methods are not comprehensively analysed. In this paper we investigate the correlation between network metrics and accuracy of different prediction methods. We selected six time-stamped real-world social networks and ten most widely used link prediction methods. The results of the experiments show that the performance of some methods has a strong correlation with certain network metrics. We managed to distinguish “prediction friendly” networks, for which most of the prediction methods give good performance, as well as “prediction unfriendly” networks, for which most of the methods result in high prediction error. Correlation analysis between network metrics and prediction accuracy of prediction methods may form the basis of a metalearning system where based on network characteristics it will be able to recommend the right prediction method for a given network.

  5. Predicting redwood productivity using biophysical data, spatial statistics and site quality indices

    Science.gov (United States)

    John-Pascal Berrill; Kevin L. O’Hara; Shawn Headley

    2017-01-01

    Coast redwood (Sequoia sempervirens (D. Don) Endl.) height growth and basal area growth are sensitive to variations in site quality. Site factors known to be correlated with redwood stand growth and yield include topographic variables such as position on slope, exposure, and the composite variable: topographic relative moisture index. Species...

  6. Impact of statistical learning methods on the predictive power of multivariate normal tissue complication probability models

    NARCIS (Netherlands)

    Xu, Cheng-Jian; van der Schaaf, Arjen; Schilstra, Cornelis; Langendijk, Johannes A.; van t Veld, Aart A.

    2012-01-01

    PURPOSE: To study the impact of different statistical learning methods on the prediction performance of multivariate normal tissue complication probability (NTCP) models. METHODS AND MATERIALS: In this study, three learning methods, stepwise selection, least absolute shrinkage and selection operator

  7. Hybrid Prediction Method for Aircraft Interior Noise, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — The goal of the project is research and development of methods for application of the Hybrid FE-SEA method to aircraft vibro-acoustic problems. This proposal...

  8. DO TIE LABORATORY BASED ASSESSMENT METHODS REALLY PREDICT FIELD EFFECTS?

    Science.gov (United States)

    Sediment Toxicity Identification and Evaluation (TIE) methods have been developed for both porewaters and whole sediments. These relatively simple laboratory methods are designed to identify specific toxicants or classes of toxicants in sediments; however, the question of whethe...

  9. Prediction of Solvent Physical Properties using the Hierarchical Clustering Method

    Science.gov (United States)

    Recently a QSAR (Quantitative Structure Activity Relationship) method, the hierarchical clustering method, was developed to estimate acute toxicity values for large, diverse datasets. This methodology has now been applied to the estimate solvent physical properties including sur...

  10. An efficient method to transcription factor binding sites imputation via simultaneous completion of multiple matrices with positional consistency.

    Science.gov (United States)

    Guo, Wei-Li; Huang, De-Shuang

    2017-08-22

    Transcription factors (TFs) are DNA-binding proteins that have a central role in regulating gene expression. Identification of DNA-binding sites of TFs is a key task in understanding transcriptional regulation, cellular processes and disease. Chromatin immunoprecipitation followed by high-throughput sequencing (ChIP-seq) enables genome-wide identification of in vivo TF binding sites. However, it is still difficult to map every TF in every cell line owing to cost and biological material availability, which poses an enormous obstacle for integrated analysis of gene regulation. To address this problem, we propose a novel computational approach, TFBSImpute, for predicting additional TF binding profiles by leveraging information from available ChIP-seq TF binding data. TFBSImpute fuses the dataset to a 3-mode tensor and imputes missing TF binding signals via simultaneous completion of multiple TF binding matrices with positional consistency. We show that signals predicted by our method achieve overall similarity with experimental data and that TFBSImpute significantly outperforms baseline approaches, by assessing the performance of imputation methods against observed ChIP-seq TF binding profiles. Besides, motif analysis shows that TFBSImpute preforms better in capturing binding motifs enriched in observed data compared with baselines, indicating that the higher performance of TFBSImpute is not simply due to averaging related samples. We anticipate that our approach will constitute a useful complement to experimental mapping of TF binding, which is beneficial for further study of regulation mechanisms and disease.

  11. Comparison of Different Height–Diameter Modelling Techniques for Prediction of Site Productivity in Natural Uneven-Aged Pure Stands

    Directory of Open Access Journals (Sweden)

    Guangshuang Duan

    2018-01-01

    Full Text Available Reliable estimates of forest site productivity are a central element of forest management. The model of height-diameter relationship of dominant trees using algebraic difference approach (ADA is a commonly used method to measure site productivity of natural uneven-aged stands. However, the existing models of this method do not recognize site type or sample plot specific variability in height curves; thus, it cannot be effectively used to estimate site type or sample plot-related site productivity for natural uneven-aged stands. Two primary subject-specific approaches, ADA with dummy variable (DV (ADA + DV and ADA with combination of dummy variable and nonlinear mixed-effects modelling (CM (ADA + CM, were proposed for height–diameter modelling. Height–diameter models developed with ADA, ADA + DV and ADA + CM were compared using data from 4161 observations on 349 permanent sample plots of four major natural uneven-aged pure stands (Spruce, Korean Larch, Mongolian Oak, and White Birch in northeastern China. It was found that models developed with ADA + CM provided the best performance, followed by the models with ADA + DV, and the models developed with ADA performed the worst. Random effects at the plot level were substantial, and their inclusion greatly improved the model’s accuracy. More importantly, the models developed with ADA + CM provide an effective method for quantifying site type- and sample plot-specific forest site productivity for uneven-aged pure stands.

  12. Predicting volume of distribution with decision tree-based regression methods using predicted tissue:plasma partition coefficients.

    Science.gov (United States)

    Freitas, Alex A; Limbu, Kriti; Ghafourian, Taravat

    2015-01-01

    Volume of distribution is an important pharmacokinetic property that indicates the extent of a drug's distribution in the body tissues. This paper addresses the problem of how to estimate the apparent volume of distribution at steady state (Vss) of chemical compounds in the human body using decision tree-based regression methods from the area of data mining (or machine learning). Hence, the pros and cons of several different types of decision tree-based regression methods have been discussed. The regression methods predict Vss using, as predictive features, both the compounds' molecular descriptors and the compounds' tissue:plasma partition coefficients (Kt:p) - often used in physiologically-based pharmacokinetics. Therefore, this work has assessed whether the data mining-based prediction of Vss can be made more accurate by using as input not only the compounds' molecular descriptors but also (a subset of) their predicted Kt:p values. Comparison of the models that used only molecular descriptors, in particular, the Bagging decision tree (mean fold error of 2.33), with those employing predicted Kt:p values in addition to the molecular descriptors, such as the Bagging decision tree using adipose Kt:p (mean fold error of 2.29), indicated that the use of predicted Kt:p values as descriptors may be beneficial for accurate prediction of Vss using decision trees if prior feature selection is applied. Decision tree based models presented in this work have an accuracy that is reasonable and similar to the accuracy of reported Vss inter-species extrapolations in the literature. The estimation of Vss for new compounds in drug discovery will benefit from methods that are able to integrate large and varied sources of data and flexible non-linear data mining methods such as decision trees, which can produce interpretable models. Graphical AbstractDecision trees for the prediction of tissue partition coefficient and volume of distribution of drugs.

  13. Prediction of Human Drug Targets and Their Interactions Using Machine Learning Methods: Current and Future Perspectives.

    Science.gov (United States)

    Nath, Abhigyan; Kumari, Priyanka; Chaube, Radha

    2018-01-01

    Identification of drug targets and drug target interactions are important steps in the drug-discovery pipeline. Successful computational prediction methods can reduce the cost and time demanded by the experimental methods. Knowledge of putative drug targets and their interactions can be very useful for drug repurposing. Supervised machine learning methods have been very useful in drug target prediction and in prediction of drug target interactions. Here, we describe the details for developing prediction models using supervised learning techniques for human drug target prediction and their interactions.

  14. A predictive estimation method for carbon dioxide transport by data-driven modeling with a physically-based data model.

    Science.gov (United States)

    Jeong, Jina; Park, Eungyu; Han, Weon Shik; Kim, Kue-Young; Jun, Seong-Chun; Choung, Sungwook; Yun, Seong-Taek; Oh, Junho; Kim, Hyun-Jun

    2017-11-01

    In this study, a data-driven method for predicting CO 2 leaks and associated concentrations from geological CO 2 sequestration is developed. Several candidate models are compared based on their reproducibility and predictive capability for CO 2 concentration measurements from the Environment Impact Evaluation Test (EIT) site in Korea. Based on the data mining results, a one-dimensional solution of the advective-dispersive equation for steady flow (i.e., Ogata-Banks solution) is found to be most representative for the test data, and this model is adopted as the data model for the developed method. In the validation step, the method is applied to estimate future CO 2 concentrations with the reference estimation by the Ogata-Banks solution, where a part of earlier data is used as the training dataset. From the analysis, it is found that the ensemble mean of multiple estimations based on the developed method shows high prediction accuracy relative to the reference estimation. In addition, the majority of the data to be predicted are included in the proposed quantile interval, which suggests adequate representation of the uncertainty by the developed method. Therefore, the incorporation of a reasonable physically-based data model enhances the prediction capability of the data-driven model. The proposed method is not confined to estimations of CO 2 concentration and may be applied to various real-time monitoring data from subsurface sites to develop automated control, management or decision-making systems. Copyright © 2017 Elsevier B.V. All rights reserved.

  15. A method to determine site-specific, anisotropic fracture toughness in biological materials

    International Nuclear Information System (INIS)

    Bechtle, Sabine; Özcoban, Hüseyin; Yilmaz, Ezgi D.; Fett, Theo; Rizzi, Gabriele; Lilleodden, Erica T.; Huber, Norbert; Schreyer, Andreas; Swain, Michael V.; Schneider, Gerold A.

    2012-01-01

    Many biological materials are hierarchically structured, with highly anisotropic structures and properties on several length scales. To characterize the mechanical properties of such materials, detailed testing methods are required that allow precise and site-specific measurements on several length scales. We propose a fracture toughness measurement technique based on notched focused ion beam prepared cantilevers of lower and medium micron size scales. Using this approach, site-specific fracture toughness values in dental enamel were determined. The usefulness and challenges of the method are discussed.

  16. Theoretical prediction method of subcooled flow boiling CHF

    Energy Technology Data Exchange (ETDEWEB)

    Kwon, Young Min; Chang, Soon Heung [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1999-12-31

    A theoretical critical heat flux (CHF ) model, based on lateral bubble coalescence on the heated wall, is proposed to predict the subcooled flow boiling CHF in a uniformly heated vertical tube. The model is based on the concept that a single layer of bubbles contacted to the heated wall prevents a bulk liquid from reaching the wall at near CHF condition. Comparisons between the model predictions and experimental data result in satisfactory agreement within less than 9.73% root-mean-square error by the appropriate choice of the critical void fraction in the bubbly layer. The present model shows comparable performance with the CHF look-up table of Groeneveld et al.. 28 refs., 11 figs., 1 tab. (Author)

  17. Machine learning methods in predicting the student academic motivation

    Directory of Open Access Journals (Sweden)

    Ivana Đurđević Babić

    2017-01-01

    Full Text Available Academic motivation is closely related to academic performance. For educators, it is equally important to detect early students with a lack of academic motivation as it is to detect those with a high level of academic motivation. In endeavouring to develop a classification model for predicting student academic motivation based on their behaviour in learning management system (LMS courses, this paper intends to establish links between the predicted student academic motivation and their behaviour in the LMS course. Students from all years at the Faculty of Education in Osijek participated in this research. Three machine learning classifiers (neural networks, decision trees, and support vector machines were used. To establish whether a significant difference in the performance of models exists, a t-test of the difference in proportions was used. Although, all classifiers were successful, the neural network model was shown to be the most successful in detecting the student academic motivation based on their behaviour in LMS course.

  18. Theoretical prediction method of subcooled flow boiling CHF

    Energy Technology Data Exchange (ETDEWEB)

    Kwon, Young Min; Chang, Soon Heung [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1998-12-31

    A theoretical critical heat flux (CHF ) model, based on lateral bubble coalescence on the heated wall, is proposed to predict the subcooled flow boiling CHF in a uniformly heated vertical tube. The model is based on the concept that a single layer of bubbles contacted to the heated wall prevents a bulk liquid from reaching the wall at near CHF condition. Comparisons between the model predictions and experimental data result in satisfactory agreement within less than 9.73% root-mean-square error by the appropriate choice of the critical void fraction in the bubbly layer. The present model shows comparable performance with the CHF look-up table of Groeneveld et al.. 28 refs., 11 figs., 1 tab. (Author)

  19. Improved Methods for Pitch Synchronous Linear Prediction Analysis of Speech

    OpenAIRE

    劉, 麗清

    2015-01-01

    Linear prediction (LP) analysis has been applied to speech system over the last few decades. LP technique is well-suited for speech analysis due to its ability to model speech production process approximately. Hence LP analysis has been widely used for speech enhancement, low-bit-rate speech coding in cellular telephony, speech recognition, characteristic parameter extraction (vocal tract resonances frequencies, fundamental frequency called pitch) and so on. However, the performance of the co...

  20. A Mechanism-Based Model for the Prediction of the Metabolic Sites of Steroids Mediated by Cytochrome P450 3A4

    Directory of Open Access Journals (Sweden)

    Zi-Ru Dai

    2015-06-01

    Full Text Available Early prediction of xenobiotic metabolism is essential for drug discovery and development. As the most important human drug-metabolizing enzyme, cytochrome P450 3A4 has a large active cavity and metabolizes a broad spectrum of substrates. The poor substrate specificity of CYP3A4 makes it a huge challenge to predict the metabolic site(s on its substrates. This study aimed to develop a mechanism-based prediction model based on two key parameters, including the binding conformation and the reaction activity of ligands, which could reveal the process of real metabolic reaction(s and the site(s of modification. The newly established model was applied to predict the metabolic site(s of steroids; a class of CYP3A4-preferred substrates. 38 steroids and 12 non-steroids were randomly divided into training and test sets. Two major metabolic reactions, including aliphatic hydroxylation and N-dealkylation, were involved in this study. At least one of the top three predicted metabolic sites was validated by the experimental data. The overall accuracy for the training and test were 82.14% and 86.36%, respectively. In summary, a mechanism-based prediction model was established for the first time, which could be used to predict the metabolic site(s of CYP3A4 on steroids with high predictive accuracy.

  1. HuMiTar: A sequence-based method for prediction of human microRNA targets

    Directory of Open Access Journals (Sweden)

    Chen Ke

    2008-12-01

    Full Text Available Abstract Background MicroRNAs (miRs are small noncoding RNAs that bind to complementary/partially complementary sites in the 3' untranslated regions of target genes to regulate protein production of the target transcript and to induce mRNA degradation or mRNA cleavage. The ability to perform accurate, high-throughput identification of physiologically active miR targets would enable functional characterization of individual miRs. Current target prediction methods include traditional approaches that are based on specific base-pairing rules in the miR's seed region and implementation of cross-species conservation of the target site, and machine learning (ML methods that explore patterns that contrast true and false miR-mRNA duplexes. However, in the case of the traditional methods research shows that some seed region matches that are conserved are false positives and that some of the experimentally validated target sites are not conserved. Results We present HuMiTar, a computational method for identifying common targets of miRs, which is based on a scoring function that considers base-pairing for both seed and non-seed positions for human miR-mRNA duplexes. Our design shows that certain non-seed miR nucleotides, such as 14, 18, 13, 11, and 17, are characterized by a strong bias towards formation of Watson-Crick pairing. We contrasted HuMiTar with several representative competing methods on two sets of human miR targets and a set of ten glioblastoma oncogenes. Comparison with the two best performing traditional methods, PicTar and TargetScanS, and a representative ML method that considers the non-seed positions, NBmiRTar, shows that HuMiTar predictions include majority of the predictions of the other three methods. At the same time, the proposed method is also capable of finding more true positive targets as a trade-off for an increased number of predictions. Genome-wide predictions show that the proposed method is characterized by 1.99 signal

  2. pLoc-mAnimal: predict subcellular localization of animal proteins with both single and multiple sites.

    Science.gov (United States)

    Cheng, Xiang; Zhao, Shu-Guang; Lin, Wei-Zhong; Xiao, Xuan; Chou, Kuo-Chen

    2017-11-15

    Cells are deemed the basic unit of life. However, many important functions of cells as well as their growth and reproduction are performed via the protein molecules located at their different organelles or locations. Facing explosive growth of protein sequences, we are challenged to develop fast and effective method to annotate their subcellular localization. However, this is by no means an easy task. Particularly, mounting evidences have indicated proteins have multi-label feature meaning that they may simultaneously exist at, or move between, two or more different subcellular location sites. Unfortunately, most of the existing computational methods can only be used to deal with the single-label proteins. Although the 'iLoc-Animal' predictor developed recently is quite powerful that can be used to deal with the animal proteins with multiple locations as well, its prediction quality needs to be improved, particularly in enhancing the absolute true rate and reducing the absolute false rate. Here we propose a new predictor called 'pLoc-mAnimal', which is superior to iLoc-Animal as shown by the compelling facts. When tested by the most rigorous cross-validation on the same high-quality benchmark dataset, the absolute true success rate achieved by the new predictor is 37% higher and the absolute false rate is four times lower in comparison with the state-of-the-art predictor. To maximize the convenience of most experimental scientists, a user-friendly web-server for the new predictor has been established at http://www.jci-bioinfo.cn/pLoc-mAnimal/, by which users can easily get their desired results without the need to go through the complicated mathematics involved. xxiao@gordonlifescience.org or kcchou@gordonlifescience.org. Supplementary data are available at Bioinformatics online. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  3. Analysis of Multi-Criteria Evaluation Method of Landfill Site Selection for Municipal Solid Waste Management

    Science.gov (United States)

    Mohammed, Habiba Ibrahim; Majid, Zulkepli; Yusof, Norhakim Bin; Bello Yamusa, Yamusa

    2018-03-01

    Landfilling remains the most common systematic technique of solid waste disposal in most of the developed and developing countries. Finding a suitable site for landfill is a very challenging task. Landfill site selection process aims to provide suitable areas that will protect the environment and public health from pollution and hazards. Therefore, various factors such as environmental, physical, socio-economic, and geological criteria must be considered before siting any landfill. This makes the site selection process vigorous and tedious because it involves the processing of large amount of spatial data, rules and regulations from different agencies and also policy from decision makers. This allows the incorporation of conflicting objectives and decision maker preferences into spatial decision models. This paper particularly analyzes the multi-criteria evaluation (MCE) method of landfill site selection for solid waste management by means of literature reviews and surveys. The study will help the decision makers and waste management authorities to choose the most effective method when considering landfill site selection.

  4. A method for the automated detection phishing websites through both site characteristics and image analysis

    Science.gov (United States)

    White, Joshua S.; Matthews, Jeanna N.; Stacy, John L.

    2012-06-01

    Phishing website analysis is largely still a time-consuming manual process of discovering potential phishing sites, verifying if suspicious sites truly are malicious spoofs and if so, distributing their URLs to the appropriate blacklisting services. Attackers increasingly use sophisticated systems for bringing phishing sites up and down rapidly at new locations, making automated response essential. In this paper, we present a method for rapid, automated detection and analysis of phishing websites. Our method relies on near real-time gathering and analysis of URLs posted on social media sites. We fetch the pages pointed to by each URL and characterize each page with a set of easily computed values such as number of images and links. We also capture a screen-shot of the rendered page image, compute a hash of the image and use the Hamming distance between these image hashes as a form of visual comparison. We provide initial results demonstrate the feasibility of our techniques by comparing legitimate sites to known fraudulent versions from Phishtank.com, by actively introducing a series of minor changes to a phishing toolkit captured in a local honeypot and by performing some initial analysis on a set of over 2.8 million URLs posted to Twitter over a 4 days in August 2011. We discuss the issues encountered during our testing such as resolvability and legitimacy of URL's posted on Twitter, the data sets used, the characteristics of the phishing sites we discovered, and our plans for future work.

  5. Analysis of Multi-Criteria Evaluation Method of Landfill Site Selection for Municipal Solid Waste Management

    Directory of Open Access Journals (Sweden)

    Ibrahim Mohammed Habiba

    2018-01-01

    Full Text Available Landfilling remains the most common systematic technique of solid waste disposal in most of the developed and developing countries. Finding a suitable site for landfill is a very challenging task. Landfill site selection process aims to provide suitable areas that will protect the environment and public health from pollution and hazards. Therefore, various factors such as environmental, physical, socio-economic, and geological criteria must be considered before siting any landfill. This makes the site selection process vigorous and tedious because it involves the processing of large amount of spatial data, rules and regulations from different agencies and also policy from decision makers. This allows the incorporation of conflicting objectives and decision maker preferences into spatial decision models. This paper particularly analyzes the multi-criteria evaluation (MCE method of landfill site selection for solid waste management by means of literature reviews and surveys. The study will help the decision makers and waste management authorities to choose the most effective method when considering landfill site selection.

  6. Development of an integrated method for long-term water quality prediction using seasonal climate forecast

    Directory of Open Access Journals (Sweden)

    J. Cho

    2016-10-01

    Full Text Available The APEC Climate Center (APCC produces climate prediction information utilizing a multi-climate model ensemble (MME technique. In this study, four different downscaling methods, in accordance with the degree of utilizing the seasonal climate prediction information, were developed in order to improve predictability and to refine the spatial scale. These methods include: (1 the Simple Bias Correction (SBC method, which directly uses APCC's dynamic prediction data with a 3 to 6 month lead time; (2 the Moving Window Regression (MWR method, which indirectly utilizes dynamic prediction data; (3 the Climate Index Regression (CIR method, which predominantly uses observation-based climate indices; and (4 the Integrated Time Regression (ITR method, which uses predictors selected from both CIR and MWR. Then, a sampling-based temporal downscaling was conducted using the Mahalanobis distance method in order to create daily weather inputs to the Soil and Water Assessment Tool (SWAT model. Long-term predictability of water quality within the Wecheon watershed of the Nakdong River Basin was evaluated. According to the Korean Ministry of Environment's Provisions of Water Quality Prediction and Response Measures, modeling-based predictability was evaluated by using 3-month lead prediction data issued in February, May, August, and November as model input of SWAT. Finally, an integrated approach, which takes into account various climate information and downscaling methods for water quality prediction, was presented. This integrated approach can be used to prevent potential problems caused by extreme climate in advance.

  7. Testing contamination risk assessment methods for toxic elements from mine waste sites

    Science.gov (United States)

    Abdaal, A.; Jordan, G.; Szilassi, P.; Kiss, J.; Detzky, G.

    2012-04-01

    Major incidents involving mine waste facilities and poor environmental management practices have left a legacy of thousands of contaminated sites like in the historic mining areas in the Carpathian Basin. Associated environmental risks have triggered the development of new EU environmental legislation to prevent and minimize the effects of such incidents. The Mine Waste Directive requires the risk-based inventory of all mine waste sites in Europe by May 2012. In order to address the mining problems a standard risk-based Pre-selection protocol has been developed by the EU Commission. This paper discusses the heavy metal contamination in acid mine drainage (AMD) for risk assessment (RA) along the Source-Pathway-Receptor chain using decision support methods which are intended to aid national and regional organizations in the inventory and assessment of potentially contaminated mine waste sites. Several recognized methods such as the European Environmental Agency (EEA) standard PRAMS model for soil contamination, US EPA-based AIMSS and Irish HMS-IRC models for RA of abandoned sites are reviewed, compared and tested for the mining waste environment. In total 145 ore mine waste sites have been selected for scientific testing using the EU Pre-selection protocol as a case study from Hungary. The proportion of uncertain to certain responses for a site and for the total number of sites may give an insight of specific and overall uncertainty in the data we use. The Pre-selection questions are efficiently linked to a GIS system as database inquiries using digital spatial data to directly generate answers. Key parameters such as distance to the nearest surface and ground water bodies, to settlements and protected areas are calculated and statistically evaluated using STATGRAPHICS® in order to calibrate the RA models. According to our scientific research results, of the 145 sites 11 sites are the most risky having foundation slope >20o, 57 sites are within distance 66 (class VI

  8. Pb and Sr isotopic compositions of ancient pottery: a method to discriminate production sites

    International Nuclear Information System (INIS)

    Zhang Xun; Chen Jiangfeng; Ma Lin; He Jianfeng; Wang Changsui; Qiu Ping

    2004-01-01

    The discriminating of production sites of ancient pottery samples using multi-isotopic systematics was described. Previous work has proven that Pb isotopic ratios can be used for discriminating the production sites of ancient pottery under certain conditions. The present work suggests that although Nd isotopic ratios are not sensitive to the production sites of ancient pottery, Sr isotopic ratios are important for the purpose. Pb isotopic ratios are indistinguishable for the pottery excavated from the Jiahu relict, Wuyang, Henan Province and for famous Qin Terra-cotta Figures. But, the 87 Sr/ 86 Sr ratios for the former (about 0.715) are significantly lower than that of the latter (0.717-0.718). The authors concluded that a combined use of Pb and Sr isotopes would be a more powerful method for discriminating the production site of ancient pottery. (authors)

  9. Documentation of archaeological sites in northern iraq using remote sensing methods

    Science.gov (United States)

    Matoušková, E.; Pavelka, K.; Nováček, K.; Starková, L.

    2015-08-01

    The MULINEM (The Medieval Urban Landscape in Northeastern Mesopotamia) project is aiming to investigate a Late Sasanian and Islamic urban network in the land of Erbil, historic province of Hidyab (Adiabene) that is located in the northern Iraq. The research of the hierarchical urban network in a defined area belongs to approaches rarely used in the study of the Islamic urbanism. The project focuses on the cluster of urban sites of the 6th-17th centuries A.D. This paper focuses on remote sensing analysis of historical sites with special interest of FORMOSAT-2 data that have been gained through a research announcement: Free FORMOSAT-2 satellite Imagery. Documentation of two archaeological sites (Makhmúr al-Qadima and Kushaf) are introduced. FORMOSAT-2 data results have been compared to historic CORONA satellite data of mentioned historical sites purchased earlier by the University of West Bohemia. Remote sensing methods were completed using in-situ measurements.

  10. iSNO-PseAAC: predict cysteine S-nitrosylation sites in proteins by incorporating position specific amino acid propensity into pseudo amino acid composition.

    Directory of Open Access Journals (Sweden)

    Yan Xu

    Full Text Available Posttranslational modifications (PTMs of proteins are responsible for sensing and transducing signals to regulate various cellular functions and signaling events. S-nitrosylation (SNO is one of the most important and universal PTMs. With the avalanche of protein sequences generated in the post-genomic age, it is highly desired to develop computational methods for timely identifying the exact SNO sites in proteins because this kind of information is very useful for both basic research and drug development. Here, a new predictor, called iSNO-PseAAC, was developed for identifying the SNO sites in proteins by incorporating the position-specific amino acid propensity (PSAAP into the general form of pseudo amino acid composition (PseAAC. The predictor was implemented using the conditional random field (CRF algorithm. As a demonstration, a benchmark dataset was constructed that contains 731 SNO sites and 810 non-SNO sites. To reduce the homology bias, none of these sites were derived from the proteins that had [Formula: see text] pairwise sequence identity to any other. It was observed that the overall cross-validation success rate achieved by iSNO-PseAAC in identifying nitrosylated proteins on an independent dataset was over 90%, indicating that the new predictor is quite promising. Furthermore, a user-friendly web-server for iSNO-PseAAC was established at http://app.aporc.org/iSNO-PseAAC/, by which users can easily obtain the desired results without the need to follow the mathematical equations involved during the process of developing the prediction method. It is anticipated that iSNO-PseAAC may become a useful high throughput tool for identifying the SNO sites, or at the very least play a complementary role to the existing methods in this area.

  11. RELIABILITY AND ACCURACY ASSESSMENT OF INVASIVE AND NON- INVASIVE SEISMIC METHODS FOR SITE CHARACTERIZATION: FEEDBACK FROM THE INTERPACIFIC PROJECT

    OpenAIRE

    Garofalo , F.; Foti , S.; Hollender , F.; Bard , P.-Y.; Cornou , C.; Cox , B.R.; Dechamp , A.; Ohrnberger , M.; Sicilia , D.; Vergniault , C.

    2017-01-01

    International audience; The InterPacific project (Intercomparison of methods for site parameter and velocity profile characterization) aims to assess the reliability of seismic site characterization methods (borehole and surface wave methods) used for estimating shear wave velocity (VS) profiles and other related parameters (e.g., VS30). Three sites, representative of different geological conditions relevant for the evaluation of seismic site response effects, have been selected: (1) a hard r...

  12. Simple methods for predicting gas leakage flows through cracks

    International Nuclear Information System (INIS)

    Ewing, D.J.F.

    1989-01-01

    This report presents closed-form approximate analytical formulae with which the flow rate out of a through-wall crack can be estimated. The crack is idealised as a rough, tapering, wedgeshaped channel and the fluid is idealised as an isothermal or polytropically-expanding perfect gas. In practice, uncertainties about the wall friction factor dominate over uncertainties caused by the fluid-dynamics simplifications. The formulae take account of crack taper and for outwardly-diverging cracks they predict flows within 12% of mathematically more accurate one-dimensional numerical models. Upper and lower estimates of wall friction are discussed. (author)

  13. Underwater Sound Propagation Modeling Methods for Predicting Marine Animal Exposure.

    Science.gov (United States)

    Hamm, Craig A; McCammon, Diana F; Taillefer, Martin L

    2016-01-01

    The offshore exploration and production (E&P) industry requires comprehensive and accurate ocean acoustic models for determining the exposure of marine life to the high levels of sound used in seismic surveys and other E&P activities. This paper reviews the types of acoustic models most useful for predicting the propagation of undersea noise sources and describes current exposure models. The severe problems caused by model sensitivity to the uncertainty in the environment are highlighted to support the conclusion that it is vital that risk assessments include transmission loss estimates with statistical measures of confidence.

  14. The calculation of site-dependent earthquake motions -3. The method of fast fourier transform

    International Nuclear Information System (INIS)

    Simpson, I.C.

    1976-10-01

    The method of Fast Fourier transform (FFT) is applied to the problem of the determination of site-dependent earthquake motions, which takes account of local geological effects. A program, VELAY 1, which uses the FFT method has been written and is described in this report. The assumptions of horizontally stratified, homogeneous, isotropic, linearly viscoelastic layers and a normally incident plane seismic wave are made. Several examples are given, using VELAY 1, of modified surface acceleration-time histories obtained using a selected input acceleration-time history and a representative system of soil layers. There is a discussion concerning the soil properties that need to be measured in order to use VELAY 1 (and similar programs described in previous reports) and hence generate site-dependent ground motions suitable for aseismic design of a nuclear power plant at a given site. (author)

  15. A method of risk assessment for a multi-plant site

    International Nuclear Information System (INIS)

    White, R.F.

    1983-06-01

    A model is presented which can be used in conjunction with probabilistic risk assessment to estimate whether a site on which there are several plants (reactors or chemical plants containing radioactive materials) meets whatever risk acceptance criteria or numerical risk guidelines are applied at the time of the assessment in relation to various groups of people and for various sources of risk. The application of the multi-plant site model to the direct and inverse methods of risk assessment is described. A method is proposed by which the potential hazard rating associated with a given plant can be quantified so that an appropriate allocation can be made when assessing the risks associated with each of the plants on a site. (author)

  16. TEMPERATURE PREDICTION IN 3013 CONTAINERS IN K AREA MATERIAL STORAGE (KAMS) FACILITY USING REGRESSION METHODS

    International Nuclear Information System (INIS)

    Gupta, N

    2008-01-01

    3013 containers are designed in accordance with the DOE-STD-3013-2004. These containers are qualified to store plutonium (Pu) bearing materials such as PuO2 for 50 years. DOT shipping packages such as the 9975 are used to store the 3013 containers in the K-Area Material Storage (KAMS) facility at Savannah River Site (SRS). DOE-STD-3013-2004 requires that a comprehensive surveillance program be set up to ensure that the 3013 container design parameters are not violated during the long term storage. To ensure structural integrity of the 3013 containers, thermal analyses using finite element models were performed to predict the contents and component temperatures for different but well defined parameters such as storage ambient temperature, PuO 2 density, fill heights, weights, and thermal loading. Interpolation is normally used to calculate temperatures if the actual parameter values are different from the analyzed values. A statistical analysis technique using regression methods is proposed to develop simple polynomial relations to predict temperatures for the actual parameter values found in the containers. The analysis shows that regression analysis is a powerful tool to develop simple relations to assess component temperatures

  17. Specification and prediction of nickel mobilization using artificial intelligence methods

    Science.gov (United States)

    Gholami, Raoof; Ziaii, Mansour; Ardejani, Faramarz Doulati; Maleki, Shahoo

    2011-12-01

    Groundwater and soil pollution from pyrite oxidation, acid mine drainage generation, and release and transport of toxic metals are common environmental problems associated with the mining industry. Nickel is one toxic metal considered to be a key pollutant in some mining setting; to date, its formation mechanism has not yet been fully evaluated. The goals of this study are 1) to describe the process of nickel mobilization in waste dumps by introducing a novel conceptual model, and 2) to predict nickel concentration using two algorithms, namely the support vector machine (SVM) and the general regression neural network (GRNN). The results obtained from this study have shown that considerable amount of nickel concentration can be arrived into the water flow system during the oxidation of pyrite and subsequent Acid Drainage (AMD) generation. It was concluded that pyrite, water, and oxygen are the most important factors for nickel pollution generation while pH condition, SO4, HCO3, TDS, EC, Mg, Fe, Zn, and Cu are measured quantities playing significant role in nickel mobilization. SVM and GRNN have predicted nickel concentration with a high degree of accuracy. Hence, SVM and GRNN can be considered as appropriate tools for environmental risk assessment.

  18. Verifying a computational method for predicting extreme ground motion

    Science.gov (United States)

    Harris, R.A.; Barall, M.; Andrews, D.J.; Duan, B.; Ma, S.; Dunham, E.M.; Gabriel, A.-A.; Kaneko, Y.; Kase, Y.; Aagaard, Brad T.; Oglesby, D.D.; Ampuero, J.-P.; Hanks, T.C.; Abrahamson, N.

    2011-01-01

    In situations where seismological data is rare or nonexistent, computer simulations may be used to predict ground motions caused by future earthquakes. This is particularly practical in the case of extreme ground motions, where engineers of special buildings may need to design for an event that has not been historically observed but which may occur in the far-distant future. Once the simulations have been performed, however, they still need to be tested. The SCEC-USGS dynamic rupture code verification exercise provides a testing mechanism for simulations that involve spontaneous earthquake rupture. We have performed this examination for the specific computer code that was used to predict maximum possible ground motion near Yucca Mountain. Our SCEC-USGS group exercises have demonstrated that the specific computer code that was used for the Yucca Mountain simulations produces similar results to those produced by other computer codes when tackling the same science problem. We also found that the 3D ground motion simulations produced smaller ground motions than the 2D simulations.

  19. Test methods for on site measurement of resistivity of concrete. A RILEM TC-154 Technical Recommendation

    NARCIS (Netherlands)

    Polder, R.B.

    2000-01-01

    This paper describes methods to assess concrete resistivity on site for various purposes related to corrosion and protection of reinforcement. It is based on a first draft of a RILEM Technical Recommendation. The electrical resistivity of concrete can be related to the two processes involved in

  20. X-ray and synchrotron methods in studies of cultural heritage sites

    Energy Technology Data Exchange (ETDEWEB)

    Koval’chuk, M. V.; Yatsishina, E. B.; Blagov, A. E.; Tereshchenko, E. Yu., E-mail: elenatereschenko@yandex.ru; Prosekov, P. A.; Dyakova, Yu. A. [National Research Centre “Kurchatov Institute” (Russian Federation)

    2016-09-15

    X-ray and synchrotron methods that are most widely used in studies of cultural heritage objects (including archaeological sites)—X-ray diffraction analysis, X-ray spectroscopy, and visualization techniques— have been considered. The reported examples show high efficiency and informativeness of natural science studies when solving most diverse problems of archaeology, history, the study of art, museology, etc.

  1. From nuclear installation to greenfield site. SCK-CEN develops a new measurement method

    International Nuclear Information System (INIS)

    2014-01-01

    The article discusses a new measurement method that has been developed by the Belgian Nuclear Research Center SCK-CEN in conjunction with the decommissioning of nuclear facilities. This measurement technique is based on on-site gamma ray spectrometry in combination with modelling and is employed for directing the flow of demolition materials in the decommissioning of nuclear facilities.

  2. Veteran Teacher Engagement in Site-Based Professional Development: A Mixed Methods Study

    Science.gov (United States)

    Houston, Biaze L.

    2016-01-01

    This research study examined how teachers self-report their levels of engagement, which factors they believe contribute most to their engagement, and which assumptions of andragogy most heavily influence teacher engagement in site-based professional development. This study employed a convergent parallel mixed methods design to study veteran…

  3. Assessment procedure and probability determination methods of aircraft crash events in siting for nuclear power plants

    International Nuclear Information System (INIS)

    Zheng Qiyan; Zhang Lijun; Huang Weiqi; Yin Qingliao

    2010-01-01

    Assessment procedure of aircraft crash events in siting for nuclear power plants, and the methods of probability determination in two different stages of prelimi- nary screening and detailed evaluation are introduced in this paper. Except for general air traffic, airport operations and aircraft in the corridor, the probability of aircraft crash by military operation in the military airspaces is considered here. (authors)

  4. X-ray and synchrotron methods in studies of cultural heritage sites

    International Nuclear Information System (INIS)

    Koval’chuk, M. V.; Yatsishina, E. B.; Blagov, A. E.; Tereshchenko, E. Yu.; Prosekov, P. A.; Dyakova, Yu. A.

    2016-01-01

    X-ray and synchrotron methods that are most widely used in studies of cultural heritage objects (including archaeological sites)—X-ray diffraction analysis, X-ray spectroscopy, and visualization techniques— have been considered. The reported examples show high efficiency and informativeness of natural science studies when solving most diverse problems of archaeology, history, the study of art, museology, etc.

  5. Using deuterated PAH amendments to validate chemical extraction methods to predict PAH bioavailability in soils

    International Nuclear Information System (INIS)

    Gomez-Eyles, Jose L.; Collins, Chris D.; Hodson, Mark E.

    2011-01-01

    Validating chemical methods to predict bioavailable fractions of polycyclic aromatic hydrocarbons (PAHs) by comparison with accumulation bioassays is problematic. Concentrations accumulated in soil organisms not only depend on the bioavailable fraction but also on contaminant properties. A historically contaminated soil was freshly spiked with deuterated PAHs (dPAHs). dPAHs have a similar fate to their respective undeuterated analogues, so chemical methods that give good indications of bioavailability should extract the fresh more readily available dPAHs and historic more recalcitrant PAHs in similar proportions to those in which they are accumulated in the tissues of test organisms. Cyclodextrin and butanol extractions predicted the bioavailable fraction for earthworms (Eisenia fetida) and plants (Lolium multiflorum) better than the exhaustive extraction. The PAHs accumulated by earthworms had a larger dPAH:PAH ratio than that predicted by chemical methods. The isotope ratio method described here provides an effective way of evaluating other chemical methods to predict bioavailability. - Research highlights: → Isotope ratios can be used to evaluate chemical methods to predict bioavailability. → Chemical methods predicted bioavailability better than exhaustive extractions. → Bioavailability to earthworms was still far from that predicted by chemical methods. - A novel method using isotope ratios to assess the ability of chemical methods to predict PAH bioavailability to soil biota.

  6. Using deuterated PAH amendments to validate chemical extraction methods to predict PAH bioavailability in soils

    Energy Technology Data Exchange (ETDEWEB)

    Gomez-Eyles, Jose L., E-mail: j.l.gomezeyles@reading.ac.uk [University of Reading, School of Human and Environmental Sciences, Soil Research Centre, Reading, RG6 6DW Berkshire (United Kingdom); Collins, Chris D.; Hodson, Mark E. [University of Reading, School of Human and Environmental Sciences, Soil Research Centre, Reading, RG6 6DW Berkshire (United Kingdom)

    2011-04-15

    Validating chemical methods to predict bioavailable fractions of polycyclic aromatic hydrocarbons (PAHs) by comparison with accumulation bioassays is problematic. Concentrations accumulated in soil organisms not only depend on the bioavailable fraction but also on contaminant properties. A historically contaminated soil was freshly spiked with deuterated PAHs (dPAHs). dPAHs have a similar fate to their respective undeuterated analogues, so chemical methods that give good indications of bioavailability should extract the fresh more readily available dPAHs and historic more recalcitrant PAHs in similar proportions to those in which they are accumulated in the tissues of test organisms. Cyclodextrin and butanol extractions predicted the bioavailable fraction for earthworms (Eisenia fetida) and plants (Lolium multiflorum) better than the exhaustive extraction. The PAHs accumulated by earthworms had a larger dPAH:PAH ratio than that predicted by chemical methods. The isotope ratio method described here provides an effective way of evaluating other chemical methods to predict bioavailability. - Research highlights: > Isotope ratios can be used to evaluate chemical methods to predict bioavailability. > Chemical methods predicted bioavailability better than exhaustive extractions. > Bioavailability to earthworms was still far from that predicted by chemical methods. - A novel method using isotope ratios to assess the ability of chemical methods to predict PAH bioavailability to soil biota.

  7. Evaluation of new location of Isfahan′s sanitary landfill site with Oleckno method

    OpenAIRE

    Maryam Salimi; Afshin Ebrahimi; Afsane Salimi

    2013-01-01

    Aims: The objective of present study was to evaluate the new location of Isfahan solid waste sanitary landfill using Geographical Information System (GIS) based on the Oleckno index method (OIM). Materials and Methods: This study was on the field- and library-based data collection and surveys of relevant data. Assessment parameters included average annual rainfall, soil type and ground water beneath and adjucent to the landfill site. To analyze data, ArcGIS version 9.3 was used. Resul...

  8. Method to predict process signals to learn for SVM

    International Nuclear Information System (INIS)

    Minowa, Hirotsugu; Gofuku, Akio

    2013-01-01

    Study of diagnostic system using machine learning to reduce the incidents of the plant is in advance because an accident causes large damage about human, economic and social loss. There is a problem that 2 performances between a classification performance and generalization performance on the machine diagnostic machine is exclusive. However, multi agent diagnostic system makes it possible to use a diagnostic machine specialized either performance by multi diagnostic machines can be used. We propose method to select optimized variables to improve classification performance. The method can also be used for other supervised learning machine but Support Vector Machine. This paper reports that our method and result of evaluation experiment applied our method to output 40% of Monju. (author)

  9. Kinetic mesh-free method for flutter prediction in turbomachines

    Indian Academy of Sciences (India)

    -based mesh-free method for unsteady flows. ... Council for Scientific and Industrial Research, National Aerospace Laboratories, Computational and Theoretical Fluid Dynamics Division, Bangalore 560 017, India; Engineering Mechanics Unit, ...

  10. Cargo flows distribution over the loading sites of enterprises by using methods of artificial intelligence

    Directory of Open Access Journals (Sweden)

    Олександр Павлович Кіркін

    2017-06-01

    Full Text Available Development of information technologies and market requirements in effective control over cargo flows, forces enterprises to look for new ways and methods of automated control over the technological operations. For rail transportation one of the most complicated tasks of automation is the cargo flows distribution over the sites of loading and unloading. In this article the solution with the use of one of the methods of artificial intelligence – a fuzzy inference has been proposed. The analysis of the last publications showed that the fuzzy inference method is effective for the solution of similar tasks, it makes it possible to accumulate experience, it is stable to temporary impacts of the environmental conditions. The existing methods of the cargo flows distribution over the sites of loading and unloading are too simplified and can lead to incorrect decisions. The purpose of the article is to create a distribution model of cargo flows of the enterprises over the sites of loading and unloading, basing on the fuzzy inference method and to automate the control. To achieve the objective a mathematical model of the cargo flows distribution over the sites of loading and unloading has been made using fuzzy logic. The key input parameters of the model are: «number of loading sites», «arrival of the next set of cars», «availability of additional operations». The output parameter is «a variety of set of cars». Application of the fuzzy inference method made it possible to reduce loading time by 15% and to reduce costs for preparatory operations before loading by 20%. Thus this method is an effective means and holds the greatest promise for railway competitiveness increase. Interaction between different types of transportation and their influence on the cargo flows distribution over the sites of loading and unloading hasn’t been considered. These sites may be busy transshipping at that very time which is characteristic of large enterprises

  11. Forecasting method for global radiation time series without training phase: Comparison with other well-known prediction methodologies

    International Nuclear Information System (INIS)

    Voyant, Cyril; Motte, Fabrice; Fouilloy, Alexis; Notton, Gilles; Paoli, Christophe; Nivet, Marie-Laure

    2017-01-01

    Integration of unpredictable renewable energy sources into electrical networks intensifies the complexity of the grid management due to their intermittent and unforeseeable nature. Because of the strong increase of solar power generation the prediction of solar yields becomes more and more important. Electrical operators need an estimation of the future production. For nowcasting and short term forecasting, the usual technics based on machine learning need large historical data sets of good quality during the training phase of predictors. However data are not always available and induce an advanced maintenance of meteorological stations, making the method inapplicable for poor instrumented or isolated sites. In this work, we propose intuitive methodologies based on the Kalman filter use (also known as linear quadratic estimation), able to predict a global radiation time series without the need of historical data. The accuracy of these methods is compared to other classical data driven methods, for different horizons of prediction and time steps. The proposed approach shows interesting capabilities allowing to improve quasi-systematically the prediction. For one to 10 h horizons Kalman model performances are competitive in comparison to more sophisticated models such as ANN which require both consistent historical data sets and computational resources. - Highlights: • Solar radiation forecasting with time series formalism. • Trainless approach compared to machine learning methods. • Very simple method dedicated to solar irradiation forecasting with high accuracy.

  12. SABER: a computational method for identifying active sites for new reactions.

    Science.gov (United States)

    Nosrati, Geoffrey R; Houk, K N

    2012-05-01

    A software suite, SABER (Selection of Active/Binding sites for Enzyme Redesign), has been developed for the analysis of atomic geometries in protein structures, using a geometric hashing algorithm (Barker and Thornton, Bioinformatics 2003;19:1644-1649). SABER is used to explore the Protein Data Bank (PDB) to locate proteins with a specific 3D arrangement of catalytic groups to identify active sites that might be redesigned to catalyze new reactions. As a proof-of-principle test, SABER was used to identify enzymes that have the same catalytic group arrangement present in o-succinyl benzoate synthase (OSBS). Among the highest-scoring scaffolds identified by the SABER search for enzymes with the same catalytic group arrangement as OSBS were L-Ala D/L-Glu epimerase (AEE) and muconate lactonizing enzyme II (MLE), both of which have been redesigned to become effective OSBS catalysts, demonstrated by experiments. Next, we used SABER to search for naturally existing active sites in the PDB with catalytic groups similar to those present in the designed Kemp elimination enzyme KE07. From over 2000 geometric matches to the KE07 active site, SABER identified 23 matches that corresponded to residues from known active sites. The best of these matches, with a 0.28 Å catalytic atom RMSD to KE07, was then redesigned to be compatible with the Kemp elimination using RosettaDesign. We also used SABER to search for potential Kemp eliminases using a theozyme predicted to provide a greater rate acceleration than the active site of KE07, and used Rosetta to create a design based on the proteins identified. Copyright © 2012 The Protein Society.

  13. Seismic PSA method for multiple nuclear power plants in a site

    Energy Technology Data Exchange (ETDEWEB)

    Hakata, Tadakuni [Nuclear Safety Commission, Tokyo (Japan)

    2007-07-15

    The maximum number of nuclear power plants in a site is eight and about 50% of power plants are built in sites with three or more plants in the world. Such nuclear sites have potential risks of simultaneous multiple plant damages especially at external events. Seismic probabilistic safety assessment method (Level-1 PSA) for multi-unit sites with up to 9 units has been developed. The models include Fault-tree linked Monte Carlo computation, taking into consideration multivariate correlations of components and systems from partial to complete, inside and across units. The models were programmed as a computer program CORAL reef. Sample analysis and sensitivity studies were performed to verify the models and algorithms and to understand some of risk insights and risk metrics, such as site core damage frequency (CDF per site-year) for multiple reactor plants. This study will contribute to realistic state of art seismic PSA, taking consideration of multiple reactor power plants, and to enhancement of seismic safety. (author)

  14. A simplified method for active-site titration of lipases immobilised on hydrophobic supports.

    Science.gov (United States)

    Nalder, Tim D; Kurtovic, Ivan; Barrow, Colin J; Marshall, Susan N

    2018-06-01

    The aim of this work was to develop a simple and accurate protocol to measure the functional active site concentration of lipases immobilised on highly hydrophobic supports. We used the potent lipase inhibitor methyl 4-methylumbelliferyl hexylphosphonate to titrate the active sites of Candida rugosa lipase (CrL) bound to three highly hydrophobic supports: octadecyl methacrylate (C18), divinylbenzene crosslinked methacrylate (DVB) and styrene. The method uses correction curves to take into account the binding of the fluorophore (4-methylumbelliferone, 4-MU) by the support materials. We showed that the uptake of the detection agent by the three supports is not linear relative to the weight of the resin, and that the uptake occurs in an equilibrium that is independent of the total fluorophore concentration. Furthermore, the percentage of bound fluorophore varied among the supports, with 50 mg of C18 and styrene resins binding approximately 64 and 94%, respectively. When the uptake of 4-MU was calculated and corrected for, the total 4-MU released via inhibition (i.e. the concentration of functional lipase active sites) could be determined via a linear relationship between immobilised lipase weight and total inhibition. It was found that the functional active site concentration of immobilised CrL varied greatly among different hydrophobic supports, with 56% for C18, compared with 14% for DVB. The described method is a simple and robust approach to measuring functional active site concentration in immobilised lipase samples. Copyright © 2018 Elsevier Inc. All rights reserved.

  15. Evaluation of Two Statistical Methods Provides Insights into the Complex Patterns of Alternative Polyadenylation Site Switching

    Science.gov (United States)

    Li, Jie; Li, Rui; You, Leiming; Xu, Anlong; Fu, Yonggui; Huang, Shengfeng

    2015-01-01

    Switching between different alternative polyadenylation (APA) sites plays an important role in the fine tuning of gene expression. New technologies for the execution of 3’-end enriched RNA-seq allow genome-wide detection of the genes that exhibit significant APA site switching between different samples. Here, we show that the independence test gives better results than the linear trend test in detecting APA site-switching events. Further examination suggests that the discrepancy between these two statistical methods arises from complex APA site-switching events that cannot be represented by a simple change of average 3’-UTR length. In theory, the linear trend test is only effective in detecting these simple changes. We classify the switching events into four switching patterns: two simple patterns (3’-UTR shortening and lengthening) and two complex patterns. By comparing the results of the two statistical methods, we show that complex patterns account for 1/4 of all observed switching events that happen between normal and cancerous human breast cell lines. Because simple and complex switching patterns may convey different biological meanings, they merit separate study. We therefore propose to combine both the independence test and the linear trend test in practice. First, the independence test should be used to detect APA site switching; second, the linear trend test should be invoked to identify simple switching events; and third, those complex switching events that pass independence testing but fail linear trend testing can be identified. PMID:25875641