WorldWideScience

Sample records for database curation lessons

  1. Research Data Curation Pilots: Lessons Learned

    Directory of Open Access Journals (Sweden)

    David Minor

    2014-07-01

    Full Text Available In the spring of 2011, the UC San Diego Research Cyberinfrastructure (RCI Implementation Team invited researchers and research teams to participate in a research curation and data management pilot program. This invitation took the form of a campus-wide solicitation. More than two dozen applications were received and, after due deliberation, the RCI Oversight Committee selected five curation-intensive projects. These projects were chosen based on a number of criteria, including how they represented campus research, varieties of topics, researcher engagement, and the various services required. The pilot process began in September 2011, and will be completed in early 2014. Extensive lessons learned from the pilots are being compiled and are being used in the on-going design and implementation of the permanent Research Data Curation Program in the UC San Diego Library. In this paper, we present specific implementation details of these various services, as well as lessons learned. The program focused on many aspects of contemporary scholarship, including data creation and storage, description and metadata creation, citation and publication, and long term preservation and access. Based on the lessons learned in our processes, the Research Data Curation Program will provide a suite of services from which campus users can pick and choose, as necessary. The program will provide support for the data management requirements from national funding agencies.

  2. DAMPD: A manually curated antimicrobial peptide database

    KAUST Repository

    Seshadri Sundararajan, Vijayaraghava

    2011-11-21

    The demand for antimicrobial peptides (AMPs) is rising because of the increased occurrence of pathogens that are tolerant or resistant to conventional antibiotics. Since naturally occurring AMPs could serve as templates for the development of new anti-infectious agents to which pathogens are not resistant, a resource that contains relevant information on AMP is of great interest. To that extent, we developed the Dragon Antimicrobial Peptide Database (DAMPD, http://apps.sanbi.ac.za/dampd) that contains 1232 manually curated AMPs. DAMPD is an update and a replacement of the ANTIMIC database. In DAMPD an integrated interface allows in a simple fashion querying based on taxonomy, species, AMP family, citation, keywords and a combination of search terms and fields (Advanced Search). A number of tools such as Blast, ClustalW, HMMER, Hydrocalculator, SignalP, AMP predictor, as well as a number of other resources that provide additional information about the results are also provided and integrated into DAMPD to augment biological analysis of AMPs. The Author(s) 2011. Published by Oxford University Press.

  3. DAMPD: A manually curated antimicrobial peptide database

    KAUST Repository

    Seshadri Sundararajan, Vijayaraghava; Gabere, Musa Nur; Pretorius, Ashley; Adam, Saleem; Christoffels, Alan; Lehvaslaiho, Minna; Archer, John A.C.; Bajic, Vladimir B.

    2011-01-01

    The demand for antimicrobial peptides (AMPs) is rising because of the increased occurrence of pathogens that are tolerant or resistant to conventional antibiotics. Since naturally occurring AMPs could serve as templates for the development of new anti-infectious agents to which pathogens are not resistant, a resource that contains relevant information on AMP is of great interest. To that extent, we developed the Dragon Antimicrobial Peptide Database (DAMPD, http://apps.sanbi.ac.za/dampd) that contains 1232 manually curated AMPs. DAMPD is an update and a replacement of the ANTIMIC database. In DAMPD an integrated interface allows in a simple fashion querying based on taxonomy, species, AMP family, citation, keywords and a combination of search terms and fields (Advanced Search). A number of tools such as Blast, ClustalW, HMMER, Hydrocalculator, SignalP, AMP predictor, as well as a number of other resources that provide additional information about the results are also provided and integrated into DAMPD to augment biological analysis of AMPs. The Author(s) 2011. Published by Oxford University Press.

  4. EMU Lessons Learned Database

    Science.gov (United States)

    Matthews, Kevin M., Jr.; Crocker, Lori; Cupples, J. Scott

    2011-01-01

    As manned space exploration takes on the task of traveling beyond low Earth orbit, many problems arise that must be solved in order to make the journey possible. One major task is protecting humans from the harsh space environment. The current method of protecting astronauts during Extravehicular Activity (EVA) is through use of the specially designed Extravehicular Mobility Unit (EMU). As more rigorous EVA conditions need to be endured at new destinations, the suit will need to be tailored and improved in order to accommodate the astronaut. The Objective behind the EMU Lessons Learned Database(LLD) is to be able to create a tool which will assist in the development of next-generation EMUs, along with maintenance and improvement of the current EMU, by compiling data from Failure Investigation and Analysis Reports (FIARs) which have information on past suit failures. FIARs use a system of codes that give more information on the aspects of the failure, but if one is unfamiliar with the EMU they will be unable to decipher the information. A goal of the EMU LLD is to not only compile the information, but to present it in a user-friendly, organized, searchable database accessible to all familiarity levels with the EMU; both newcomers and veterans alike. The EMU LLD originally started as an Excel database, which allowed easy navigation and analysis of the data through pivot charts. Creating an entry requires access to the Problem Reporting And Corrective Action database (PRACA), which contains the original FIAR data for all hardware. FIAR data are then transferred to, defined, and formatted in the LLD. Work is being done to create a web-based version of the LLD in order to increase accessibility to all of Johnson Space Center (JSC), which includes converting entries from Excel to the HTML format. FIARs related to the EMU have been completed in the Excel version, and now focus has shifted to expanding FIAR data in the LLD to include EVA tools and support hardware such as

  5. CCDB: a curated database of genes involved in cervix cancer.

    Science.gov (United States)

    Agarwal, Subhash M; Raghav, Dhwani; Singh, Harinder; Raghava, G P S

    2011-01-01

    The Cervical Cancer gene DataBase (CCDB, http://crdd.osdd.net/raghava/ccdb) is a manually curated catalog of experimentally validated genes that are thought, or are known to be involved in the different stages of cervical carcinogenesis. In spite of the large women population that is presently affected from this malignancy still at present, no database exists that catalogs information on genes associated with cervical cancer. Therefore, we have compiled 537 genes in CCDB that are linked with cervical cancer causation processes such as methylation, gene amplification, mutation, polymorphism and change in expression level, as evident from published literature. Each record contains details related to gene like architecture (exon-intron structure), location, function, sequences (mRNA/CDS/protein), ontology, interacting partners, homology to other eukaryotic genomes, structure and links to other public databases, thus augmenting CCDB with external data. Also, manually curated literature references have been provided to support the inclusion of the gene in the database and establish its association with cervix cancer. In addition, CCDB provides information on microRNA altered in cervical cancer as well as search facility for querying, several browse options and an online tool for sequence similarity search, thereby providing researchers with easy access to the latest information on genes involved in cervix cancer.

  6. MIPS: curated databases and comprehensive secondary data resources in 2010.

    Science.gov (United States)

    Mewes, H Werner; Ruepp, Andreas; Theis, Fabian; Rattei, Thomas; Walter, Mathias; Frishman, Dmitrij; Suhre, Karsten; Spannagl, Manuel; Mayer, Klaus F X; Stümpflen, Volker; Antonov, Alexey

    2011-01-01

    The Munich Information Center for Protein Sequences (MIPS at the Helmholtz Center for Environmental Health, Neuherberg, Germany) has many years of experience in providing annotated collections of biological data. Selected data sets of high relevance, such as model genomes, are subjected to careful manual curation, while the bulk of high-throughput data is annotated by automatic means. High-quality reference resources developed in the past and still actively maintained include Saccharomyces cerevisiae, Neurospora crassa and Arabidopsis thaliana genome databases as well as several protein interaction data sets (MPACT, MPPI and CORUM). More recent projects are PhenomiR, the database on microRNA-related phenotypes, and MIPS PlantsDB for integrative and comparative plant genome research. The interlinked resources SIMAP and PEDANT provide homology relationships as well as up-to-date and consistent annotation for 38,000,000 protein sequences. PPLIPS and CCancer are versatile tools for proteomics and functional genomics interfacing to a database of compilations from gene lists extracted from literature. A novel literature-mining tool, EXCERBT, gives access to structured information on classified relations between genes, proteins, phenotypes and diseases extracted from Medline abstracts by semantic analysis. All databases described here, as well as the detailed descriptions of our projects can be accessed through the MIPS WWW server (http://mips.helmholtz-muenchen.de).

  7. Integration of curated databases to identify genotype-phenotype associations

    Directory of Open Access Journals (Sweden)

    Li Jianrong

    2006-10-01

    Full Text Available Abstract Background The ability to rapidly characterize an unknown microorganism is critical in both responding to infectious disease and biodefense. To do this, we need some way of anticipating an organism's phenotype based on the molecules encoded by its genome. However, the link between molecular composition (i.e. genotype and phenotype for microbes is not obvious. While there have been several studies that address this challenge, none have yet proposed a large-scale method integrating curated biological information. Here we utilize a systematic approach to discover genotype-phenotype associations that combines phenotypic information from a biomedical informatics database, GIDEON, with the molecular information contained in National Center for Biotechnology Information's Clusters of Orthologous Groups database (NCBI COGs. Results Integrating the information in the two databases, we are able to correlate the presence or absence of a given protein in a microbe with its phenotype as measured by certain morphological characteristics or survival in a particular growth media. With a 0.8 correlation score threshold, 66% of the associations found were confirmed by the literature and at a 0.9 correlation threshold, 86% were positively verified. Conclusion Our results suggest possible phenotypic manifestations for proteins biochemically associated with sugar metabolism and electron transport. Moreover, we believe our approach can be extended to linking pathogenic phenotypes with functionally related proteins.

  8. The curation paradigm and application tool used for manual curation of the scientific literature at the Comparative Toxicogenomics Database

    Science.gov (United States)

    Davis, Allan Peter; Wiegers, Thomas C.; Murphy, Cynthia G.; Mattingly, Carolyn J.

    2011-01-01

    The Comparative Toxicogenomics Database (CTD) is a public resource that promotes understanding about the effects of environmental chemicals on human health. CTD biocurators read the scientific literature and convert free-text information into a structured format using official nomenclature, integrating third party controlled vocabularies for chemicals, genes, diseases and organisms, and a novel controlled vocabulary for molecular interactions. Manual curation produces a robust, richly annotated dataset of highly accurate and detailed information. Currently, CTD describes over 349 000 molecular interactions between 6800 chemicals, 20 900 genes (for 330 organisms) and 4300 diseases that have been manually curated from over 25 400 peer-reviewed articles. This manually curated data are further integrated with other third party data (e.g. Gene Ontology, KEGG and Reactome annotations) to generate a wealth of toxicogenomic relationships. Here, we describe our approach to manual curation that uses a powerful and efficient paradigm involving mnemonic codes. This strategy allows biocurators to quickly capture detailed information from articles by generating simple statements using codes to represent the relationships between data types. The paradigm is versatile, expandable, and able to accommodate new data challenges that arise. We have incorporated this strategy into a web-based curation tool to further increase efficiency and productivity, implement quality control in real-time and accommodate biocurators working remotely. Database URL: http://ctd.mdibl.org PMID:21933848

  9. The art and science of data curation: Lessons learned from constructing a virtual collection

    Science.gov (United States)

    Bugbee, Kaylin; Ramachandran, Rahul; Maskey, Manil; Gatlin, Patrick

    2018-03-01

    A digital, or virtual, collection is a value added service developed by libraries that curates information and resources around a topic, theme or organization. Adoption of the virtual collection concept as an Earth science data service improves the discoverability, accessibility and usability of data both within individual data centers but also across data centers and disciplines. In this paper, we introduce a methodology for systematically and rigorously curating Earth science data and information into a cohesive virtual collection. This methodology builds on the geocuration model of searching, selecting and synthesizing Earth science data, metadata and other information into a single and useful collection. We present our experiences curating a virtual collection for one of NASA's twelve Distributed Active Archive Centers (DAACs), the Global Hydrology Resource Center (GHRC), and describe lessons learned as a result of this curation effort. We also provide recommendations and best practices for data centers and data providers who wish to curate virtual collections for the Earth sciences.

  10. TreeFam: a curated database of phylogenetic trees of animal gene families

    DEFF Research Database (Denmark)

    Li, Heng; Coghlan, Avril; Ruan, Jue

    2006-01-01

    TreeFam is a database of phylogenetic trees of gene families found in animals. It aims to develop a curated resource that presents the accurate evolutionary history of all animal gene families, as well as reliable ortholog and paralog assignments. Curated families are being added progressively......, based on seed alignments and trees in a similar fashion to Pfam. Release 1.1 of TreeFam contains curated trees for 690 families and automatically generated trees for another 11 646 families. These represent over 128 000 genes from nine fully sequenced animal genomes and over 45 000 other animal proteins...

  11. A curated database of cyanobacterial strains relevant for modern taxonomy and phylogenetic studies

    OpenAIRE

    Ramos, Vitor; Morais, Jo?o; Vasconcelos, Vitor M.

    2017-01-01

    The dataset herein described lays the groundwork for an online database of relevant cyanobacterial strains, named CyanoType (http://lege.ciimar.up.pt/cyanotype). It is a database that includes categorized cyanobacterial strains useful for taxonomic, phylogenetic or genomic purposes, with associated information obtained by means of a literature-based curation. The dataset lists 371 strains and represents the first version of the database (CyanoType v.1). Information for each strain includes st...

  12. Text mining facilitates database curation - extraction of mutation-disease associations from Bio-medical literature.

    Science.gov (United States)

    Ravikumar, Komandur Elayavilli; Wagholikar, Kavishwar B; Li, Dingcheng; Kocher, Jean-Pierre; Liu, Hongfang

    2015-06-06

    Advances in the next generation sequencing technology has accelerated the pace of individualized medicine (IM), which aims to incorporate genetic/genomic information into medicine. One immediate need in interpreting sequencing data is the assembly of information about genetic variants and their corresponding associations with other entities (e.g., diseases or medications). Even with dedicated effort to capture such information in biological databases, much of this information remains 'locked' in the unstructured text of biomedical publications. There is a substantial lag between the publication and the subsequent abstraction of such information into databases. Multiple text mining systems have been developed, but most of them focus on the sentence level association extraction with performance evaluation based on gold standard text annotations specifically prepared for text mining systems. We developed and evaluated a text mining system, MutD, which extracts protein mutation-disease associations from MEDLINE abstracts by incorporating discourse level analysis, using a benchmark data set extracted from curated database records. MutD achieves an F-measure of 64.3% for reconstructing protein mutation disease associations in curated database records. Discourse level analysis component of MutD contributed to a gain of more than 10% in F-measure when compared against the sentence level association extraction. Our error analysis indicates that 23 of the 64 precision errors are true associations that were not captured by database curators and 68 of the 113 recall errors are caused by the absence of associated disease entities in the abstract. After adjusting for the defects in the curated database, the revised F-measure of MutD in association detection reaches 81.5%. Our quantitative analysis reveals that MutD can effectively extract protein mutation disease associations when benchmarking based on curated database records. The analysis also demonstrates that incorporating

  13. HSC-explorer: a curated database for hematopoietic stem cells.

    Science.gov (United States)

    Montrone, Corinna; Kokkaliaris, Konstantinos D; Loeffler, Dirk; Lechner, Martin; Kastenmüller, Gabi; Schroeder, Timm; Ruepp, Andreas

    2013-01-01

    HSC-Explorer (http://mips.helmholtz-muenchen.de/HSC/) is a publicly available, integrative database containing detailed information about the early steps of hematopoiesis. The resource aims at providing fast and easy access to relevant information, in particular to the complex network of interacting cell types and molecules, from the wealth of publications in the field through visualization interfaces. It provides structured information on more than 7000 experimentally validated interactions between molecules, bioprocesses and environmental factors. Information is manually derived by critical reading of the scientific literature from expert annotators. Hematopoiesis-relevant interactions are accompanied with context information such as model organisms and experimental methods for enabling assessment of reliability and relevance of experimental results. Usage of established vocabularies facilitates downstream bioinformatics applications and to convert the results into complex networks. Several predefined datasets (Selected topics) offer insights into stem cell behavior, the stem cell niche and signaling processes supporting hematopoietic stem cell maintenance. HSC-Explorer provides a versatile web-based resource for scientists entering the field of hematopoiesis enabling users to inspect the associated biological processes through interactive graphical presentation.

  14. HSC-explorer: a curated database for hematopoietic stem cells.

    Directory of Open Access Journals (Sweden)

    Corinna Montrone

    Full Text Available HSC-Explorer (http://mips.helmholtz-muenchen.de/HSC/ is a publicly available, integrative database containing detailed information about the early steps of hematopoiesis. The resource aims at providing fast and easy access to relevant information, in particular to the complex network of interacting cell types and molecules, from the wealth of publications in the field through visualization interfaces. It provides structured information on more than 7000 experimentally validated interactions between molecules, bioprocesses and environmental factors. Information is manually derived by critical reading of the scientific literature from expert annotators. Hematopoiesis-relevant interactions are accompanied with context information such as model organisms and experimental methods for enabling assessment of reliability and relevance of experimental results. Usage of established vocabularies facilitates downstream bioinformatics applications and to convert the results into complex networks. Several predefined datasets (Selected topics offer insights into stem cell behavior, the stem cell niche and signaling processes supporting hematopoietic stem cell maintenance. HSC-Explorer provides a versatile web-based resource for scientists entering the field of hematopoiesis enabling users to inspect the associated biological processes through interactive graphical presentation.

  15. The SIB Swiss Institute of Bioinformatics' resources: focus on curated databases

    OpenAIRE

    Bultet, Lisandra Aguilar; Aguilar Rodriguez, Jose; Ahrens, Christian H; Ahrne, Erik Lennart; Ai, Ni; Aimo, Lucila; Akalin, Altuna; Aleksiev, Tyanko; Alocci, Davide; Altenhoff, Adrian; Alves, Isabel; Ambrosini, Giovanna; Pedone, Pascale Anderle; Angelina, Paolo; Anisimova, Maria

    2016-01-01

    The SIB Swiss Institute of Bioinformatics (www.isb-sib.ch) provides world-class bioinformatics databases, software tools, services and training to the international life science community in academia and industry. These solutions allow life scientists to turn the exponentially growing amount of data into knowledge. Here, we provide an overview of SIB's resources and competence areas, with a strong focus on curated databases and SIB's most popular and widely used resources. In particular, SIB'...

  16. XML technology planning database : lessons learned

    Science.gov (United States)

    Some, Raphael R.; Neff, Jon M.

    2005-01-01

    A hierarchical Extensible Markup Language(XML) database called XCALIBR (XML Analysis LIBRary) has been developed by Millennium Program to assist in technology investment (ROI) analysis and technology Language Capability the New return on portfolio optimization. The database contains mission requirements and technology capabilities, which are related by use of an XML dictionary. The XML dictionary codifies a standardized taxonomy for space missions, systems, subsystems and technologies. In addition to being used for ROI analysis, the database is being examined for use in project planning, tracking and documentation. During the past year, the database has moved from development into alpha testing. This paper describes the lessons learned during construction and testing of the prototype database and the motivation for moving from an XML taxonomy to a standard XML-based ontology.

  17. The MIntAct project--IntAct as a common curation platform for 11 molecular interaction databases

    OpenAIRE

    Orchard, S; Ammari, M; Aranda, B; Breuza, L; Briganti, L; Broackes-Carter, F; Campbell, N; Chavali, G; Chen, C; del-Toro, N; Duesbury, M; Dumousseau, M; Galeota, E; Hinz, U; Iannuccelli, M

    2014-01-01

    IntAct (freely available at http://www.ebi.ac.uk/intact) is an open-source, open data molecular interaction database populated by data either curated from the literature or from direct data depositions. IntAct has developed a sophisticated web-based curation tool, capable of supporting both IMEx- and MIMIx-level curation. This tool is now utilized by multiple additional curation teams, all of whom annotate data directly into the IntAct database. Members of the IntAct team supply appropriate l...

  18. CPAD, Curated Protein Aggregation Database: A Repository of Manually Curated Experimental Data on Protein and Peptide Aggregation.

    Science.gov (United States)

    Thangakani, A Mary; Nagarajan, R; Kumar, Sandeep; Sakthivel, R; Velmurugan, D; Gromiha, M Michael

    2016-01-01

    Accurate distinction between peptide sequences that can form amyloid-fibrils or amorphous β-aggregates, identification of potential aggregation prone regions in proteins, and prediction of change in aggregation rate of a protein upon mutation(s) are critical to research on protein misfolding diseases, such as Alzheimer's and Parkinson's, as well as biotechnological production of protein based therapeutics. We have developed a Curated Protein Aggregation Database (CPAD), which has collected results from experimental studies performed by scientific community aimed at understanding protein/peptide aggregation. CPAD contains more than 2300 experimentally observed aggregation rates upon mutations in known amyloidogenic proteins. Each entry includes numerical values for the following parameters: change in rate of aggregation as measured by fluorescence intensity or turbidity, name and source of the protein, Uniprot and Protein Data Bank codes, single point as well as multiple mutations, and literature citation. The data in CPAD has been supplemented with five different types of additional information: (i) Amyloid fibril forming hexa-peptides, (ii) Amorphous β-aggregating hexa-peptides, (iii) Amyloid fibril forming peptides of different lengths, (iv) Amyloid fibril forming hexa-peptides whose crystal structures are available in the Protein Data Bank (PDB) and (v) Experimentally validated aggregation prone regions found in amyloidogenic proteins. Furthermore, CPAD is linked to other related databases and resources, such as Uniprot, Protein Data Bank, PUBMED, GAP, TANGO, WALTZ etc. We have set up a web interface with different search and display options so that users have the ability to get the data in multiple ways. CPAD is freely available at http://www.iitm.ac.in/bioinfo/CPAD/. The potential applications of CPAD have also been discussed.

  19. Estimating the annotation error rate of curated GO database sequence annotations

    Directory of Open Access Journals (Sweden)

    Brown Alfred L

    2007-05-01

    Full Text Available Abstract Background Annotations that describe the function of sequences are enormously important to researchers during laboratory investigations and when making computational inferences. However, there has been little investigation into the data quality of sequence function annotations. Here we have developed a new method of estimating the error rate of curated sequence annotations, and applied this to the Gene Ontology (GO sequence database (GOSeqLite. This method involved artificially adding errors to sequence annotations at known rates, and used regression to model the impact on the precision of annotations based on BLAST matched sequences. Results We estimated the error rate of curated GO sequence annotations in the GOSeqLite database (March 2006 at between 28% and 30%. Annotations made without use of sequence similarity based methods (non-ISS had an estimated error rate of between 13% and 18%. Annotations made with the use of sequence similarity methodology (ISS had an estimated error rate of 49%. Conclusion While the overall error rate is reasonably low, it would be prudent to treat all ISS annotations with caution. Electronic annotators that use ISS annotations as the basis of predictions are likely to have higher false prediction rates, and for this reason designers of these systems should consider avoiding ISS annotations where possible. Electronic annotators that use ISS annotations to make predictions should be viewed sceptically. We recommend that curators thoroughly review ISS annotations before accepting them as valid. Overall, users of curated sequence annotations from the GO database should feel assured that they are using a comparatively high quality source of information.

  20. The MIntAct project—IntAct as a common curation platform for 11 molecular interaction databases

    Science.gov (United States)

    Orchard, Sandra; Ammari, Mais; Aranda, Bruno; Breuza, Lionel; Briganti, Leonardo; Broackes-Carter, Fiona; Campbell, Nancy H.; Chavali, Gayatri; Chen, Carol; del-Toro, Noemi; Duesbury, Margaret; Dumousseau, Marine; Galeota, Eugenia; Hinz, Ursula; Iannuccelli, Marta; Jagannathan, Sruthi; Jimenez, Rafael; Khadake, Jyoti; Lagreid, Astrid; Licata, Luana; Lovering, Ruth C.; Meldal, Birgit; Melidoni, Anna N.; Milagros, Mila; Peluso, Daniele; Perfetto, Livia; Porras, Pablo; Raghunath, Arathi; Ricard-Blum, Sylvie; Roechert, Bernd; Stutz, Andre; Tognolli, Michael; van Roey, Kim; Cesareni, Gianni; Hermjakob, Henning

    2014-01-01

    IntAct (freely available at http://www.ebi.ac.uk/intact) is an open-source, open data molecular interaction database populated by data either curated from the literature or from direct data depositions. IntAct has developed a sophisticated web-based curation tool, capable of supporting both IMEx- and MIMIx-level curation. This tool is now utilized by multiple additional curation teams, all of whom annotate data directly into the IntAct database. Members of the IntAct team supply appropriate levels of training, perform quality control on entries and take responsibility for long-term data maintenance. Recently, the MINT and IntAct databases decided to merge their separate efforts to make optimal use of limited developer resources and maximize the curation output. All data manually curated by the MINT curators have been moved into the IntAct database at EMBL-EBI and are merged with the existing IntAct dataset. Both IntAct and MINT are active contributors to the IMEx consortium (http://www.imexconsortium.org). PMID:24234451

  1. NSDNA: a manually curated database of experimentally supported ncRNAs associated with nervous system diseases.

    Science.gov (United States)

    Wang, Jianjian; Cao, Yuze; Zhang, Huixue; Wang, Tianfeng; Tian, Qinghua; Lu, Xiaoyu; Lu, Xiaoyan; Kong, Xiaotong; Liu, Zhaojun; Wang, Ning; Zhang, Shuai; Ma, Heping; Ning, Shangwei; Wang, Lihua

    2017-01-04

    The Nervous System Disease NcRNAome Atlas (NSDNA) (http://www.bio-bigdata.net/nsdna/) is a manually curated database that provides comprehensive experimentally supported associations about nervous system diseases (NSDs) and noncoding RNAs (ncRNAs). NSDs represent a common group of disorders, some of which are characterized by high morbidity and disabilities. The pathogenesis of NSDs at the molecular level remains poorly understood. ncRNAs are a large family of functionally important RNA molecules. Increasing evidence shows that diverse ncRNAs play a critical role in various NSDs. Mining and summarizing NSD-ncRNA association data can help researchers discover useful information. Hence, we developed an NSDNA database that documents 24 713 associations between 142 NSDs and 8593 ncRNAs in 11 species, curated from more than 1300 articles. This database provides a user-friendly interface for browsing and searching and allows for data downloading flexibility. In addition, NSDNA offers a submission page for researchers to submit novel NSD-ncRNA associations. It represents an extremely useful and valuable resource for researchers who seek to understand the functions and molecular mechanisms of ncRNA involved in NSDs. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  2. IMPPAT: A curated database of Indian Medicinal Plants, Phytochemistry And Therapeutics.

    Science.gov (United States)

    Mohanraj, Karthikeyan; Karthikeyan, Bagavathy Shanmugam; Vivek-Ananth, R P; Chand, R P Bharath; Aparna, S R; Mangalapandi, Pattulingam; Samal, Areejit

    2018-03-12

    Phytochemicals of medicinal plants encompass a diverse chemical space for drug discovery. India is rich with a flora of indigenous medicinal plants that have been used for centuries in traditional Indian medicine to treat human maladies. A comprehensive online database on the phytochemistry of Indian medicinal plants will enable computational approaches towards natural product based drug discovery. In this direction, we present, IMPPAT, a manually curated database of 1742 Indian Medicinal Plants, 9596 Phytochemicals, And 1124 Therapeutic uses spanning 27074 plant-phytochemical associations and 11514 plant-therapeutic associations. Notably, the curation effort led to a non-redundant in silico library of 9596 phytochemicals with standard chemical identifiers and structure information. Using cheminformatic approaches, we have computed the physicochemical, ADMET (absorption, distribution, metabolism, excretion, toxicity) and drug-likeliness properties of the IMPPAT phytochemicals. We show that the stereochemical complexity and shape complexity of IMPPAT phytochemicals differ from libraries of commercial compounds or diversity-oriented synthesis compounds while being similar to other libraries of natural products. Within IMPPAT, we have filtered a subset of 960 potential druggable phytochemicals, of which majority have no significant similarity to existing FDA approved drugs, and thus, rendering them as good candidates for prospective drugs. IMPPAT database is openly accessible at: https://cb.imsc.res.in/imppat .

  3. Text Mining Genotype-Phenotype Relationships from Biomedical Literature for Database Curation and Precision Medicine.

    Science.gov (United States)

    Singhal, Ayush; Simmons, Michael; Lu, Zhiyong

    2016-11-01

    The practice of precision medicine will ultimately require databases of genes and mutations for healthcare providers to reference in order to understand the clinical implications of each patient's genetic makeup. Although the highest quality databases require manual curation, text mining tools can facilitate the curation process, increasing accuracy, coverage, and productivity. However, to date there are no available text mining tools that offer high-accuracy performance for extracting such triplets from biomedical literature. In this paper we propose a high-performance machine learning approach to automate the extraction of disease-gene-variant triplets from biomedical literature. Our approach is unique because we identify the genes and protein products associated with each mutation from not just the local text content, but from a global context as well (from the Internet and from all literature in PubMed). Our approach also incorporates protein sequence validation and disease association using a novel text-mining-based machine learning approach. We extract disease-gene-variant triplets from all abstracts in PubMed related to a set of ten important diseases (breast cancer, prostate cancer, pancreatic cancer, lung cancer, acute myeloid leukemia, Alzheimer's disease, hemochromatosis, age-related macular degeneration (AMD), diabetes mellitus, and cystic fibrosis). We then evaluate our approach in two ways: (1) a direct comparison with the state of the art using benchmark datasets; (2) a validation study comparing the results of our approach with entries in a popular human-curated database (UniProt) for each of the previously mentioned diseases. In the benchmark comparison, our full approach achieves a 28% improvement in F1-measure (from 0.62 to 0.79) over the state-of-the-art results. For the validation study with UniProt Knowledgebase (KB), we present a thorough analysis of the results and errors. Across all diseases, our approach returned 272 triplets (disease

  4. Text Mining Genotype-Phenotype Relationships from Biomedical Literature for Database Curation and Precision Medicine.

    Directory of Open Access Journals (Sweden)

    Ayush Singhal

    2016-11-01

    Full Text Available The practice of precision medicine will ultimately require databases of genes and mutations for healthcare providers to reference in order to understand the clinical implications of each patient's genetic makeup. Although the highest quality databases require manual curation, text mining tools can facilitate the curation process, increasing accuracy, coverage, and productivity. However, to date there are no available text mining tools that offer high-accuracy performance for extracting such triplets from biomedical literature. In this paper we propose a high-performance machine learning approach to automate the extraction of disease-gene-variant triplets from biomedical literature. Our approach is unique because we identify the genes and protein products associated with each mutation from not just the local text content, but from a global context as well (from the Internet and from all literature in PubMed. Our approach also incorporates protein sequence validation and disease association using a novel text-mining-based machine learning approach. We extract disease-gene-variant triplets from all abstracts in PubMed related to a set of ten important diseases (breast cancer, prostate cancer, pancreatic cancer, lung cancer, acute myeloid leukemia, Alzheimer's disease, hemochromatosis, age-related macular degeneration (AMD, diabetes mellitus, and cystic fibrosis. We then evaluate our approach in two ways: (1 a direct comparison with the state of the art using benchmark datasets; (2 a validation study comparing the results of our approach with entries in a popular human-curated database (UniProt for each of the previously mentioned diseases. In the benchmark comparison, our full approach achieves a 28% improvement in F1-measure (from 0.62 to 0.79 over the state-of-the-art results. For the validation study with UniProt Knowledgebase (KB, we present a thorough analysis of the results and errors. Across all diseases, our approach returned 272 triplets

  5. The Neotoma Paleoecology Database: An International Community-Curated Resource for Paleoecological and Paleoenvironmental Data

    Science.gov (United States)

    Williams, J. W.; Grimm, E. C.; Ashworth, A. C.; Blois, J.; Charles, D. F.; Crawford, S.; Davis, E.; Goring, S. J.; Graham, R. W.; Miller, D. A.; Smith, A. J.; Stryker, M.; Uhen, M. D.

    2017-12-01

    The Neotoma Paleoecology Database supports global change research at the intersection of geology and ecology by providing a high-quality, community-curated data repository for paleoecological data. These data are widely used to study biological responses and feedbacks to past environmental change at local to global scales. The Neotoma data model is flexible and can store multiple kinds of fossil, biogeochemical, or physical variables measured from sedimentary archives. Data additions to Neotoma are growing and include >3.5 million observations, >16,000 datasets, and >8,500 sites. Dataset types include fossil pollen, vertebrates, diatoms, ostracodes, macroinvertebrates, plant macrofossils, insects, testate amoebae, geochronological data, and the recently added organic biomarkers, stable isotopes, and specimen-level data. Neotoma data can be found and retrieved in multiple ways, including the Explorer map-based interface, a RESTful Application Programming Interface, the neotoma R package, and digital object identifiers. Neotoma has partnered with the Paleobiology Database to produce a common data portal for paleobiological data, called the Earth Life Consortium. A new embargo management is designed to allow investigators to put their data into Neotoma and then make use of Neotoma's value-added services. Neotoma's distributed scientific governance model is flexible and scalable, with many open pathways for welcoming new members, data contributors, stewards, and research communities. As the volume and variety of scientific data grow, community-curated data resources such as Neotoma have become foundational infrastructure for big data science.

  6. CORE: a phylogenetically-curated 16S rDNA database of the core oral microbiome.

    Directory of Open Access Journals (Sweden)

    Ann L Griffen

    2011-04-01

    Full Text Available Comparing bacterial 16S rDNA sequences to GenBank and other large public databases via BLAST often provides results of little use for identification and taxonomic assignment of the organisms of interest. The human microbiome, and in particular the oral microbiome, includes many taxa, and accurate identification of sequence data is essential for studies of these communities. For this purpose, a phylogenetically curated 16S rDNA database of the core oral microbiome, CORE, was developed. The goal was to include a comprehensive and minimally redundant representation of the bacteria that regularly reside in the human oral cavity with computationally robust classification at the level of species and genus. Clades of cultivated and uncultivated taxa were formed based on sequence analyses using multiple criteria, including maximum-likelihood-based topology and bootstrap support, genetic distance, and previous naming. A number of classification inconsistencies for previously named species, especially at the level of genus, were resolved. The performance of the CORE database for identifying clinical sequences was compared to that of three publicly available databases, GenBank nr/nt, RDP and HOMD, using a set of sequencing reads that had not been used in creation of the database. CORE offered improved performance compared to other public databases for identification of human oral bacterial 16S sequences by a number of criteria. In addition, the CORE database and phylogenetic tree provide a framework for measures of community divergence, and the focused size of the database offers advantages of efficiency for BLAST searching of large datasets. The CORE database is available as a searchable interface and for download at http://microbiome.osu.edu.

  7. SolCyc: a database hub at the Sol Genomics Network (SGN) for the manual curation of metabolic networks in Solanum and Nicotiana specific databases

    Science.gov (United States)

    Foerster, Hartmut; Bombarely, Aureliano; Battey, James N D; Sierro, Nicolas; Ivanov, Nikolai V; Mueller, Lukas A

    2018-01-01

    Abstract SolCyc is the entry portal to pathway/genome databases (PGDBs) for major species of the Solanaceae family hosted at the Sol Genomics Network. Currently, SolCyc comprises six organism-specific PGDBs for tomato, potato, pepper, petunia, tobacco and one Rubiaceae, coffee. The metabolic networks of those PGDBs have been computationally predicted by the pathologic component of the pathway tools software using the manually curated multi-domain database MetaCyc (http://www.metacyc.org/) as reference. SolCyc has been recently extended by taxon-specific databases, i.e. the family-specific SolanaCyc database, containing only curated data pertinent to species of the nightshade family, and NicotianaCyc, a genus-specific database that stores all relevant metabolic data of the Nicotiana genus. Through manual curation of the published literature, new metabolic pathways have been created in those databases, which are complemented by the continuously updated, relevant species-specific pathways from MetaCyc. At present, SolanaCyc comprises 199 pathways and 29 superpathways and NicotianaCyc accounts for 72 pathways and 13 superpathways. Curator-maintained, taxon-specific databases such as SolanaCyc and NicotianaCyc are characterized by an enrichment of data specific to these taxa and free of falsely predicted pathways. Both databases have been used to update recently created Nicotiana-specific databases for Nicotiana tabacum, Nicotiana benthamiana, Nicotiana sylvestris and Nicotiana tomentosiformis by propagating verifiable data into those PGDBs. In addition, in-depth curation of the pathways in N.tabacum has been carried out which resulted in the elimination of 156 pathways from the 569 pathways predicted by pathway tools. Together, in-depth curation of the predicted pathway network and the supplementation with curated data from taxon-specific databases has substantially improved the curation status of the species–specific N.tabacum PGDB. The implementation of this

  8. A curated database of cyanobacterial strains relevant for modern taxonomy and phylogenetic studies.

    Science.gov (United States)

    Ramos, Vitor; Morais, João; Vasconcelos, Vitor M

    2017-04-25

    The dataset herein described lays the groundwork for an online database of relevant cyanobacterial strains, named CyanoType (http://lege.ciimar.up.pt/cyanotype). It is a database that includes categorized cyanobacterial strains useful for taxonomic, phylogenetic or genomic purposes, with associated information obtained by means of a literature-based curation. The dataset lists 371 strains and represents the first version of the database (CyanoType v.1). Information for each strain includes strain synonymy and/or co-identity, strain categorization, habitat, accession numbers for molecular data, taxonomy and nomenclature notes according to three different classification schemes, hierarchical automatic classification, phylogenetic placement according to a selection of relevant studies (including this), and important bibliographic references. The database will be updated periodically, namely by adding new strains meeting the criteria for inclusion and by revising and adding up-to-date metadata for strains already listed. A global 16S rDNA-based phylogeny is provided in order to assist users when choosing the appropriate strains for their studies.

  9. The Developmental Brain Disorders Database (DBDB): a curated neurogenetics knowledge base with clinical and research applications.

    Science.gov (United States)

    Mirzaa, Ghayda M; Millen, Kathleen J; Barkovich, A James; Dobyns, William B; Paciorkowski, Alex R

    2014-06-01

    The number of single genes associated with neurodevelopmental disorders has increased dramatically over the past decade. The identification of causative genes for these disorders is important to clinical outcome as it allows for accurate assessment of prognosis, genetic counseling, delineation of natural history, inclusion in clinical trials, and in some cases determines therapy. Clinicians face the challenge of correctly identifying neurodevelopmental phenotypes, recognizing syndromes, and prioritizing the best candidate genes for testing. However, there is no central repository of definitions for many phenotypes, leading to errors of diagnosis. Additionally, there is no system of levels of evidence linking genes to phenotypes, making it difficult for clinicians to know which genes are most strongly associated with a given condition. We have developed the Developmental Brain Disorders Database (DBDB: https://www.dbdb.urmc.rochester.edu/home), a publicly available, online-curated repository of genes, phenotypes, and syndromes associated with neurodevelopmental disorders. DBDB contains the first referenced ontology of developmental brain phenotypes, and uses a novel system of levels of evidence for gene-phenotype associations. It is intended to assist clinicians in arriving at the correct diagnosis, select the most appropriate genetic test for that phenotype, and improve the care of patients with developmental brain disorders. For researchers interested in the discovery of novel genes for developmental brain disorders, DBDB provides a well-curated source of important genes against which research sequencing results can be compared. Finally, DBDB allows novel observations about the landscape of the neurogenetics knowledge base. © 2014 Wiley Periodicals, Inc.

  10. Using random forests for assistance in the curation of G-protein coupled receptor databases.

    Science.gov (United States)

    Shkurin, Aleksei; Vellido, Alfredo

    2017-08-18

    Biology is experiencing a gradual but fast transformation from a laboratory-centred science towards a data-centred one. As such, it requires robust data engineering and the use of quantitative data analysis methods as part of database curation. This paper focuses on G protein-coupled receptors, a large and heterogeneous super-family of cell membrane proteins of interest to biology in general. One of its families, Class C, is of particular interest to pharmacology and drug design. This family is quite heterogeneous on its own, and the discrimination of its several sub-families is a challenging problem. In the absence of known crystal structure, such discrimination must rely on their primary amino acid sequences. We are interested not as much in achieving maximum sub-family discrimination accuracy using quantitative methods, but in exploring sequence misclassification behavior. Specifically, we are interested in isolating those sequences showing consistent misclassification, that is, sequences that are very often misclassified and almost always to the same wrong sub-family. Random forests are used for this analysis due to their ensemble nature, which makes them naturally suited to gauge the consistency of misclassification. This consistency is here defined through the voting scheme of their base tree classifiers. Detailed consistency results for the random forest ensemble classification were obtained for all receptors and for all data transformations of their unaligned primary sequences. Shortlists of the most consistently misclassified receptors for each subfamily and transformation, as well as an overall shortlist including those cases that were consistently misclassified across transformations, were obtained. The latter should be referred to experts for further investigation as a data curation task. The automatic discrimination of the Class C sub-families of G protein-coupled receptors from their unaligned primary sequences shows clear limits. This study has

  11. dinoref: A curated dinoflagellate (Dinophyceae) reference database for the 18S rRNA gene.

    Science.gov (United States)

    Mordret, Solenn; Piredda, Roberta; Vaulot, Daniel; Montresor, Marina; Kooistra, Wiebe H C F; Sarno, Diana

    2018-03-30

    Dinoflagellates are a heterogeneous group of protists present in all aquatic ecosystems where they occupy various ecological niches. They play a major role as primary producers, but many species are mixotrophic or heterotrophic. Environmental metabarcoding based on high-throughput sequencing is increasingly applied to assess diversity and abundance of planktonic organisms, and reference databases are definitely needed to taxonomically assign the huge number of sequences. We provide an updated 18S rRNA reference database of dinoflagellates: dinoref. Sequences were downloaded from genbank and filtered based on stringent quality criteria. All sequences were taxonomically curated, classified taking into account classical morphotaxonomic studies and molecular phylogenies, and linked to a series of metadata. dinoref includes 1,671 sequences representing 149 genera and 422 species. The taxonomic assignation of 468 sequences was revised. The largest number of sequences belongs to Gonyaulacales and Suessiales that include toxic and symbiotic species. dinoref provides an opportunity to test the level of taxonomic resolution of different 18S barcode markers based on a large number of sequences and species. As an example, when only the V4 region is considered, 374 of the 422 species included in dinoref can still be unambiguously identified. Clustering the V4 sequences at 98% similarity, a threshold that is commonly applied in metabarcoding studies, resulted in a considerable underestimation of species diversity. © 2018 John Wiley & Sons Ltd.

  12. AtlasT4SS: a curated database for type IV secretion systems.

    Science.gov (United States)

    Souza, Rangel C; del Rosario Quispe Saji, Guadalupe; Costa, Maiana O C; Netto, Diogo S; Lima, Nicholas C B; Klein, Cecília C; Vasconcelos, Ana Tereza R; Nicolás, Marisa F

    2012-08-09

    The type IV secretion system (T4SS) can be classified as a large family of macromolecule transporter systems, divided into three recognized sub-families, according to the well-known functions. The major sub-family is the conjugation system, which allows transfer of genetic material, such as a nucleoprotein, via cell contact among bacteria. Also, the conjugation system can transfer genetic material from bacteria to eukaryotic cells; such is the case with the T-DNA transfer of Agrobacterium tumefaciens to host plant cells. The system of effector protein transport constitutes the second sub-family, and the third one corresponds to the DNA uptake/release system. Genome analyses have revealed numerous T4SS in Bacteria and Archaea. The purpose of this work was to organize, classify, and integrate the T4SS data into a single database, called AtlasT4SS - the first public database devoted exclusively to this prokaryotic secretion system. The AtlasT4SS is a manual curated database that describes a large number of proteins related to the type IV secretion system reported so far in Gram-negative and Gram-positive bacteria, as well as in Archaea. The database was created using the RDBMS MySQL and the Catalyst Framework based in the Perl programming language and using the Model-View-Controller (MVC) design pattern for Web. The current version holds a comprehensive collection of 1,617 T4SS proteins from 58 Bacteria (49 Gram-negative and 9 Gram-Positive), one Archaea and 11 plasmids. By applying the bi-directional best hit (BBH) relationship in pairwise genome comparison, it was possible to obtain a core set of 134 clusters of orthologous genes encoding T4SS proteins. In our database we present one way of classifying orthologous groups of T4SSs in a hierarchical classification scheme with three levels. The first level comprises four classes that are based on the organization of genetic determinants, shared homologies, and evolutionary relationships: (i) F-T4SS, (ii) P-T4SS, (iii

  13. Text mining effectively scores and ranks the literature for improving chemical-gene-disease curation at the comparative toxicogenomics database.

    Directory of Open Access Journals (Sweden)

    Allan Peter Davis

    Full Text Available The Comparative Toxicogenomics Database (CTD; http://ctdbase.org/ is a public resource that curates interactions between environmental chemicals and gene products, and their relationships to diseases, as a means of understanding the effects of environmental chemicals on human health. CTD provides a triad of core information in the form of chemical-gene, chemical-disease, and gene-disease interactions that are manually curated from scientific articles. To increase the efficiency, productivity, and data coverage of manual curation, we have leveraged text mining to help rank and prioritize the triaged literature. Here, we describe our text-mining process that computes and assigns each article a document relevancy score (DRS, wherein a high DRS suggests that an article is more likely to be relevant for curation at CTD. We evaluated our process by first text mining a corpus of 14,904 articles triaged for seven heavy metals (cadmium, cobalt, copper, lead, manganese, mercury, and nickel. Based upon initial analysis, a representative subset corpus of 3,583 articles was then selected from the 14,094 articles and sent to five CTD biocurators for review. The resulting curation of these 3,583 articles was analyzed for a variety of parameters, including article relevancy, novel data content, interaction yield rate, mean average precision, and biological and toxicological interpretability. We show that for all measured parameters, the DRS is an effective indicator for scoring and improving the ranking of literature for the curation of chemical-gene-disease information at CTD. Here, we demonstrate how fully incorporating text mining-based DRS scoring into our curation pipeline enhances manual curation by prioritizing more relevant articles, thereby increasing data content, productivity, and efficiency.

  14. Text Mining Effectively Scores and Ranks the Literature for Improving Chemical-Gene-Disease Curation at the Comparative Toxicogenomics Database

    Science.gov (United States)

    Johnson, Robin J.; Lay, Jean M.; Lennon-Hopkins, Kelley; Saraceni-Richards, Cynthia; Sciaky, Daniela; Murphy, Cynthia Grondin; Mattingly, Carolyn J.

    2013-01-01

    The Comparative Toxicogenomics Database (CTD; http://ctdbase.org/) is a public resource that curates interactions between environmental chemicals and gene products, and their relationships to diseases, as a means of understanding the effects of environmental chemicals on human health. CTD provides a triad of core information in the form of chemical-gene, chemical-disease, and gene-disease interactions that are manually curated from scientific articles. To increase the efficiency, productivity, and data coverage of manual curation, we have leveraged text mining to help rank and prioritize the triaged literature. Here, we describe our text-mining process that computes and assigns each article a document relevancy score (DRS), wherein a high DRS suggests that an article is more likely to be relevant for curation at CTD. We evaluated our process by first text mining a corpus of 14,904 articles triaged for seven heavy metals (cadmium, cobalt, copper, lead, manganese, mercury, and nickel). Based upon initial analysis, a representative subset corpus of 3,583 articles was then selected from the 14,094 articles and sent to five CTD biocurators for review. The resulting curation of these 3,583 articles was analyzed for a variety of parameters, including article relevancy, novel data content, interaction yield rate, mean average precision, and biological and toxicological interpretability. We show that for all measured parameters, the DRS is an effective indicator for scoring and improving the ranking of literature for the curation of chemical-gene-disease information at CTD. Here, we demonstrate how fully incorporating text mining-based DRS scoring into our curation pipeline enhances manual curation by prioritizing more relevant articles, thereby increasing data content, productivity, and efficiency. PMID:23613709

  15. Designing a database for performance assessment: Lessons learned from WIPP

    International Nuclear Information System (INIS)

    Martell, M.A.; Schenker, A.

    1997-01-01

    The Waste Isolation Pilot Plant (WIPP) Compliance Certification Application (CCA) Performance Assessment (PA) used a relational database that was originally designed only to supply the input parameters required for implementation of the PA codes. Reviewers used the database as a point of entry to audit quality assurance measures for control, traceability, and retrievability of input information used for analysis, and output/work products. During these audits it became apparent that modifications to the architecture and scope of the database would benefit the EPA regulator and other stakeholders when reviewing the recertification application. This paper contains a discussion of the WPP PA CCA database and lessons learned for designing a database

  16. Lessons Learned From Developing Reactor Pressure Vessel Steel Embrittlement Database

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Jy-An John [ORNL

    2010-08-01

    Materials behaviors caused by neutron irradiation under fission and/or fusion environments can be little understood without practical examination. Easily accessible material information system with large material database using effective computers is necessary for design of nuclear materials and analyses or simulations of the phenomena. The developed Embrittlement Data Base (EDB) at ORNL is this comprehensive collection of data. EDB database contains power reactor pressure vessel surveillance data, the material test reactor data, foreign reactor data (through bilateral agreements authorized by NRC), and the fracture toughness data. The lessons learned from building EDB program and the associated database management activity regarding Material Database Design Methodology, Architecture and the Embedded QA Protocol are described in this report. The development of IAEA International Database on Reactor Pressure Vessel Materials (IDRPVM) and the comparison of EDB database and IAEA IDRPVM database are provided in the report. The recommended database QA protocol and database infrastructure are also stated in the report.

  17. MSDD: a manually curated database of experimentally supported associations among miRNAs, SNPs and human diseases

    OpenAIRE

    Yue, Ming; Zhou, Dianshuang; Zhi, Hui; Wang, Peng; Zhang, Yan; Gao, Yue; Guo, Maoni; Li, Xin; Wang, Yanxia; Zhang, Yunpeng; Ning, Shangwei; Li, Xia

    2017-01-01

    Abstract The MiRNA SNP Disease Database (MSDD, http://www.bio-bigdata.com/msdd/) is a manually curated database that provides comprehensive experimentally supported associations among microRNAs (miRNAs), single nucleotide polymorphisms (SNPs) and human diseases. SNPs in miRNA-related functional regions such as mature miRNAs, promoter regions, pri-miRNAs, pre-miRNAs and target gene 3′-UTRs, collectively called ‘miRSNPs’, represent a novel category of functional molecules. miRSNPs can lead to m...

  18. miRSponge: a manually curated database for experimentally supported miRNA sponges and ceRNAs.

    Science.gov (United States)

    Wang, Peng; Zhi, Hui; Zhang, Yunpeng; Liu, Yue; Zhang, Jizhou; Gao, Yue; Guo, Maoni; Ning, Shangwei; Li, Xia

    2015-01-01

    In this study, we describe miRSponge, a manually curated database, which aims at providing an experimentally supported resource for microRNA (miRNA) sponges. Recent evidence suggests that miRNAs are themselves regulated by competing endogenous RNAs (ceRNAs) or 'miRNA sponges' that contain miRNA binding sites. These competitive molecules can sequester miRNAs to prevent them interacting with their natural targets to play critical roles in various biological and pathological processes. It has become increasingly important to develop a high quality database to record and store ceRNA data to support future studies. To this end, we have established the experimentally supported miRSponge database that contains data on 599 miRNA-sponge interactions and 463 ceRNA relationships from 11 species following manual curating from nearly 1200 published articles. Database classes include endogenously generated molecules including coding genes, pseudogenes, long non-coding RNAs and circular RNAs, along with exogenously introduced molecules including viral RNAs and artificial engineered sponges. Approximately 70% of the interactions were identified experimentally in disease states. miRSponge provides a user-friendly interface for convenient browsing, retrieval and downloading of dataset. A submission page is also included to allow researchers to submit newly validated miRNA sponge data. Database URL: http://www.bio-bigdata.net/miRSponge. © The Author(s) 2015. Published by Oxford University Press.

  19. Bisphosphonate adverse effects, lessons from large databases

    DEFF Research Database (Denmark)

    Abrahamsen, Bo

    2010-01-01

    To review the latest findings on bisphosphonate safety from health databases, in particular sources that can provide incidence rates for stress fractures, osteonecrosis of the jaw (ONJ), atrial fibrillation and gastrointestinal lesions including esophageal cancer. The main focus is on bisphosphon...

  20. Lnc2Cancer: a manually curated database of experimentally supported lncRNAs associated with various human cancers.

    Science.gov (United States)

    Ning, Shangwei; Zhang, Jizhou; Wang, Peng; Zhi, Hui; Wang, Jianjian; Liu, Yue; Gao, Yue; Guo, Maoni; Yue, Ming; Wang, Lihua; Li, Xia

    2016-01-04

    Lnc2Cancer (http://www.bio-bigdata.net/lnc2cancer) is a manually curated database of cancer-associated long non-coding RNAs (lncRNAs) with experimental support that aims to provide a high-quality and integrated resource for exploring lncRNA deregulation in various human cancers. LncRNAs represent a large category of functional RNA molecules that play a significant role in human cancers. A curated collection and summary of deregulated lncRNAs in cancer is essential to thoroughly understand the mechanisms and functions of lncRNAs. Here, we developed the Lnc2Cancer database, which contains 1057 manually curated associations between 531 lncRNAs and 86 human cancers. Each association includes lncRNA and cancer name, the lncRNA expression pattern, experimental techniques, a brief functional description, the original reference and additional annotation information. Lnc2Cancer provides a user-friendly interface to conveniently browse, retrieve and download data. Lnc2Cancer also offers a submission page for researchers to submit newly validated lncRNA-cancer associations. With the rapidly increasing interest in lncRNAs, Lnc2Cancer will significantly improve our understanding of lncRNA deregulation in cancer and has the potential to be a timely and valuable resource. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  1. MortalityPredictors.org: a manually-curated database of published biomarkers of human all-cause mortality.

    Science.gov (United States)

    Peto, Maximus V; De la Guardia, Carlos; Winslow, Ksenia; Ho, Andrew; Fortney, Kristen; Morgen, Eric

    2017-08-31

    Biomarkers of all-cause mortality are of tremendous clinical and research interest. Because of the long potential duration of prospective human lifespan studies, such biomarkers can play a key role in quantifying human aging and quickly evaluating any potential therapies. Decades of research into mortality biomarkers have resulted in numerous associations documented across hundreds of publications. Here, we present MortalityPredictors.org , a manually-curated, publicly accessible database, housing published, statistically-significant relationships between biomarkers and all-cause mortality in population-based or generally healthy samples. To gather the information for this database, we searched PubMed for appropriate research papers and then manually curated relevant data from each paper. We manually curated 1,576 biomarker associations, involving 471 distinct biomarkers. Biomarkers ranged in type from hematologic (red blood cell distribution width) to molecular (DNA methylation changes) to physical (grip strength). Via the web interface, the resulting data can be easily browsed, searched, and downloaded for further analysis. MortalityPredictors.org provides comprehensive results on published biomarkers of human all-cause mortality that can be used to compare biomarkers, facilitate meta-analysis, assist with the experimental design of aging studies, and serve as a central resource for analysis. We hope that it will facilitate future research into human mortality and aging.

  2. submitter BioSharing: curated and crowd-sourced metadata standards, databases and data policies in the life sciences

    CERN Document Server

    McQuilton, Peter; Rocca-Serra, Philippe; Thurston, Milo; Lister, Allyson; Maguire, Eamonn; Sansone, Susanna-Assunta

    2016-01-01

    BioSharing (http://www.biosharing.org) is a manually curated, searchable portal of three linked registries. These resources cover standards (terminologies, formats and models, and reporting guidelines), databases, and data policies in the life sciences, broadly encompassing the biological, environmental and biomedical sciences. Launched in 2011 and built by the same core team as the successful MIBBI portal, BioSharing harnesses community curation to collate and cross-reference resources across the life sciences from around the world. BioSharing makes these resources findable and accessible (the core of the FAIR principle). Every record is designed to be interlinked, providing a detailed description not only on the resource itself, but also on its relations with other life science infrastructures. Serving a variety of stakeholders, BioSharing cultivates a growing community, to which it offers diverse benefits. It is a resource for funding bodies and journal publishers to navigate the metadata landscape of the ...

  3. The art of curation at a biological database: Principles and application

    Directory of Open Access Journals (Sweden)

    Sarah G. Odell

    2017-09-01

    Full Text Available The variety and quantity of data being produced by biological research has grown dramatically in recent years, resulting in an expansion of our understanding of biological systems. However, this abundance of data has brought new challenges, especially in curation. The role of biocurators is in part to filter research outcomes as they are generated, not only so that information is formatted and consolidated into locations that can provide long-term data sustainability, but also to ensure that the relevant data that was captured is reliable, reusable, and accessible. In many ways, biocuration lies somewhere between an art and a science. At GrainGenes (https://wheat.pw.usda.gov;https://graingenes.org, a long-time, stably-funded centralized repository for data about wheat, barley, rye, oat, and other small grains, curators have implemented a workflow for locating, parsing, and uploading new data so that the most important, peer-reviewed, high-quality research is available to users as quickly as possible with rich links to past research outcomes. In this report, we illustrate the principles and practical considerations of curation that we follow at GrainGenes with three case studies for curating a gene, a quantitative trait locus (QTL, and genomic elements. These examples demonstrate how our work allows users, i.e., small grains geneticists and breeders, to harness high-quality small grains data at GrainGenes to help them develop plants with enhanced agronomic traits.

  4. Improving the Discoverability and Availability of Sample Data and Imagery in NASA's Astromaterials Curation Digital Repository Using a New Common Architecture for Sample Databases

    Science.gov (United States)

    Todd, N. S.; Evans, C.

    2015-01-01

    The Astromaterials Acquisition and Curation Office at NASA's Johnson Space Center (JSC) is the designated facility for curating all of NASA's extraterrestrial samples. The suite of collections includes the lunar samples from the Apollo missions, cosmic dust particles falling into the Earth's atmosphere, meteorites collected in Antarctica, comet and interstellar dust particles from the Stardust mission, asteroid particles from the Japanese Hayabusa mission, and solar wind atoms collected during the Genesis mission. To support planetary science research on these samples, NASA's Astromaterials Curation Office hosts the Astromaterials Curation Digital Repository, which provides descriptions of the missions and collections, and critical information about each individual sample. Our office is implementing several informatics initiatives with the goal of better serving the planetary research community. One of these initiatives aims to increase the availability and discoverability of sample data and images through the use of a newly designed common architecture for Astromaterials Curation databases.

  5. Data Curation for the Exploitation of Large Earth Observation Products Databases - The MEA system

    Science.gov (United States)

    Mantovani, Simone; Natali, Stefano; Barboni, Damiano; Cavicchi, Mario; Della Vecchia, Andrea

    2014-05-01

    National Space Agencies under the umbrella of the European Space Agency are performing a strong activity to handle and provide solutions to Big Data and related knowledge (metadata, software tools and services) management and exploitation. The continuously increasing amount of long-term and of historic data in EO facilities in the form of online datasets and archives, the incoming satellite observation platforms that will generate an impressive amount of new data and the new EU approach on the data distribution policy make necessary to address technologies for the long-term management of these data sets, including their consolidation, preservation, distribution, continuation and curation across multiple missions. The management of long EO data time series of continuing or historic missions - with more than 20 years of data available already today - requires technical solutions and technologies which differ considerably from the ones exploited by existing systems. Several tools, both open source and commercial, are already providing technologies to handle data and metadata preparation, access and visualization via OGC standard interfaces. This study aims at describing the Multi-sensor Evolution Analysis (MEA) system and the Data Curation concept as approached and implemented within the ASIM and EarthServer projects, funded by the European Space Agency and the European Commission, respectively.

  6. Geroprotectors.org: a new, structured and curated database of current therapeutic interventions in aging and age-related disease

    Science.gov (United States)

    Moskalev, Alexey; Chernyagina, Elizaveta; de Magalhães, João Pedro; Barardo, Diogo; Thoppil, Harikrishnan; Shaposhnikov, Mikhail; Budovsky, Arie; Fraifeld, Vadim E.; Garazha, Andrew; Tsvetkov, Vasily; Bronovitsky, Evgeny; Bogomolov, Vladislav; Scerbacov, Alexei; Kuryan, Oleg; Gurinovich, Roman; Jellen, Leslie C.; Kennedy, Brian; Mamoshina, Polina; Dobrovolskaya, Evgeniya; Aliper, Alex; Kaminsky, Dmitry; Zhavoronkov, Alex

    2015-01-01

    As the level of interest in aging research increases, there is a growing number of geroprotectors, or therapeutic interventions that aim to extend the healthy lifespan and repair or reduce aging-related damage in model organisms and, eventually, in humans. There is a clear need for a manually-curated database of geroprotectors to compile and index their effects on aging and age-related diseases and link these effects to relevant studies and multiple biochemical and drug databases. Here, we introduce the first such resource, Geroprotectors (http://geroprotectors.org). Geroprotectors is a public, rapidly explorable database that catalogs over 250 experiments involving over 200 known or candidate geroprotectors that extend lifespan in model organisms. Each compound has a comprehensive profile complete with biochemistry, mechanisms, and lifespan effects in various model organisms, along with information ranging from chemical structure, side effects, and toxicity to FDA drug status. These are presented in a visually intuitive, efficient framework fit for casual browsing or in-depth research alike. Data are linked to the source studies or databases, providing quick and convenient access to original data. The Geroprotectors database facilitates cross-study, cross-organism, and cross-discipline analysis and saves countless hours of inefficient literature and web searching. Geroprotectors is a one-stop, knowledge-sharing, time-saving resource for researchers seeking healthy aging solutions. PMID:26342919

  7. EVLncRNAs: a manually curated database for long non-coding RNAs validated by low-throughput experiments

    Science.gov (United States)

    Zhao, Huiying; Yu, Jiafeng; Guo, Chengang; Dou, Xianghua; Song, Feng; Hu, Guodong; Cao, Zanxia; Qu, Yuanxu

    2018-01-01

    Abstract Long non-coding RNAs (lncRNAs) play important functional roles in various biological processes. Early databases were utilized to deposit all lncRNA candidates produced by high-throughput experimental and/or computational techniques to facilitate classification, assessment and validation. As more lncRNAs are validated by low-throughput experiments, several databases were established for experimentally validated lncRNAs. However, these databases are small in scale (with a few hundreds of lncRNAs only) and specific in their focuses (plants, diseases or interactions). Thus, it is highly desirable to have a comprehensive dataset for experimentally validated lncRNAs as a central repository for all of their structures, functions and phenotypes. Here, we established EVLncRNAs by curating lncRNAs validated by low-throughput experiments (up to 1 May 2016) and integrating specific databases (lncRNAdb, LncRANDisease, Lnc2Cancer and PLNIncRBase) with additional functional and disease-specific information not covered previously. The current version of EVLncRNAs contains 1543 lncRNAs from 77 species that is 2.9 times larger than the current largest database for experimentally validated lncRNAs. Seventy-four percent lncRNA entries are partially or completely new, comparing to all existing experimentally validated databases. The established database allows users to browse, search and download as well as to submit experimentally validated lncRNAs. The database is available at http://biophy.dzu.edu.cn/EVLncRNAs. PMID:28985416

  8. NEMiD: a web-based curated microbial diversity database with geo-based plotting.

    Science.gov (United States)

    Bhattacharjee, Kaushik; Joshi, Santa Ram

    2014-01-01

    The majority of the Earth's microbes remain unknown, and that their potential utility cannot be exploited until they are discovered and characterized. They provide wide scope for the development of new strains as well as biotechnological uses. The documentation and bioprospection of microorganisms carry enormous significance considering their relevance to human welfare. This calls for an urgent need to develop a database with emphasis on the microbial diversity of the largest untapped reservoirs in the biosphere. The data annotated in the North-East India Microbial database (NEMiD) were obtained by the isolation and characterization of microbes from different parts of the Eastern Himalayan region. The database was constructed as a relational database management system (RDBMS) for data storage in MySQL in the back-end on a Linux server and implemented in an Apache/PHP environment. This database provides a base for understanding the soil microbial diversity pattern in this megabiodiversity hotspot and indicates the distribution patterns of various organisms along with identification. The NEMiD database is freely available at www.mblabnehu.info/nemid/.

  9. NEMiD: A Web-Based Curated Microbial Diversity Database with Geo-Based Plotting

    Science.gov (United States)

    Bhattacharjee, Kaushik; Joshi, Santa Ram

    2014-01-01

    The majority of the Earth's microbes remain unknown, and that their potential utility cannot be exploited until they are discovered and characterized. They provide wide scope for the development of new strains as well as biotechnological uses. The documentation and bioprospection of microorganisms carry enormous significance considering their relevance to human welfare. This calls for an urgent need to develop a database with emphasis on the microbial diversity of the largest untapped reservoirs in the biosphere. The data annotated in the North-East India Microbial database (NEMiD) were obtained by the isolation and characterization of microbes from different parts of the Eastern Himalayan region. The database was constructed as a relational database management system (RDBMS) for data storage in MySQL in the back-end on a Linux server and implemented in an Apache/PHP environment. This database provides a base for understanding the soil microbial diversity pattern in this megabiodiversity hotspot and indicates the distribution patterns of various organisms along with identification. The NEMiD database is freely available at www.mblabnehu.info/nemid/. PMID:24714636

  10. MSDD: a manually curated database of experimentally supported associations among miRNAs, SNPs and human diseases.

    Science.gov (United States)

    Yue, Ming; Zhou, Dianshuang; Zhi, Hui; Wang, Peng; Zhang, Yan; Gao, Yue; Guo, Maoni; Li, Xin; Wang, Yanxia; Zhang, Yunpeng; Ning, Shangwei; Li, Xia

    2018-01-04

    The MiRNA SNP Disease Database (MSDD, http://www.bio-bigdata.com/msdd/) is a manually curated database that provides comprehensive experimentally supported associations among microRNAs (miRNAs), single nucleotide polymorphisms (SNPs) and human diseases. SNPs in miRNA-related functional regions such as mature miRNAs, promoter regions, pri-miRNAs, pre-miRNAs and target gene 3'-UTRs, collectively called 'miRSNPs', represent a novel category of functional molecules. miRSNPs can lead to miRNA and its target gene dysregulation, and resulting in susceptibility to or onset of human diseases. A curated collection and summary of miRSNP-associated diseases is essential for a thorough understanding of the mechanisms and functions of miRSNPs. Here, we describe MSDD, which currently documents 525 associations among 182 human miRNAs, 197 SNPs, 153 genes and 164 human diseases through a review of more than 2000 published papers. Each association incorporates information on the miRNAs, SNPs, miRNA target genes and disease names, SNP locations and alleles, the miRNA dysfunctional pattern, experimental techniques, a brief functional description, the original reference and additional annotation. MSDD provides a user-friendly interface to conveniently browse, retrieve, download and submit novel data. MSDD will significantly improve our understanding of miRNA dysfunction in disease, and thus, MSDD has the potential to serve as a timely and valuable resource. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  11. DREMECELS: A Curated Database for Base Excision and Mismatch Repair Mechanisms Associated Human Malignancies.

    Directory of Open Access Journals (Sweden)

    Ankita Shukla

    Full Text Available DNA repair mechanisms act as a warrior combating various damaging processes that ensue critical malignancies. DREMECELS was designed considering the malignancies with frequent alterations in DNA repair pathways, that is, colorectal and endometrial cancers, associated with Lynch syndrome (also known as HNPCC. Since lynch syndrome carries high risk (~40-60% for both cancers, therefore we decided to cover all three diseases in this portal. Although a large population is presently affected by these malignancies, many resources are available for various cancer types but no database archives information on the genes specifically for only these cancers and disorders. The database contains 156 genes and two repair mechanisms, base excision repair (BER and mismatch repair (MMR. Other parameters include some of the regulatory processes that have roles in these disease progressions due to incompetent repair mechanisms, specifically BER and MMR. However, our unique database mainly provides qualitative and quantitative information on these cancer types along with methylation, drug sensitivity, miRNAs, copy number variation (CNV and somatic mutations data. This database would serve the scientific community by providing integrated information on these disease types, thus sustaining diagnostic and therapeutic processes. This repository would serve as an excellent accompaniment for researchers and biomedical professionals and facilitate in understanding such critical diseases. DREMECELS is publicly available at http://www.bioinfoindia.org/dremecels.

  12. dbEM: A database of epigenetic modifiers curated from cancerous and normal genomes

    Science.gov (United States)

    Singh Nanda, Jagpreet; Kumar, Rahul; Raghava, Gajendra P. S.

    2016-01-01

    We have developed a database called dbEM (database of Epigenetic Modifiers) to maintain the genomic information of about 167 epigenetic modifiers/proteins, which are considered as potential cancer targets. In dbEM, modifiers are classified on functional basis and comprise of 48 histone methyl transferases, 33 chromatin remodelers and 31 histone demethylases. dbEM maintains the genomic information like mutations, copy number variation and gene expression in thousands of tumor samples, cancer cell lines and healthy samples. This information is obtained from public resources viz. COSMIC, CCLE and 1000-genome project. Gene essentiality data retrieved from COLT database further highlights the importance of various epigenetic proteins for cancer survival. We have also reported the sequence profiles, tertiary structures and post-translational modifications of these epigenetic proteins in cancer. It also contains information of 54 drug molecules against different epigenetic proteins. A wide range of tools have been integrated in dbEM e.g. Search, BLAST, Alignment and Profile based prediction. In our analysis, we found that epigenetic proteins DNMT3A, HDAC2, KDM6A, and TET2 are highly mutated in variety of cancers. We are confident that dbEM will be very useful in cancer research particularly in the field of epigenetic proteins based cancer therapeutics. This database is available for public at URL: http://crdd.osdd.net/raghava/dbem.

  13. The Coral Trait Database, a curated database of trait information for coral species from the global oceans

    Science.gov (United States)

    Madin, Joshua S.; Anderson, Kristen D.; Andreasen, Magnus Heide; Bridge, Tom C. L.; Cairns, Stephen D.; Connolly, Sean R.; Darling, Emily S.; Diaz, Marcela; Falster, Daniel S.; Franklin, Erik C.; Gates, Ruth D.; Hoogenboom, Mia O.; Huang, Danwei; Keith, Sally A.; Kosnik, Matthew A.; Kuo, Chao-Yang; Lough, Janice M.; Lovelock, Catherine E.; Luiz, Osmar; Martinelli, Julieta; Mizerek, Toni; Pandolfi, John M.; Pochon, Xavier; Pratchett, Morgan S.; Putnam, Hollie M.; Roberts, T. Edward; Stat, Michael; Wallace, Carden C.; Widman, Elizabeth; Baird, Andrew H.

    2016-03-01

    Trait-based approaches advance ecological and evolutionary research because traits provide a strong link to an organism’s function and fitness. Trait-based research might lead to a deeper understanding of the functions of, and services provided by, ecosystems, thereby improving management, which is vital in the current era of rapid environmental change. Coral reef scientists have long collected trait data for corals; however, these are difficult to access and often under-utilized in addressing large-scale questions. We present the Coral Trait Database initiative that aims to bring together physiological, morphological, ecological, phylogenetic and biogeographic trait information into a single repository. The database houses species- and individual-level data from published field and experimental studies alongside contextual data that provide important framing for analyses. In this data descriptor, we release data for 56 traits for 1547 species, and present a collaborative platform on which other trait data are being actively federated. Our overall goal is for the Coral Trait Database to become an open-source, community-led data clearinghouse that accelerates coral reef research.

  14. The Coral Trait Database, a curated database of trait information for coral species from the global oceans.

    Science.gov (United States)

    Madin, Joshua S; Anderson, Kristen D; Andreasen, Magnus Heide; Bridge, Tom C L; Cairns, Stephen D; Connolly, Sean R; Darling, Emily S; Diaz, Marcela; Falster, Daniel S; Franklin, Erik C; Gates, Ruth D; Harmer, Aaron; Hoogenboom, Mia O; Huang, Danwei; Keith, Sally A; Kosnik, Matthew A; Kuo, Chao-Yang; Lough, Janice M; Lovelock, Catherine E; Luiz, Osmar; Martinelli, Julieta; Mizerek, Toni; Pandolfi, John M; Pochon, Xavier; Pratchett, Morgan S; Putnam, Hollie M; Roberts, T Edward; Stat, Michael; Wallace, Carden C; Widman, Elizabeth; Baird, Andrew H

    2016-03-29

    Trait-based approaches advance ecological and evolutionary research because traits provide a strong link to an organism's function and fitness. Trait-based research might lead to a deeper understanding of the functions of, and services provided by, ecosystems, thereby improving management, which is vital in the current era of rapid environmental change. Coral reef scientists have long collected trait data for corals; however, these are difficult to access and often under-utilized in addressing large-scale questions. We present the Coral Trait Database initiative that aims to bring together physiological, morphological, ecological, phylogenetic and biogeographic trait information into a single repository. The database houses species- and individual-level data from published field and experimental studies alongside contextual data that provide important framing for analyses. In this data descriptor, we release data for 56 traits for 1547 species, and present a collaborative platform on which other trait data are being actively federated. Our overall goal is for the Coral Trait Database to become an open-source, community-led data clearinghouse that accelerates coral reef research.

  15. A site-specific curated database for the microorganisms of activated sludge and anaerobic digesters

    DEFF Research Database (Denmark)

    McIlroy, Simon Jon; Kirkegaard, Rasmus Hansen; McIlroy, Bianca

    RNA gene amplicon sequencing (V1-3 region), including full-scale AS (20 plants, 8 years) and AD systems (36 reactors, 18 plants, 4 years). Surveys also include the Archaea (V3-5 region). The MiDAS field guide is intended as a collaborative platform for researchers and wastewater treatment practitioners...... taxonomy, proposes putative names for each genus-level-taxon that can be used as a common vocabulary for all researchers in the field. The online database covers >250 genera found to be abundant and/or important in biological nutrient removal treatment plants, based on extensive in-house surveys with 16S r...

  16. A curated gluten protein sequence database to support development of proteomics methods for determination of gluten in gluten-free foods.

    Science.gov (United States)

    Bromilow, Sophie; Gethings, Lee A; Buckley, Mike; Bromley, Mike; Shewry, Peter R; Langridge, James I; Clare Mills, E N

    2017-06-23

    The unique physiochemical properties of wheat gluten enable a diverse range of food products to be manufactured. However, gluten triggers coeliac disease, a condition which is treated using a gluten-free diet. Analytical methods are required to confirm if foods are gluten-free, but current immunoassay-based methods can unreliable and proteomic methods offer an alternative but require comprehensive and well annotated sequence databases which are lacking for gluten. A manually a curated database (GluPro V1.0) of gluten proteins, comprising 630 discrete unique full length protein sequences has been compiled. It is representative of the different types of gliadin and glutenin components found in gluten. An in silico comparison of their coeliac toxicity was undertaken by analysing the distribution of coeliac toxic motifs. This demonstrated that whilst the α-gliadin proteins contained more toxic motifs, these were distributed across all gluten protein sub-types. Comparison of annotations observed using a discovery proteomics dataset acquired using ion mobility MS/MS showed that more reliable identifications were obtained using the GluPro V1.0 database compared to the complete reviewed Viridiplantae database. This highlights the value of a curated sequence database specifically designed to support the proteomic workflows and the development of methods to detect and quantify gluten. We have constructed the first manually curated open-source wheat gluten protein sequence database (GluPro V1.0) in a FASTA format to support the application of proteomic methods for gluten protein detection and quantification. We have also analysed the manually verified sequences to give the first comprehensive overview of the distribution of sequences able to elicit a reaction in coeliac disease, the prevalent form of gluten intolerance. Provision of this database will improve the reliability of gluten protein identification by proteomic analysis, and aid the development of targeted mass

  17. From field to database : a user-oriented approche to promote cyber-curating of scientific drilling cores

    Science.gov (United States)

    Pignol, C.; Arnaud, F.; Godinho, E.; Galabertier, B.; Caillo, A.; Billy, I.; Augustin, L.; Calzas, M.; Rousseau, D. D.; Crosta, X.

    2016-12-01

    Managing scientific data is probably one the most challenging issues in modern science. In plaeosciences the question is made even more sensitive with the need of preserving and managing high value fragile geological samples: cores. Large international scientific programs, such as IODP or ICDP led intense effort to solve this problem and proposed detailed high standard work- and dataflows thorough core handling and curating. However many paleoscience results derived from small-scale research programs in which data and sample management is too often managed only locally - when it is… In this paper we present a national effort leads in France to develop an integrated system to curate ice and sediment cores. Under the umbrella of the national excellence equipment program CLIMCOR, we launched a reflexion about core curating and the management of associated fieldwork data. Our aim was then to conserve all data from fieldwork in an integrated cyber-environment which will evolve toward laboratory-acquired data storage in a near future. To do so, our demarche was conducted through an intimate relationship with field operators as well laboratory core curators in order to propose user-oriented solutions. The national core curating initiative proposes a single web portal in which all teams can store their fieldwork data. This portal is used as a national hub to attribute IGSNs. For legacy samples, this requires the establishment of a dedicated core list with associated metadata. However, for forthcoming core data, we developed a mobile application to capture technical and scientific data directly on the field. This application is linked with a unique coring-tools library and is adapted to most coring devices (gravity, drilling, percussion etc.) including multiple sections and holes coring operations. Those field data can be uploaded automatically to the national portal, but also referenced through international standards (IGSN and INSPIRE) and displayed in international

  18. Database Systems and Oracle: Experiences and Lessons Learned

    Science.gov (United States)

    Dunn, Deborah

    2005-01-01

    In a tight job market, IT professionals with database experience are likely to be in great demand. Companies need database personnel who can help improve access to and security of data. The events of September 11 have increased business' awareness of the need for database security, backup, and recovery procedures. It is our responsibility to…

  19. XML: James Webb Space Telescope Database Issues, Lessons, and Status

    Science.gov (United States)

    Detter, Ryan; Mooney, Michael; Fatig, Curtis

    2003-01-01

    This paper will present the current concept using extensible Markup Language (XML) as the underlying structure for the James Webb Space Telescope (JWST) database. The purpose of using XML is to provide a JWST database, independent of any portion of the ground system, yet still compatible with the various systems using a variety of different structures. The testing of the JWST Flight Software (FSW) started in 2002, yet the launch is scheduled for 2011 with a planned 5-year mission and a 5-year follow on option. The initial database and ground system elements, including the commands, telemetry, and ground system tools will be used for 19 years, plus post mission activities. During the Integration and Test (I&T) phases of the JWST development, 24 distinct laboratories, each geographically dispersed, will have local database tools with an XML database. Each of these laboratories database tools will be used for the exporting and importing of data both locally and to a central database system, inputting data to the database certification process, and providing various reports. A centralized certified database repository will be maintained by the Space Telescope Science Institute (STScI), in Baltimore, Maryland, USA. One of the challenges for the database is to be flexible enough to allow for the upgrade, addition or changing of individual items without effecting the entire ground system. Also, using XML should allow for the altering of the import and export formats needed by the various elements, tracking the verification/validation of each database item, allow many organizations to provide database inputs, and the merging of the many existing database processes into one central database structure throughout the JWST program. Many National Aeronautics and Space Administration (NASA) projects have attempted to take advantage of open source and commercial technology. Often this causes a greater reliance on the use of Commercial-Off-The-Shelf (COTS), which is often limiting

  20. Data Curation

    Science.gov (United States)

    Mallon, Melissa, Ed.

    2012-01-01

    In their Top Trends of 2012, the Association of College and Research Libraries (ACRL) named data curation as one of the issues to watch in academic libraries in the near future (ACRL, 2012, p. 312). Data curation can be summarized as "the active and ongoing management of data through its life cycle of interest and usefulness to scholarship,…

  1. Consensus coding sequence (CCDS) database: a standardized set of human and mouse protein-coding regions supported by expert curation.

    Science.gov (United States)

    Pujar, Shashikant; O'Leary, Nuala A; Farrell, Catherine M; Loveland, Jane E; Mudge, Jonathan M; Wallin, Craig; Girón, Carlos G; Diekhans, Mark; Barnes, If; Bennett, Ruth; Berry, Andrew E; Cox, Eric; Davidson, Claire; Goldfarb, Tamara; Gonzalez, Jose M; Hunt, Toby; Jackson, John; Joardar, Vinita; Kay, Mike P; Kodali, Vamsi K; Martin, Fergal J; McAndrews, Monica; McGarvey, Kelly M; Murphy, Michael; Rajput, Bhanu; Rangwala, Sanjida H; Riddick, Lillian D; Seal, Ruth L; Suner, Marie-Marthe; Webb, David; Zhu, Sophia; Aken, Bronwen L; Bruford, Elspeth A; Bult, Carol J; Frankish, Adam; Murphy, Terence; Pruitt, Kim D

    2018-01-04

    The Consensus Coding Sequence (CCDS) project provides a dataset of protein-coding regions that are identically annotated on the human and mouse reference genome assembly in genome annotations produced independently by NCBI and the Ensembl group at EMBL-EBI. This dataset is the product of an international collaboration that includes NCBI, Ensembl, HUGO Gene Nomenclature Committee, Mouse Genome Informatics and University of California, Santa Cruz. Identically annotated coding regions, which are generated using an automated pipeline and pass multiple quality assurance checks, are assigned a stable and tracked identifier (CCDS ID). Additionally, coordinated manual review by expert curators from the CCDS collaboration helps in maintaining the integrity and high quality of the dataset. The CCDS data are available through an interactive web page (https://www.ncbi.nlm.nih.gov/CCDS/CcdsBrowse.cgi) and an FTP site (ftp://ftp.ncbi.nlm.nih.gov/pub/CCDS/). In this paper, we outline the ongoing work, growth and stability of the CCDS dataset and provide updates on new collaboration members and new features added to the CCDS user interface. We also present expert curation scenarios, with specific examples highlighting the importance of an accurate reference genome assembly and the crucial role played by input from the research community. Published by Oxford University Press on behalf of Nucleic Acids Research 2017.

  2. PFR²: a curated database of planktonic foraminifera 18S ribosomal DNA as a resource for studies of plankton ecology, biogeography and evolution.

    Science.gov (United States)

    Morard, Raphaël; Darling, Kate F; Mahé, Frédéric; Audic, Stéphane; Ujiié, Yurika; Weiner, Agnes K M; André, Aurore; Seears, Heidi A; Wade, Christopher M; Quillévéré, Frédéric; Douady, Christophe J; Escarguel, Gilles; de Garidel-Thoron, Thibault; Siccha, Michael; Kucera, Michal; de Vargas, Colomban

    2015-11-01

    Planktonic foraminifera (Rhizaria) are ubiquitous marine pelagic protists producing calcareous shells with conspicuous morphology. They play an important role in the marine carbon cycle, and their exceptional fossil record serves as the basis for biochronostratigraphy and past climate reconstructions. A major worldwide sampling effort over the last two decades has resulted in the establishment of multiple large collections of cryopreserved individual planktonic foraminifera samples. Thousands of 18S rDNA partial sequences have been generated, representing all major known morphological taxa across their worldwide oceanic range. This comprehensive data coverage provides an opportunity to assess patterns of molecular ecology and evolution in a holistic way for an entire group of planktonic protists. We combined all available published and unpublished genetic data to build PFR(2), the Planktonic foraminifera Ribosomal Reference database. The first version of the database includes 3322 reference 18S rDNA sequences belonging to 32 of the 47 known morphospecies of extant planktonic foraminifera, collected from 460 oceanic stations. All sequences have been rigorously taxonomically curated using a six-rank annotation system fully resolved to the morphological species level and linked to a series of metadata. The PFR(2) website, available at http://pfr2.sb-roscoff.fr, allows downloading the entire database or specific sections, as well as the identification of new planktonic foraminiferal sequences. Its novel, fully documented curation process integrates advances in morphological and molecular taxonomy. It allows for an increase in its taxonomic resolution and assures that integrity is maintained by including a complete contingency tracking of annotations and assuring that the annotations remain internally consistent. © 2015 John Wiley & Sons Ltd.

  3. Lessons Learned from resolving massive IPS database change for SPADES+

    International Nuclear Information System (INIS)

    Kim, Jin-Soo

    2016-01-01

    Safety Parameter Display and Evaluation System+ (SPADES+) was implemented to meet the requirements for Safety Parameter Display System (SPDS) which are related to TMI Action Plan requirements. SPADES+ monitors continuously the critical safety function during normal, abnormal, and emergency operation mode and generates the alarm output to the alarm server when the tolerance related to safety functions are not satisfied. The alarm algorithm for critical safety function is performed in the NSSS Application Software (NAPS) server of the Information Process System (IPS) and the calculation result will be displayed on the flat panel display (FPD) of the IPS. SPADES+ provides the critical variable to the control room operators to aid them in rapidly and reliable determining the safety status of the plant. Many database point ID names (518 points) were changed. POINT_ID is used in the programming source code, the related documents such as SDS and SRS, and Graphic database. To reduce human errors, computer program and office program’s Macro are used. Though the automatic methods are used for changing POINT_IDs, it takes lots of time to resolve for editing the change list except for making computerized solutions. In IPS, there are many more programs than SPADES+ and over 30,000 POINT_IDs are in IPS database. Changing POINT_IDs could be a burden to software engineers. In case of Ovation system database, there is the Alias field to prevent this kind of problem. The Alias is a kind of secondary key in database

  4. Lessons Learned from resolving massive IPS database change for SPADES+

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jin-Soo [KEPCO Engineering and Construction Co., Deajeon (Korea, Republic of)

    2016-10-15

    Safety Parameter Display and Evaluation System+ (SPADES+) was implemented to meet the requirements for Safety Parameter Display System (SPDS) which are related to TMI Action Plan requirements. SPADES+ monitors continuously the critical safety function during normal, abnormal, and emergency operation mode and generates the alarm output to the alarm server when the tolerance related to safety functions are not satisfied. The alarm algorithm for critical safety function is performed in the NSSS Application Software (NAPS) server of the Information Process System (IPS) and the calculation result will be displayed on the flat panel display (FPD) of the IPS. SPADES+ provides the critical variable to the control room operators to aid them in rapidly and reliable determining the safety status of the plant. Many database point ID names (518 points) were changed. POINT{sub I}D is used in the programming source code, the related documents such as SDS and SRS, and Graphic database. To reduce human errors, computer program and office program’s Macro are used. Though the automatic methods are used for changing POINT{sub I}Ds, it takes lots of time to resolve for editing the change list except for making computerized solutions. In IPS, there are many more programs than SPADES+ and over 30,000 POINT{sub I}Ds are in IPS database. Changing POINT{sub I}Ds could be a burden to software engineers. In case of Ovation system database, there is the Alias field to prevent this kind of problem. The Alias is a kind of secondary key in database.

  5. PhytoREF: a reference database of the plastidial 16S rRNA gene of photosynthetic eukaryotes with curated taxonomy.

    Science.gov (United States)

    Decelle, Johan; Romac, Sarah; Stern, Rowena F; Bendif, El Mahdi; Zingone, Adriana; Audic, Stéphane; Guiry, Michael D; Guillou, Laure; Tessier, Désiré; Le Gall, Florence; Gourvil, Priscillia; Dos Santos, Adriana L; Probert, Ian; Vaulot, Daniel; de Vargas, Colomban; Christen, Richard

    2015-11-01

    Photosynthetic eukaryotes have a critical role as the main producers in most ecosystems of the biosphere. The ongoing environmental metabarcoding revolution opens the perspective for holistic ecosystems biological studies of these organisms, in particular the unicellular microalgae that often lack distinctive morphological characters and have complex life cycles. To interpret environmental sequences, metabarcoding necessarily relies on taxonomically curated databases containing reference sequences of the targeted gene (or barcode) from identified organisms. To date, no such reference framework exists for photosynthetic eukaryotes. In this study, we built the PhytoREF database that contains 6490 plastidial 16S rDNA reference sequences that originate from a large diversity of eukaryotes representing all known major photosynthetic lineages. We compiled 3333 amplicon sequences available from public databases and 879 sequences extracted from plastidial genomes, and generated 411 novel sequences from cultured marine microalgal strains belonging to different eukaryotic lineages. A total of 1867 environmental Sanger 16S rDNA sequences were also included in the database. Stringent quality filtering and a phylogeny-based taxonomic classification were applied for each 16S rDNA sequence. The database mainly focuses on marine microalgae, but sequences from land plants (representing half of the PhytoREF sequences) and freshwater taxa were also included to broaden the applicability of PhytoREF to different aquatic and terrestrial habitats. PhytoREF, accessible via a web interface (http://phytoref.fr), is a new resource in molecular ecology to foster the discovery, assessment and monitoring of the diversity of photosynthetic eukaryotes using high-throughput sequencing. © 2015 John Wiley & Sons Ltd.

  6. AT_CHLORO, a comprehensive chloroplast proteome database with subplastidial localization and curated information on envelope proteins.

    Science.gov (United States)

    Ferro, Myriam; Brugière, Sabine; Salvi, Daniel; Seigneurin-Berny, Daphné; Court, Magali; Moyet, Lucas; Ramus, Claire; Miras, Stéphane; Mellal, Mourad; Le Gall, Sophie; Kieffer-Jaquinod, Sylvie; Bruley, Christophe; Garin, Jérôme; Joyard, Jacques; Masselon, Christophe; Rolland, Norbert

    2010-06-01

    Recent advances in the proteomics field have allowed a series of high throughput experiments to be conducted on chloroplast samples, and the data are available in several public databases. However, the accurate localization of many chloroplast proteins often remains hypothetical. This is especially true for envelope proteins. We went a step further into the knowledge of the chloroplast proteome by focusing, in the same set of experiments, on the localization of proteins in the stroma, the thylakoids, and envelope membranes. LC-MS/MS-based analyses first allowed building the AT_CHLORO database (http://www.grenoble.prabi.fr/protehome/grenoble-plant-proteomics/), a comprehensive repertoire of the 1323 proteins, identified by 10,654 unique peptide sequences, present in highly purified chloroplasts and their subfractions prepared from Arabidopsis thaliana leaves. This database also provides extensive proteomics information (peptide sequences and molecular weight, chromatographic retention times, MS/MS spectra, and spectral count) for a unique chloroplast protein accurate mass and time tag database gathering identified peptides with their respective and precise analytical coordinates, molecular weight, and retention time. We assessed the partitioning of each protein in the three chloroplast compartments by using a semiquantitative proteomics approach (spectral count). These data together with an in-depth investigation of the literature were compiled to provide accurate subplastidial localization of previously known and newly identified proteins. A unique knowledge base containing extensive information on the proteins identified in envelope fractions was thus obtained, allowing new insights into this membrane system to be revealed. Altogether, the data we obtained provide unexpected information about plastidial or subplastidial localization of some proteins that were not suspected to be associated to this membrane system. The spectral counting-based strategy was further

  7. AllergenOnline: A peer-reviewed, curated allergen database to assess novel food proteins for potential cross-reactivity.

    Science.gov (United States)

    Goodman, Richard E; Ebisawa, Motohiro; Ferreira, Fatima; Sampson, Hugh A; van Ree, Ronald; Vieths, Stefan; Baumert, Joseph L; Bohle, Barbara; Lalithambika, Sreedevi; Wise, John; Taylor, Steve L

    2016-05-01

    Increasingly regulators are demanding evaluation of potential allergenicity of foods prior to marketing. Primary risks are the transfer of allergens or potentially cross-reactive proteins into new foods. AllergenOnline was developed in 2005 as a peer-reviewed bioinformatics platform to evaluate risks of new dietary proteins in genetically modified organisms (GMO) and novel foods. The process used to identify suspected allergens and evaluate the evidence of allergenicity was refined between 2010 and 2015. Candidate proteins are identified from the NCBI database using keyword searches, the WHO/IUIS nomenclature database and peer reviewed publications. Criteria to classify proteins as allergens are described. Characteristics of the protein, the source and human subjects, test methods and results are evaluated by our expert panel and archived. Food, inhalant, salivary, venom, and contact allergens are included. Users access allergen sequences through links to the NCBI database and relevant references are listed online. Version 16 includes 1956 sequences from 778 taxonomic-protein groups that are accepted with evidence of allergic serum IgE-binding and/or biological activity. AllergenOnline provides a useful peer-reviewed tool for identifying the primary potential risks of allergy for GMOs and novel foods based on criteria described by the Codex Alimentarius Commission (2003). © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  8. Affiliation to the work market after curative treatment of head-and-neck cancer: a population-based study from the DAHANCA database.

    Science.gov (United States)

    Kjær, Trille; Bøje, Charlotte Rotbøl; Olsen, Maja Halgren; Overgaard, Jens; Johansen, Jørgen; Ibfelt, Else; Steding-Jessen, Marianne; Johansen, Christoffer; Dalton, Susanne O

    2013-02-01

    Survivors of squamous cell carcinoma of the head and neck (HNSCC) are more severely affected in regard to affiliation to the work market than other cancer survivors. Few studies have investigated associations between socioeconomic and disease-related factors and work market affiliation after curative treatment of HNSCC. We investigated the factors for early retirement pension due to disability and unemployment in patients who had been available for work one year before diagnosis. In a nationwide, population-based cohort study, data on 2436 HNSCC patients treated curatively in 1992-2008 were obtained from the Danish Head and Neck Cancer Group database and linked to Danish administrative population-based registries to obtain demographic and socioeconomic variables. We used multivariate logistic regression models to assess associations between socioeconomic factors (education, income and cohabitating status), cancer-specific variables such as tumour site and stage, comorbidity, early retirement pension and unemployment, with adjustment for age, gender and year of diagnosis. Short education [odds ratio (OR) 4.8; 95% confidence interval (CI) 2.2-10.4], low income (OR 3.2; 95% CI 1.8-5.8), living alone (OR 3.0; 95% CI 2.1-4.4) and having a Charlson comorbidity index score of 3 or more (OR 5.9; 95% CI 3.1-11) were significantly associated with early retirement overall and in all site groups. For the subgroup of patients who were employed before diagnosis, the risk pattern was similar. Tumour stage was not associated with early retirement or unemployment. Cancer-related factors were less strongly associated with early retirement and unemployment than socioeconomic factors and comorbidity. Clinicians treating HNSCC patients should be aware of the socioeconomic factors related to work market affiliation in order to provide more intensive social support or targeted rehabilitation for this patient group.

  9. BioM2MetDisease: a manually curated database for associations between microRNAs, metabolites, small molecules and metabolic diseases.

    Science.gov (United States)

    Xu, Yanjun; Yang, Haixiu; Wu, Tan; Dong, Qun; Sun, Zeguo; Shang, Desi; Li, Feng; Xu, Yingqi; Su, Fei; Liu, Siyao; Zhang, Yunpeng; Li, Xia

    2017-01-01

    BioM2MetDisease is a manually curated database that aims to provide a comprehensive and experimentally supported resource of associations between metabolic diseases and various biomolecules. Recently, metabolic diseases such as diabetes have become one of the leading threats to people’s health. Metabolic disease associated with alterations of multiple types of biomolecules such as miRNAs and metabolites. An integrated and high-quality data source that collection of metabolic disease associated biomolecules is essential for exploring the underlying molecular mechanisms and discovering novel therapeutics. Here, we developed the BioM2MetDisease database, which currently documents 2681 entries of relationships between 1147 biomolecules (miRNAs, metabolites and small molecules/drugs) and 78 metabolic diseases across 14 species. Each entry includes biomolecule category, species, biomolecule name, disease name, dysregulation pattern, experimental technique, a brief description of metabolic disease-biomolecule relationships, the reference, additional annotation information etc. BioM2MetDisease provides a user-friendly interface to explore and retrieve all data conveniently. A submission page was also offered for researchers to submit new associations between biomolecules and metabolic diseases. BioM2MetDisease provides a comprehensive resource for studying biology molecules act in metabolic diseases, and it is helpful for understanding the molecular mechanisms and developing novel therapeutics for metabolic diseases. http://www.bio-bigdata.com/BioM2MetDisease/. © The Author(s) 2017. Published by Oxford University Press.

  10. Learning lessons from Natech accidents - the eNATECH accident database

    Science.gov (United States)

    Krausmann, Elisabeth; Girgin, Serkan

    2016-04-01

    When natural hazards impact industrial facilities that house or process hazardous materials, fires, explosions and toxic releases can occur. This type of accident is commonly referred to as Natech accident. In order to prevent the recurrence of accidents or to better mitigate their consequences, lessons-learned type studies using available accident data are usually carried out. Through post-accident analysis, conclusions can be drawn on the most common damage and failure modes and hazmat release paths, particularly vulnerable storage and process equipment, and the hazardous materials most commonly involved in these types of accidents. These analyses also lend themselves to identifying technical and organisational risk-reduction measures that require improvement or are missing. Industrial accident databases are commonly used for retrieving sets of Natech accident case histories for further analysis. These databases contain accident data from the open literature, government authorities or in-company sources. The quality of reported information is not uniform and exhibits different levels of detail and accuracy. This is due to the difficulty of finding qualified information sources, especially in situations where accident reporting by the industry or by authorities is not compulsory, e.g. when spill quantities are below the reporting threshold. Data collection has then to rely on voluntary record keeping often by non-experts. The level of detail is particularly non-uniform for Natech accident data depending on whether the consequences of the Natech event were major or minor, and whether comprehensive information was available for reporting. In addition to the reporting bias towards high-consequence events, industrial accident databases frequently lack information on the severity of the triggering natural hazard, as well as on failure modes that led to the hazmat release. This makes it difficult to reconstruct the dynamics of the accident and renders the development of

  11. Lnc2Meth: a manually curated database of regulatory relationships between long non-coding RNAs and DNA methylation associated with human disease.

    Science.gov (United States)

    Zhi, Hui; Li, Xin; Wang, Peng; Gao, Yue; Gao, Baoqing; Zhou, Dianshuang; Zhang, Yan; Guo, Maoni; Yue, Ming; Shen, Weitao; Ning, Shangwei; Jin, Lianhong; Li, Xia

    2018-01-04

    Lnc2Meth (http://www.bio-bigdata.com/Lnc2Meth/), an interactive resource to identify regulatory relationships between human long non-coding RNAs (lncRNAs) and DNA methylation, is not only a manually curated collection and annotation of experimentally supported lncRNAs-DNA methylation associations but also a platform that effectively integrates tools for calculating and identifying the differentially methylated lncRNAs and protein-coding genes (PCGs) in diverse human diseases. The resource provides: (i) advanced search possibilities, e.g. retrieval of the database by searching the lncRNA symbol of interest, DNA methylation patterns, regulatory mechanisms and disease types; (ii) abundant computationally calculated DNA methylation array profiles for the lncRNAs and PCGs; (iii) the prognostic values for each hit transcript calculated from the patients clinical data; (iv) a genome browser to display the DNA methylation landscape of the lncRNA transcripts for a specific type of disease; (v) tools to re-annotate probes to lncRNA loci and identify the differential methylation patterns for lncRNAs and PCGs with user-supplied external datasets; (vi) an R package (LncDM) to complete the differentially methylated lncRNAs identification and visualization with local computers. Lnc2Meth provides a timely and valuable resource that can be applied to significantly expand our understanding of the regulatory relationships between lncRNAs and DNA methylation in various human diseases. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  12. tRNA sequence data, annotation data and curation data - tRNADB-CE | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available switchLanguage; BLAST Search Image Search Home About Archive Update History Data List Contact us tRNAD... tRNA sequence data, annotation data and curation data - tRNADB-CE | LSDB Archive ...

  13. How should the completeness and quality of curated nanomaterial data be evaluated?

    Science.gov (United States)

    Marchese Robinson, Richard L.; Lynch, Iseult; Peijnenburg, Willie; Rumble, John; Klaessig, Fred; Marquardt, Clarissa; Rauscher, Hubert; Puzyn, Tomasz; Purian, Ronit; Åberg, Christoffer; Karcher, Sandra; Vriens, Hanne; Hoet, Peter; Hoover, Mark D.; Hendren, Christine Ogilvie; Harper, Stacey L.

    2016-05-01

    Nanotechnology is of increasing significance. Curation of nanomaterial data into electronic databases offers opportunities to better understand and predict nanomaterials' behaviour. This supports innovation in, and regulation of, nanotechnology. It is commonly understood that curated data need to be sufficiently complete and of sufficient quality to serve their intended purpose. However, assessing data completeness and quality is non-trivial in general and is arguably especially difficult in the nanoscience area, given its highly multidisciplinary nature. The current article, part of the Nanomaterial Data Curation Initiative series, addresses how to assess the completeness and quality of (curated) nanomaterial data. In order to address this key challenge, a variety of related issues are discussed: the meaning and importance of data completeness and quality, existing approaches to their assessment and the key challenges associated with evaluating the completeness and quality of curated nanomaterial data. Considerations which are specific to the nanoscience area and lessons which can be learned from other relevant scientific disciplines are considered. Hence, the scope of this discussion ranges from physicochemical characterisation requirements for nanomaterials and interference of nanomaterials with nanotoxicology assays to broader issues such as minimum information checklists, toxicology data quality schemes and computational approaches that facilitate evaluation of the completeness and quality of (curated) data. This discussion is informed by a literature review and a survey of key nanomaterial data curation stakeholders. Finally, drawing upon this discussion, recommendations are presented concerning the central question: how should the completeness and quality of curated nanomaterial data be evaluated?Nanotechnology is of increasing significance. Curation of nanomaterial data into electronic databases offers opportunities to better understand and predict

  14. Improving the Acquisition and Management of Sample Curation Data

    Science.gov (United States)

    Todd, Nancy S.; Evans, Cindy A.; Labasse, Dan

    2011-01-01

    This paper discusses the current sample documentation processes used during and after a mission, examines the challenges and special considerations needed for designing effective sample curation data systems, and looks at the results of a simulated sample result mission and the lessons learned from this simulation. In addition, it introduces a new data architecture for an integrated sample Curation data system being implemented at the NASA Astromaterials Acquisition and Curation department and discusses how it improves on existing data management systems.

  15. Lessons from an enterprise-wide technical and administrative database using CASE and GUI front-ends

    International Nuclear Information System (INIS)

    Chan, A.; Crane, G.; MacGregor, I.; Meyer, S.

    1995-07-01

    An enterprise-wide database built via Oracle*CASE is a hot topic. The authors describe the PEP-II/BABAR Project-Wide Database, and the lessons learned in delivering and developing this system with a small team averaging two and one half people. They also give some details of providing World Wide Web (WWW) access to the information, and using Oracle*CASE and Oracle Forms4. The B Factory at the Stanford Linear Accelerator Center (SLAC) is a project built to study the physics of matter and anti-matter. It consists of two accelerator storage rings (PEP-II) and a detector (BABAR)--a project of approximately $250 million with collaboration by many labs worldwide. Foremost among these lessons is that the support and vision of management are key to the successful design and implementation of an enterprise-wide database. The authors faced the challenge of integrating both administrative and technical data into one CASE enterprise design. The goal, defined at the project's inception in late 1992, was to use a central database as a tool for the collaborating labs to: (1) track quality assurance during construction of the accelerator storage rings and detectors; (2) track down problems faster when they develop; and (3) facilitate the construction process. The focus of the project database, therefore, is on technical data which is less well-defined than administrative data

  16. METHODS OF CONTENTS CURATOR

    Directory of Open Access Journals (Sweden)

    V. Kukharenko

    2013-03-01

    Full Text Available Content curated - a new activity (started in 2008 qualified network users with process large amounts of information to represent her social network users. To prepare content curators developed 7 weeks distance course, which examines the functions, methods and tools curator. Courses showed a significant relationship success learning on the availability of advanced personal learning environment and the ability to process and analyze information.

  17. Databases

    Digital Repository Service at National Institute of Oceanography (India)

    Kunte, P.D.

    Information on bibliographic as well as numeric/textual databases relevant to coastal geomorphology has been included in a tabular form. Databases cover a broad spectrum of related subjects like coastal environment and population aspects, coastline...

  18. Meeting Curation Challenges in a Neuroimaging Group

    Directory of Open Access Journals (Sweden)

    Angus Whyte

    2008-08-01

    Full Text Available The SCARP project is a series of short studies with two aims; firstly to discover more about disciplinary approaches and attitudes to digital curation through ‘immersion’ in selected cases; secondly to apply known good practice, and where possible, identify new lessons from practice in the selected discipline areas. The study summarised here is of the Neuroimaging Group in the University of Edinburgh’s Division of Psychiatry, which plays a leading role in eScience collaborations to improve the infrastructure for neuroimaging data integration and reuse. The Group also aims to address growing data storage and curation needs, given the capabilities afforded by new infrastructure. The study briefly reviews the policy context and current challenges to data integration and sharing in the neuroimaging field. It then describes how curation and preservation risks and opportunities for change were identified throughout the curation lifecycle; and their context appreciated through field study in the research site. The results are consistent with studies of neuroimaging eInfrastructure that emphasise the role of local data sharing and reuse practices. These sustain mutual awareness of datasets and experimental protocols through sharing peer to peer, and among senior researchers and students, enabling continuity in research and flexibility in project work. This “human infrastructure” is taken into account in considering next steps for curation and preservation of the Group’s datasets and a phased approach to supporting data documentation.

  19. Databases

    Directory of Open Access Journals (Sweden)

    Nick Ryan

    2004-01-01

    Full Text Available Databases are deeply embedded in archaeology, underpinning and supporting many aspects of the subject. However, as well as providing a means for storing, retrieving and modifying data, databases themselves must be a result of a detailed analysis and design process. This article looks at this process, and shows how the characteristics of data models affect the process of database design and implementation. The impact of the Internet on the development of databases is examined, and the article concludes with a discussion of a range of issues associated with the recording and management of archaeological data.

  20. An Integrated Database of Unit Training Performance: Description an Lessons Learned

    National Research Council Canada - National Science Library

    Leibrecht, Bruce

    1997-01-01

    The Army Research Institute (ARI) has developed a prototype relational database for processing and archiving unit performance data from home station, training area, simulation based, and Combat Training Center training exercises...

  1. Lessons learned while building the Deepwater Horizon Database: Toward improved data sharing in coastal science

    Science.gov (United States)

    Thessen, Anne E.; McGinnis, Sean; North, Elizabeth W.

    2016-02-01

    Process studies and coupled-model validation efforts in geosciences often require integration of multiple data types across time and space. For example, improved prediction of hydrocarbon fate and transport is an important societal need which fundamentally relies upon synthesis of oceanography and hydrocarbon chemistry. Yet, there are no publically accessible databases which integrate these diverse data types in a georeferenced format, nor are there guidelines for developing such a database. The objective of this research was to analyze the process of building one such database to provide baseline information on data sources and data sharing and to document the challenges and solutions that arose during this major undertaking. The resulting Deepwater Horizon Database was approximately 2.4 GB in size and contained over 8 million georeferenced data points collected from industry, government databases, volunteer networks, and individual researchers. The major technical challenges that were overcome were reconciliation of terms, units, and quality flags which were necessary to effectively integrate the disparate data sets. Assembling this database required the development of relationships with individual researchers and data managers which often involved extensive e-mail contacts. The average number of emails exchanged per data set was 7.8. Of the 95 relevant data sets that were discovered, 38 (40%) were obtained, either in whole or in part. Over one third (36%) of the requests for data went unanswered. The majority of responses were received after the first request (64%) and within the first week of the first request (67%). Although fewer than half of the potentially relevant datasets were incorporated into the database, the level of sharing (40%) was high compared to some other disciplines where sharing can be as low as 10%. Our suggestions for building integrated databases include budgeting significant time for e-mail exchanges, being cognizant of the cost versus

  2. The curation of genetic variants: difficulties and possible solutions.

    Science.gov (United States)

    Pandey, Kapil Raj; Maden, Narendra; Poudel, Barsha; Pradhananga, Sailendra; Sharma, Amit Kumar

    2012-12-01

    The curation of genetic variants from biomedical articles is required for various clinical and research purposes. Nowadays, establishment of variant databases that include overall information about variants is becoming quite popular. These databases have immense utility, serving as a user-friendly information storehouse of variants for information seekers. While manual curation is the gold standard method for curation of variants, it can turn out to be time-consuming on a large scale thus necessitating the need for automation. Curation of variants described in biomedical literature may not be straightforward mainly due to various nomenclature and expression issues. Though current trends in paper writing on variants is inclined to the standard nomenclature such that variants can easily be retrieved, we have a massive store of variants in the literature that are present as non-standard names and the online search engines that are predominantly used may not be capable of finding them. For effective curation of variants, knowledge about the overall process of curation, nature and types of difficulties in curation, and ways to tackle the difficulties during the task are crucial. Only by effective curation, can variants be correctly interpreted. This paper presents the process and difficulties of curation of genetic variants with possible solutions and suggestions from our work experience in the field including literature support. The paper also highlights aspects of interpretation of genetic variants and the importance of writing papers on variants following standard and retrievable methods. Copyright © 2012. Published by Elsevier Ltd.

  3. Constructing Data Curation Profiles

    Directory of Open Access Journals (Sweden)

    Michael Witt

    2009-12-01

    Full Text Available This paper presents a brief literature review and then introduces the methods, design, and construction of the Data Curation Profile, an instrument that can be used to provide detailed information on particular data forms that might be curated by an academic library. These data forms are presented in the context of the related sub-disciplinary research area, and they provide the flow of the research process from which these data are generated. The profiles also represent the needs for data curation from the perspective of the data producers, using their own language. As such, they support the exploration of data curation across different research domains in real and practical terms. With the sponsorship of the Institute of Museum and Library Services, investigators from Purdue University and the University of Illinois interviewed 19 faculty subjects to identify needs for discovery, access, preservation, and reuse of their research data. For each subject, a profile was constructed that includes information about his or her general research, data forms and stages, value of data, data ingest, intellectual property, organization and description of data, tools, interoperability, impact and prestige, data management, and preservation. Each profile also presents a specific dataset supplied by the subject to serve as a concrete example. The Data Curation Profiles are being published to a public wiki for questions and discussion, and a blank template will be disseminated with guidelines for others to create and share their own profiles. This study was conducted primarily from the viewpoint of librarians interacting with faculty researchers; however, it is expected that these findings will complement a wide variety of data curation research and practice outside of librarianship and the university environment.

  4. Searching Harvard Business Review Online. . . Lessons in Searching a Full Text Database.

    Science.gov (United States)

    Tenopir, Carol

    1985-01-01

    This article examines the Harvard Business Review Online (HBRO) database (bibliographic description fields, abstracts, extracted information, full text, subject descriptors) and reports on 31 sample HBRO searches conducted in Bibliographic Retrieval Services to test differences between searching full text and searching bibliographic record. Sample…

  5. Can we replace curation with information extraction software?

    Science.gov (United States)

    Karp, Peter D

    2016-01-01

    Can we use programs for automated or semi-automated information extraction from scientific texts as practical alternatives to professional curation? I show that error rates of current information extraction programs are too high to replace professional curation today. Furthermore, current IEP programs extract single narrow slivers of information, such as individual protein interactions; they cannot extract the large breadth of information extracted by professional curators for databases such as EcoCyc. They also cannot arbitrate among conflicting statements in the literature as curators can. Therefore, funding agencies should not hobble the curation efforts of existing databases on the assumption that a problem that has stymied Artificial Intelligence researchers for more than 60 years will be solved tomorrow. Semi-automated extraction techniques appear to have significantly more potential based on a review of recent tools that enhance curator productivity. But a full cost-benefit analysis for these tools is lacking. Without such analysis it is possible to expend significant effort developing information-extraction tools that automate small parts of the overall curation workflow without achieving a significant decrease in curation costs.Database URL. © The Author(s) 2016. Published by Oxford University Press.

  6. Experience and Lessons learnt from running High Availability Databases on Network Attached Storage

    CERN Document Server

    Guijarro, Manuel

    2008-01-01

    The Database and Engineering Services Group of CERN's Information Technology Department supplies the Oracle Central Database services used in many activities at CERN. In order to provide High Availability and ease management for those services, a NAS (Network Attached Storage) based infrastructure has been setup. It runs several instances of the Oracle RAC (Real Application Cluster) using NFS (Network File System) as shared disk space for RAC purposes and Data hosting. It is composed of two private LANs (Local Area Network), one to provide access to the NAS filers and a second to implement the Oracle RAC private interconnect, both using Network Bonding. NAS filers are configured in partnership to prevent having single points of failure and to provide automatic NAS filer fail-over.

  7. Experience and lessons learnt from running high availability databases on network attached storage

    International Nuclear Information System (INIS)

    Guijarro, M; Gaspar, R

    2008-01-01

    The Database and Engineering Services Group of CERN's Information Technology Department supplies the Oracle Central Database services used in many activities at CERN. In order to provide High Availability and ease management for those services, a NAS (Network Attached Storage) based infrastructure has been setup. It runs several instances of the Oracle RAC (Real Application Cluster) using NFS (Network File System) as shared disk space for RAC purposes and Data hosting. It is composed of two private LANs (Local Area Network), one to provide access to the NAS filers and a second to implement the Oracle RAC private interconnect, both using Network Bonding. NAS filers are configured in partnership to prevent having single points of failure and to provide automatic NAS filer fail-over

  8. DMPD: Infectious non-self recognition in invertebrates: lessons from Drosophila andother insect models. [Dynamic Macrophage Pathway CSML Database

    Lifescience Database Archive (English)

    Full Text Available 15476918 Infectious non-self recognition in invertebrates: lessons from Drosophila ...fectious non-self recognition in invertebrates: lessons from Drosophila andother insect models. PubmedID 154...76918 Title Infectious non-self recognition in invertebrates: lessons from Drosop

  9. Experience and Lessons learnt from running high availability databases on Network Attached Storage

    CERN Document Server

    Guijarro, Juan Manuel; Segura Chinchilla, Nilo

    2008-01-01

    The Database and Engineering Services Group of CERN's Information Technology Department provides the Oracle based Central Data Base services used in many activities at CERN. In order to provide High Availability and ease management for those services, a NAS (Network Attached Storage) based infrastructure has been set up. It runs several instances of the Oracle RAC (Real Application Cluster) using NFS as share disk space for RAC purposes and Data hosting. It is composed of two private LAN's to provide access to the NAS file servers and Oracle RAC interconnect, both using network bonding. NAS nodes are configured in partnership to prevent having single points of failure and to provide automatic NAS fail-over. This presentation describes that infrastructure and gives some advice on how to automate its management and setup using a Fabric Management framework such as Quattor. It also covers aspects related with NAS Performance and Monitoring as well Data Backup and Archive of such facility using already existing i...

  10. XTCE and XML Database Evolution and Lessons from JWST, LandSat, and Constellation

    Science.gov (United States)

    Gal-Edd, Jonathan; Kreistle, Steven; Fatig. Cirtos; Jones, Ronald

    2008-01-01

    The database organizations within three different NASA projects have advanced current practices by creating database synergy between the various spacecraft life cycle stakeholders and educating users in the benefits of the Consultative Committee for Space Data Systems (CCSDS) XML Telemetry and Command Exchange (XTCE) format. The combination of XML for managing program data and CCSDS XTCE for exchange is a robust approach that will meet all user requirements using Standards and Non proprietary tools. COTS tools for XTCEKML are very wide and varied. To combine together various low cost and free tools can be more expensive in the long run than choosing a more expensive COTS tool that meets all the needs. This was especially important when deploying in 32 remote sites with no need for licenses. A common mission XTCEKML format between dissimilar systems is possible and is not difficult. Command XMLKTCE is more complex than telemetry and the use of XTCEKML metadata to describe pages and scripts is needed due to the proprietary nature of most current ground systems. Other mission and science products such as spacecraft loads, science image catalogs, and mission operation procedures can all be described with XML as well to increase there flexibility as systems evolve and change. Figure 10 is an example of a spacecraft table load. The word is out and the XTCE community is growing, The f sXt TCE user group was held in October and in addition to ESAESOC, SC02000, and CNES identified several systems based on XTCE. The second XTCE user group is scheduled for March 10, 2008 with LDMC and others joining. As the experience with XTCE grows and the user community receives the promised benefits of using XTCE and XML the interest is growing fast.

  11. Screen Practice in Curating

    DEFF Research Database (Denmark)

    Toft, Tanya Søndergaard

    2014-01-01

    During the past one and a half decade, a curatorial orientation towards "screen practice" has expanded the moving image and digital art into the public domain, exploring alternative artistic uses of the screen. The emergence of urban LED screens in the late 1990s provided a new venue that allowed...... for digital art to expand into public space. It also offered a political point of departure, inviting for confrontation with the Spectacle and with the politics and ideology of the screen as a mass communication medium that instrumentalized spectator positions. In this article I propose that screen practice...... to the dispositif of screen practice in curating, resulting in a medium-based curatorial discourse. With reference to the nomadic exhibition project Nordic Outbreak that I co-curated with Nina Colosi in 2013 and 2014, I suggest that the topos of the defined visual display area, frequently still known as "the screen...

  12. Curating Gothic Nightmares

    Directory of Open Access Journals (Sweden)

    Heather Tilley

    2007-10-01

    Full Text Available This review takes the occasion of a workshop given by Martin Myrone, curator of Gothic Nightmares: Fuseli, Blake, and the Romantic Imagination (Tate Britain, 2006 as a starting point to reflect on the practice of curating, and its relation to questions of the verbal and the visual in contemporary art historical practice. The exhibition prompted an engagement with questions of the genre of Gothic, through a dramatic display of the differences between ‘the Gothic' in literature and ‘the Gothic' in the visual arts within eighteenth- and early nineteenth-century culture. I also address the various ways in which 'the Gothic' was interpreted and reinscribed by visitors, especially those who dressed up for the exhibition. Finally, I consider some of the show's ‘marginalia' (specifically the catalogue, exploring the ways in which these extra events and texts shaped, and continue to shape, the cultural effect of the exhibition.

  13. The DCC Curation Lifecycle Model

    Directory of Open Access Journals (Sweden)

    Sarah Higgins

    2008-08-01

    Full Text Available Lifecycle management of digital materials is necessary to ensure their continuity. The DCC Curation Lifecycle Model has been developed as a generic, curation-specific, tool which can be used, in conjunction with relevant standards, to plan curation and preservation activities to different levels of granularity. The DCC will use the model: as a training tool for data creators, data curators and data users; to organise and plan their resources; and to help organisations identify risks to their digital assets and plan management strategies for their successful curation.

  14. Qrator: A web-based curation tool for glycan structures

    Science.gov (United States)

    Eavenson, Matthew; Kochut, Krys J; Miller, John A; Ranzinger, René; Tiemeyer, Michael; Aoki, Kazuhiro; York, William S

    2015-01-01

    Most currently available glycan structure databases use their own proprietary structure representation schema and contain numerous annotation errors. These cause problems when glycan databases are used for the annotation or mining of data generated in the laboratory. Due to the complexity of glycan structures, curating these databases is often a tedious and labor-intensive process. However, rigorously validating glycan structures can be made easier with a curation workflow that incorporates a structure-matching algorithm that compares candidate glycans to a canonical tree that embodies structural features consistent with established mechanisms for the biosynthesis of a particular class of glycans. To this end, we have implemented Qrator, a web-based application that uses a combination of external literature and database references, user annotations and canonical trees to assist and guide researchers in making informed decisions while curating glycans. Using this application, we have started the curation of large numbers of N-glycans, O-glycans and glycosphingolipids. Our curation workflow allows creating and extending canonical trees for these classes of glycans, which have subsequently been used to improve the curation workflow. PMID:25165068

  15. Research resources: curating the new eagle-i discovery system

    Science.gov (United States)

    Vasilevsky, Nicole; Johnson, Tenille; Corday, Karen; Torniai, Carlo; Brush, Matthew; Segerdell, Erik; Wilson, Melanie; Shaffer, Chris; Robinson, David; Haendel, Melissa

    2012-01-01

    Development of biocuration processes and guidelines for new data types or projects is a challenging task. Each project finds its way toward defining annotation standards and ensuring data consistency with varying degrees of planning and different tools to support and/or report on consistency. Further, this process may be data type specific even within the context of a single project. This article describes our experiences with eagle-i, a 2-year pilot project to develop a federated network of data repositories in which unpublished, unshared or otherwise ‘invisible’ scientific resources could be inventoried and made accessible to the scientific community. During the course of eagle-i development, the main challenges we experienced related to the difficulty of collecting and curating data while the system and the data model were simultaneously built, and a deficiency and diversity of data management strategies in the laboratories from which the source data was obtained. We discuss our approach to biocuration and the importance of improving information management strategies to the research process, specifically with regard to the inventorying and usage of research resources. Finally, we highlight the commonalities and differences between eagle-i and similar efforts with the hope that our lessons learned will assist other biocuration endeavors. Database URL: www.eagle-i.net PMID:22434835

  16. Data Curation Education in Research Centers (DCERC)

    Science.gov (United States)

    Marlino, M. R.; Mayernik, M. S.; Kelly, K.; Allard, S.; Tenopir, C.; Palmer, C.; Varvel, V. E., Jr.

    2012-12-01

    Digital data both enable and constrain scientific research. Scientists are enabled by digital data to develop new research methods, utilize new data sources, and investigate new topics, but they also face new data collection, management, and preservation burdens. The current data workforce consists primarily of scientists who receive little formal training in data management and data managers who are typically educated through on-the-job training. The Data Curation Education in Research Centers (DCERC) program is investigating a new model for educating data professionals to contribute to scientific research. DCERC is a collaboration between the University of Illinois at Urbana-Champaign Graduate School of Library and Information Science, the University of Tennessee School of Information Sciences, and the National Center for Atmospheric Research. The program is organized around a foundations course in data curation and provides field experiences in research and data centers for both master's and doctoral students. This presentation will outline the aims and the structure of the DCERC program and discuss results and lessons learned from the first set of summer internships in 2012. Four masters students participated and worked with both data mentors and science mentors, gaining first hand experiences in the issues, methods, and challenges of scientific data curation. They engaged in a diverse set of topics, including climate model metadata, observational data management workflows, and data cleaning, documentation, and ingest processes within a data archive. The students learned current data management practices and challenges while developing expertise and conducting research. They also made important contributions to NCAR data and science teams by evaluating data management workflows and processes, preparing data sets to be archived, and developing recommendations for particular data management activities. The master's student interns will return in summer of 2013

  17. Curating NASA's Past, Present, and Future Astromaterial Sample Collections

    Science.gov (United States)

    Zeigler, R. A.; Allton, J. H.; Evans, C. A.; Fries, M. D.; McCubbin, F. M.; Nakamura-Messenger, K.; Righter, K.; Zolensky, M.; Stansbery, E. K.

    2016-01-01

    The Astromaterials Acquisition and Curation Office at NASA Johnson Space Center (hereafter JSC curation) is responsible for curating all of NASA's extraterrestrial samples. JSC presently curates 9 different astromaterials collections in seven different clean-room suites: (1) Apollo Samples (ISO (International Standards Organization) class 6 + 7); (2) Antarctic Meteorites (ISO 6 + 7); (3) Cosmic Dust Particles (ISO 5); (4) Microparticle Impact Collection (ISO 7; formerly called Space-Exposed Hardware); (5) Genesis Solar Wind Atoms (ISO 4); (6) Stardust Comet Particles (ISO 5); (7) Stardust Interstellar Particles (ISO 5); (8) Hayabusa Asteroid Particles (ISO 5); (9) OSIRIS-REx Spacecraft Coupons and Witness Plates (ISO 7). Additional cleanrooms are currently being planned to house samples from two new collections, Hayabusa 2 (2021) and OSIRIS-REx (2023). In addition to the labs that house the samples, we maintain a wide variety of infra-structure facilities required to support the clean rooms: HEPA-filtered air-handling systems, ultrapure dry gaseous nitrogen systems, an ultrapure water system, and cleaning facilities to provide clean tools and equipment for the labs. We also have sample preparation facilities for making thin sections, microtome sections, and even focused ion-beam sections. We routinely monitor the cleanliness of our clean rooms and infrastructure systems, including measurements of inorganic or organic contamination, weekly airborne particle counts, compositional and isotopic monitoring of liquid N2 deliveries, and daily UPW system monitoring. In addition to the physical maintenance of the samples, we track within our databases the current and ever changing characteristics (weight, location, etc.) of more than 250,000 individually numbered samples across our various collections, as well as more than 100,000 images, and countless "analog" records that record the sample processing records of each individual sample. JSC Curation is co-located with JSC

  18. Scaling drug indication curation through crowdsourcing.

    Science.gov (United States)

    Khare, Ritu; Burger, John D; Aberdeen, John S; Tresner-Kirsch, David W; Corrales, Theodore J; Hirchman, Lynette; Lu, Zhiyong

    2015-01-01

    Motivated by the high cost of human curation of biological databases, there is an increasing interest in using computational approaches to assist human curators and accelerate the manual curation process. Towards the goal of cataloging drug indications from FDA drug labels, we recently developed LabeledIn, a human-curated drug indication resource for 250 clinical drugs. Its development required over 40 h of human effort across 20 weeks, despite using well-defined annotation guidelines. In this study, we aim to investigate the feasibility of scaling drug indication annotation through a crowdsourcing technique where an unknown network of workers can be recruited through the technical environment of Amazon Mechanical Turk (MTurk). To translate the expert-curation task of cataloging indications into human intelligence tasks (HITs) suitable for the average workers on MTurk, we first simplify the complex task such that each HIT only involves a worker making a binary judgment of whether a highlighted disease, in context of a given drug label, is an indication. In addition, this study is novel in the crowdsourcing interface design where the annotation guidelines are encoded into user options. For evaluation, we assess the ability of our proposed method to achieve high-quality annotations in a time-efficient and cost-effective manner. We posted over 3000 HITs drawn from 706 drug labels on MTurk. Within 8 h of posting, we collected 18 775 judgments from 74 workers, and achieved an aggregated accuracy of 96% on 450 control HITs (where gold-standard answers are known), at a cost of $1.75 per drug label. On the basis of these results, we conclude that our crowdsourcing approach not only results in significant cost and time saving, but also leads to accuracy comparable to that of domain experts. Published by Oxford University Press 2015. This work is written by US Government employees and is in the public domain in the US.

  19. Curating the Poster

    DEFF Research Database (Denmark)

    Christensen, Line Hjorth

    2017-01-01

    Parallel to the primary functions performed by posters in the urban environment, we find a range of curatorial practices that tie the poster, a mass-produced graphic design media, to the museum institution. Yet little research has attempted to uncover the diverse subject of curatorial work...... and the process where posters created to live in a real-world environment are relocated in a museum. According to Peter Bil’ak (2006), it creates a situation where ”the entire raison d’être of the work is lost as a side effect of losing the context of the work”. The article investigates how environmental...... structures can work as guidelines for curating posters and graphic design in a museum context. By applying an ecological view to design, specifically the semiotic notion “counter-ability”, it stresses the reciprocal relationship of humans and their built and product-designed environments. It further suggests...

  20. Database Description - Arabidopsis Phenome Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Arabidopsis Phenome Database Database Description General information of database Database n... BioResource Center Hiroshi Masuya Database classification Plant databases - Arabidopsis thaliana Organism T...axonomy Name: Arabidopsis thaliana Taxonomy ID: 3702 Database description The Arabidopsis thaliana phenome i...heir effective application. We developed the new Arabidopsis Phenome Database integrating two novel database...seful materials for their experimental research. The other, the “Database of Curated Plant Phenome” focusing

  1. Integration of data: the Nanomaterial Registry project and data curation

    International Nuclear Information System (INIS)

    Guzan, K A; Mills, K C; Gupta, V; Murry, D; Ostraat, M L; Scheier, C N; Willis, D A

    2013-01-01

    Due to the use of nanomaterials in multiple fields of applied science and technology, there is a need for accelerated understanding of any potential implications of using these unique and promising materials. There is a multitude of research data that, if integrated, can be leveraged to drive toward a better understanding. Integration can be achieved by applying nanoinformatics concepts. The Nanomaterial Registry is using applied minimal information about nanomaterials to support a robust data curation process in order to promote integration across a diverse data set. This paper describes the evolution of the curation methodology used in the Nanomaterial Registry project as well as the current procedure that is used. Some of the lessons learned about curation of nanomaterial data are also discussed. (paper)

  2. Text Mining to Support Gene Ontology Curation and Vice Versa.

    Science.gov (United States)

    Ruch, Patrick

    2017-01-01

    In this chapter, we explain how text mining can support the curation of molecular biology databases dealing with protein functions. We also show how curated data can play a disruptive role in the developments of text mining methods. We review a decade of efforts to improve the automatic assignment of Gene Ontology (GO) descriptors, the reference ontology for the characterization of genes and gene products. To illustrate the high potential of this approach, we compare the performances of an automatic text categorizer and show a large improvement of +225 % in both precision and recall on benchmarked data. We argue that automatic text categorization functions can ultimately be embedded into a Question-Answering (QA) system to answer questions related to protein functions. Because GO descriptors can be relatively long and specific, traditional QA systems cannot answer such questions. A new type of QA system, so-called Deep QA which uses machine learning methods trained with curated contents, is thus emerging. Finally, future advances of text mining instruments are directly dependent on the availability of high-quality annotated contents at every curation step. Databases workflows must start recording explicitly all the data they curate and ideally also some of the data they do not curate.

  3. Curation of complex, context-dependent immunological data

    Directory of Open Access Journals (Sweden)

    Sidney John

    2006-07-01

    Full Text Available Abstract Background The Immune Epitope Database and Analysis Resource (IEDB is dedicated to capturing, housing and analyzing complex immune epitope related data http://www.immuneepitope.org. Description To identify and extract relevant data from the scientific literature in an efficient and accurate manner, novel processes were developed for manual and semi-automated annotation. Conclusion Formalized curation strategies enable the processing of a large volume of context-dependent data, which are now available to the scientific community in an accessible and transparent format. The experiences described herein are applicable to other databases housing complex biological data and requiring a high level of curation expertise.

  4. Curating research data

    DEFF Research Database (Denmark)

    Nielsen, Hans Jørn; Hjørland, Birger

    2014-01-01

    libraries may be the best place to select, keep, organize and use research data. To prepare for this task, research libraries should be actively involved in domain-specific analytic studies of their respective domains. Originality/value – This paper offers a theoretical analysis and clarification......Purpose – A key issue in the literature about research libraries concerns their potential role in managing research data. The aim of this paper is to study the arguments for and against associating this task with libraries and the impact such an association would have on information professionals......, and consider the competitors to libraries in this field. Design/methodology/approach – This paper considers the nature of data and discusses data typologies, the kinds of data contained within databases and the implications of criticisms of the data-information-knowledge (DIK) hierarchy. It outlines the many...

  5. Lessons learned from process incident databases and the process safety incident database (PSID) approach sponsored by the Center for Chemical Process Safety

    International Nuclear Information System (INIS)

    Sepeda, Adrian L.

    2006-01-01

    Learning from the experiences of others has long been recognized as a valued and relatively painless process. In the world of process safety, this learning method is an essential tool since industry has neither the time and resources nor the willingness to experience an incident before taking corrective or preventative steps. This paper examines the need for and value of process safety incident databases that collect incidents of high learning value and structure them so that needed information can be easily and quickly extracted. It also explores how they might be used to prevent incidents by increasing awareness and by being a tool for conducting PHAs and incident investigations. The paper then discusses how the CCPS PSID meets those requirements, how PSID is structured and managed, and its attributes and features

  6. Teacher Training in Curative Education.

    Science.gov (United States)

    Juul, Kristen D.; Maier, Manfred

    1992-01-01

    This article considers the application of the philosophical and educational principles of Rudolf Steiner, called "anthroposophy," to the training of teachers and curative educators in the Waldorf schools. Special emphasis is on the Camphill movement which focuses on therapeutic schools and communities for children with special needs. (DB)

  7. Lessons learned bulletin

    International Nuclear Information System (INIS)

    1994-05-01

    During the past four years, the Department of Energy -- Savannah River Operations Office and the Westinghouse Savannah River Company (WSRC) Environmental Restoration (ER) Program completed various activities ranging from waste site investigations to closure and post closure projects. Critiques for lessons learned regarding project activities are performed at the completion of each project milestone, and this critique interval allows for frequent recognition of lessons learned. In addition to project related lessons learned, ER also performs lessons learned critiques. T'he Savannah River Site (SRS) also obtains lessons learned information from general industry, commercial nuclear industry, naval nuclear programs, and other DOE sites within the complex. Procedures are approved to administer the lessons learned program, and a database is available to catalog applicable lessons learned regarding environmental remediation, restoration, and administrative activities. ER will continue to use this database as a source of information available to SRS personnel

  8. Multi-centre, multi-database studies with common protocols : lessons learnt from the IMI PROTECT project

    NARCIS (Netherlands)

    Klungel, Olaf H; Kurz, Xavier; de Groot, Mark C H; Schlienger, Raymond G; Tcherny-Lessenot, Stephanie; Grimaldi, Lamiae; Ibáñez, Luisa; Groenwold, Rolf H H; Reynolds, Robert F

    2016-01-01

    PURPOSE: To assess the impact of a variety of methodological parameters on the association between six drug classes and five key adverse events in multiple databases. METHODS: The selection of Drug-Adverse Event pairs was based on public health impact, regulatory relevance, and the possibility to

  9. Enhancing Clinical Content and Race/Ethnicity Data in Statewide Hospital Administrative Databases: Obstacles Encountered, Strategies Adopted, and Lessons Learned.

    Science.gov (United States)

    Pine, Michael; Kowlessar, Niranjana M; Salemi, Jason L; Miyamura, Jill; Zingmond, David S; Katz, Nicole E; Schindler, Joe

    2015-08-01

    Eight grant teams used Agency for Healthcare Research and Quality infrastructure development research grants to enhance the clinical content of and improve race/ethnicity identifiers in statewide all-payer hospital administrative databases. Grantees faced common challenges, including recruiting data partners and ensuring their continued effective participation, acquiring and validating the accuracy and utility of new data elements, and linking data from multiple sources to create internally consistent enhanced administrative databases. Successful strategies to overcome these challenges included aggressively engaging with providers of critical sources of data, emphasizing potential benefits to participants, revising requirements to lessen burdens associated with participation, maintaining continuous communication with participants, being flexible when responding to participants' difficulties in meeting program requirements, and paying scrupulous attention to preparing data specifications and creating and implementing protocols for data auditing, validation, cleaning, editing, and linking. In addition to common challenges, grantees also had to contend with unique challenges from local environmental factors that shaped the strategies they adopted. The creation of enhanced administrative databases to support comparative effectiveness research is difficult, particularly in the face of numerous challenges with recruiting data partners such as competing demands on information technology resources. Excellent communication, flexibility, and attention to detail are essential ingredients in accomplishing this task. Additional research is needed to develop strategies for maintaining these databases when initial funding is exhausted. © Health Research and Educational Trust.

  10. Digital curation theory and practice

    CERN Document Server

    Hedges, Mark

    2016-01-01

    Digital curation is a multi-skilled profession with a key role to play not only in domains traditionally associated with the management of information, such as libraries and archives, but in a broad range of market sectors. Digital information is a defining feature of our age. As individuals we increasingly communicate and record our lives and memories in digital form, whether consciously or as a by-product of broader social, cultural and business activities. Throughout government and industry, there is a pressing need to manage complex information assets and to exploit their social, cultural and commercial value. This book addresses the key strategic, technical and practical issues around digital curation, curatorial practice, and locating the discussions within an appropriate theoretical context.

  11. Curative radiotherapy of supraglottic cancer

    International Nuclear Information System (INIS)

    Kim, Yong Ho; Chai, Gyu Young

    1998-01-01

    The purpose of this study was to evaluate the efficacy of curative radiotherapy in the management of supraglottic cancer. Twenty-one patients with squamous cell carcinoma of the supraglottis were treated with radiotherapy at Gyeongsang National University Hospital between 1990 and 1994. Median follow-up period was 36 months and 95% were observed for at least 2 years. Actuarial survival rate at 5 years was 39.3% for 21 patients. The 5-year actuarial survival rate was 75.0% in Stage I, 42.9% in Stage II, 33.3% in Stage III, and 28.6% in Stage IV(p=0.54). The 5-year local control rate was 52.0% for 21 patients. The 5-year local control rate was 75.0% in Stage I, 57.1% in Stage II, 66.7% in Stage III, and 28.6% in Stage IV(p=0.33). Double primary cancer was developed in 3 patients and those were all esophageal cancers. In early stage(Stage I and II) supraglottic cancer, curative radiotherapy would be a treatment of choice and surgery would be better to be reserved for salvage of radiotherapy failure. In advanced stage(Stage III and IV), radiotherapy alone is inadequate for curative therapy and combination with surgery should be done in operable patients. This report emphasizes the importance of esophagoscopy and esophagogram at the follow-up of patients with supraglottic cancer

  12. Outcomes research in amyotrophic lateral sclerosis: lessons learned from the amyotrophic lateral sclerosis clinical assessment, research, and education database.

    Science.gov (United States)

    Miller, Robert G; Anderson, Fred; Brooks, Benjamin Rix; Mitsumoto, Hiroshi; Bradley, Walter G; Ringel, Steven P

    2009-01-01

    To examine the care of patients with ALS following the publication of the standardized recommendations for the management of patients with amyotrophic lateral sclerosis (ALS) published in 1999 by the American Academy of Neurology. Specific aspects of ALS patient management have been evaluated serially using a national Amyotrophic Lateral Sclerosis Clinical Assessment, Research, and Education (ALS CARE) database to encourage compliance with these recommendations and to assure continuing quality improvement. The most recent analysis of 5,600 patients shows interesting epidemiological observations and treatment trends. Proper management of many ALS symptoms has increased substantially since the first publication of the guidelines, and awareness of pseudobulbar affect has increased. Other recommendations are underutilized: Only 9% undergo percutaneous endoscopic gastrostomy, although this procedure was recommended in 22% of patients; and noninvasive positive pressure ventilation was used by only 21% of patients despite being associated with improved 5-year survival rates. This observational database has been a useful tool in monitoring compliance with the standard of care for patients with ALS and may have resulted in greater adherence to guidelines.

  13. Curating Big Data Made Simple: Perspectives from Scientific Communities.

    Science.gov (United States)

    Sowe, Sulayman K; Zettsu, Koji

    2014-03-01

    The digital universe is exponentially producing an unprecedented volume of data that has brought benefits as well as fundamental challenges for enterprises and scientific communities alike. This trend is inherently exciting for the development and deployment of cloud platforms to support scientific communities curating big data. The excitement stems from the fact that scientists can now access and extract value from the big data corpus, establish relationships between bits and pieces of information from many types of data, and collaborate with a diverse community of researchers from various domains. However, despite these perceived benefits, to date, little attention is focused on the people or communities who are both beneficiaries and, at the same time, producers of big data. The technical challenges posed by big data are as big as understanding the dynamics of communities working with big data, whether scientific or otherwise. Furthermore, the big data era also means that big data platforms for data-intensive research must be designed in such a way that research scientists can easily search and find data for their research, upload and download datasets for onsite/offsite use, perform computations and analysis, share their findings and research experience, and seamlessly collaborate with their colleagues. In this article, we present the architecture and design of a cloud platform that meets some of these requirements, and a big data curation model that describes how a community of earth and environmental scientists is using the platform to curate data. Motivation for developing the platform, lessons learnt in overcoming some challenges associated with supporting scientists to curate big data, and future research directions are also presented.

  14. Identification and Removal of Contaminant Sequences From Ribosomal Gene Databases: Lessons From the Census of Deep Life.

    Science.gov (United States)

    Sheik, Cody S; Reese, Brandi Kiel; Twing, Katrina I; Sylvan, Jason B; Grim, Sharon L; Schrenk, Matthew O; Sogin, Mitchell L; Colwell, Frederick S

    2018-01-01

    Earth's subsurface environment is one of the largest, yet least studied, biomes on Earth, and many questions remain regarding what microorganisms are indigenous to the subsurface. Through the activity of the Census of Deep Life (CoDL) and the Deep Carbon Observatory, an open access 16S ribosomal RNA gene sequence database from diverse subsurface environments has been compiled. However, due to low quantities of biomass in the deep subsurface, the potential for incorporation of contaminants from reagents used during sample collection, processing, and/or sequencing is high. Thus, to understand the ecology of subsurface microorganisms (i.e., the distribution, richness, or survival), it is necessary to minimize, identify, and remove contaminant sequences that will skew the relative abundances of all taxa in the sample. In this meta-analysis, we identify putative contaminants associated with the CoDL dataset, recommend best practices for removing contaminants from samples, and propose a series of best practices for subsurface microbiology sampling. The most abundant putative contaminant genera observed, independent of evenness across samples, were Propionibacterium , Aquabacterium , Ralstonia , and Acinetobacter . While the top five most frequently observed genera were Pseudomonas , Propionibacterium , Acinetobacter , Ralstonia , and Sphingomonas . The majority of the most frequently observed genera (high evenness) were associated with reagent or potential human contamination. Additionally, in DNA extraction blanks, we observed potential archaeal contaminants, including methanogens, which have not been discussed in previous contamination studies. Such contaminants would directly affect the interpretation of subsurface molecular studies, as methanogenesis is an important subsurface biogeochemical process. Utilizing previously identified contaminant genera, we found that ∼27% of the total dataset were identified as contaminant sequences that likely originate from DNA

  15. Advances in Astromaterials Curation: Supporting Future Sample Return Missions

    Science.gov (United States)

    Evans, C. A.; Zeigler, R. A.; Fries, M. D..; Righter, K.; Allton, J. H.; Zolensky, M. E.; Calaway, M. J.; Bell, M. S.

    2015-01-01

    NASA's Astromaterials, curated at the Johnson Space Center in Houston, are the most extensive, best-documented, and leastcontaminated extraterrestrial samples that are provided to the worldwide research community. These samples include lunar samples from the Apollo missions, meteorites collected over nearly 40 years of expeditions to Antarctica (providing samples of dozens of asteroid bodies, the Moon, and Mars), Genesis solar wind samples, cosmic dust collected by NASA's high altitude airplanes, Comet Wild 2 and interstellar dust samples from the Stardust mission, and asteroid samples from JAXA's Hayabusa mission. A full account of NASA's curation efforts for these collections is provided by Allen, et al [1]. On average, we annually allocate about 1500 individual samples from NASA's astromaterials collections to hundreds of researchers from around the world, including graduate students and post-doctoral scientists; our allocation rate has roughly doubled over the past 10 years. The curation protocols developed for the lunar samples returned from the Apollo missions remain relevant and are adapted to new and future missions. Several lessons from the Apollo missions, including the need for early involvement of curation scientists in mission planning [1], have been applied to all subsequent sample return campaigns. From the 2013 National Academy of Sciences report [2]: "Curation is the critical interface between sample return missions and laboratory research. Proper curation has maintained the scientific integrity and utility of the Apollo, Antarctic meteorite, and cosmic dust collections for decades. Each of these collections continues to yield important new science. In the past decade, new state-of-the-art curatorial facilities for the Genesis and Stardust missions were key to the scientific breakthroughs provided by these missions." The results speak for themselves: research on NASA's astromaterials result in hundreds of papers annually, yield fundamental

  16. miRandola 2017: a curated knowledge base of non-invasive biomarkers

    DEFF Research Database (Denmark)

    Russo, Francesco; Di Bella, Sebastiano; Vannini, Federica

    2018-01-01

    databases. Data are manually curated from 314 articles that describe miRNAs, long non-coding RNAs and circular RNAs. Fourteen organisms are now included in the database, and associations of ncRNAs with 25 drugs, 47 sample types and 197 diseases. miRandola also classifies extracellular RNAs based...

  17. Immunisation in a curative setting

    DEFF Research Database (Denmark)

    Kofoed, Poul-Erik; Nielsen, B; Rahman, A K

    1990-01-01

    OBJECTIVE: To study the uptake of vaccination offered to women and children attending a curative health facility. DESIGN: Prospective survey over eight months of the uptake of vaccination offered to unimmunised women and children attending a diarrhoeal treatment centre as patients or attendants....... SETTING: The International Centre for Diarrhoeal Disease Research, Dhaka, Bangladesh. SUBJECTS: An estimated 19,349 unimmunised women aged 15 to 45 and 17,372 children attending the centre for treatment or accompanying patients between 1 January and 31 August 1989. MAIN OUTCOME MEASURES: The number...... of women and children who were unimmunised or incompletely immunised was calculated and the percentage of this target population accepting vaccination was recorded. RESULTS: 7530 (84.2%) Of 8944 eligible children and 7730 (40.4%) of 19,138 eligible women were vaccinated. Of the children, 63.8% were boys...

  18. Sharing and community curation of mass spectrometry data with GNPS

    Science.gov (United States)

    Nguyen, Don Duy; Watrous, Jeramie; Kapono, Clifford A; Luzzatto-Knaan, Tal; Porto, Carla; Bouslimani, Amina; Melnik, Alexey V; Meehan, Michael J; Liu, Wei-Ting; Crüsemann, Max; Boudreau, Paul D; Esquenazi, Eduardo; Sandoval-Calderón, Mario; Kersten, Roland D; Pace, Laura A; Quinn, Robert A; Duncan, Katherine R; Hsu, Cheng-Chih; Floros, Dimitrios J; Gavilan, Ronnie G; Kleigrewe, Karin; Northen, Trent; Dutton, Rachel J; Parrot, Delphine; Carlson, Erin E; Aigle, Bertrand; Michelsen, Charlotte F; Jelsbak, Lars; Sohlenkamp, Christian; Pevzner, Pavel; Edlund, Anna; McLean, Jeffrey; Piel, Jörn; Murphy, Brian T; Gerwick, Lena; Liaw, Chih-Chuang; Yang, Yu-Liang; Humpf, Hans-Ulrich; Maansson, Maria; Keyzers, Robert A; Sims, Amy C; Johnson, Andrew R.; Sidebottom, Ashley M; Sedio, Brian E; Klitgaard, Andreas; Larson, Charles B; P., Cristopher A Boya; Torres-Mendoza, Daniel; Gonzalez, David J; Silva, Denise B; Marques, Lucas M; Demarque, Daniel P; Pociute, Egle; O'Neill, Ellis C; Briand, Enora; Helfrich, Eric J. N.; Granatosky, Eve A; Glukhov, Evgenia; Ryffel, Florian; Houson, Hailey; Mohimani, Hosein; Kharbush, Jenan J; Zeng, Yi; Vorholt, Julia A; Kurita, Kenji L; Charusanti, Pep; McPhail, Kerry L; Nielsen, Kristian Fog; Vuong, Lisa; Elfeki, Maryam; Traxler, Matthew F; Engene, Niclas; Koyama, Nobuhiro; Vining, Oliver B; Baric, Ralph; Silva, Ricardo R; Mascuch, Samantha J; Tomasi, Sophie; Jenkins, Stefan; Macherla, Venkat; Hoffman, Thomas; Agarwal, Vinayak; Williams, Philip G; Dai, Jingqui; Neupane, Ram; Gurr, Joshua; Rodríguez, Andrés M. C.; Lamsa, Anne; Zhang, Chen; Dorrestein, Kathleen; Duggan, Brendan M; Almaliti, Jehad; Allard, Pierre-Marie; Phapale, Prasad; Nothias, Louis-Felix; Alexandrov, Theodore; Litaudon, Marc; Wolfender, Jean-Luc; Kyle, Jennifer E; Metz, Thomas O; Peryea, Tyler; Nguyen, Dac-Trung; VanLeer, Danielle; Shinn, Paul; Jadhav, Ajit; Müller, Rolf; Waters, Katrina M; Shi, Wenyuan; Liu, Xueting; Zhang, Lixin; Knight, Rob; Jensen, Paul R; Palsson, Bernhard O; Pogliano, Kit; Linington, Roger G; Gutiérrez, Marcelino; Lopes, Norberto P; Gerwick, William H; Moore, Bradley S; Dorrestein, Pieter C; Bandeira, Nuno

    2017-01-01

    The potential of the diverse chemistries present in natural products (NP) for biotechnology and medicine remains untapped because NP databases are not searchable with raw data and the NP community has no way to share data other than in published papers. Although mass spectrometry techniques are well-suited to high-throughput characterization of natural products, there is a pressing need for an infrastructure to enable sharing and curation of data. We present Global Natural Products Social molecular networking (GNPS, http://gnps.ucsd.edu), an open-access knowledge base for community wide organization and sharing of raw, processed or identified tandem mass (MS/MS) spectrometry data. In GNPS crowdsourced curation of freely available community-wide reference MS libraries will underpin improved annotations. Data-driven social-networking should facilitate identification of spectra and foster collaborations. We also introduce the concept of ‘living data’ through continuous reanalysis of deposited data. PMID:27504778

  19. Gene Name Thesaurus - Gene Name Thesaurus | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available 08/lsdba.nbdc00966-001 Description of data contents Curators who have expertize in biological research edit ...onym information fields in various gene/genome databases. 2. The curators who have expertise in biological research

  20. MET network in PubMed: a text-mined network visualization and curation system.

    Science.gov (United States)

    Dai, Hong-Jie; Su, Chu-Hsien; Lai, Po-Ting; Huang, Ming-Siang; Jonnagaddala, Jitendra; Rose Jue, Toni; Rao, Shruti; Chou, Hui-Jou; Milacic, Marija; Singh, Onkar; Syed-Abdul, Shabbir; Hsu, Wen-Lian

    2016-01-01

    Metastasis is the dissemination of a cancer/tumor from one organ to another, and it is the most dangerous stage during cancer progression, causing more than 90% of cancer deaths. Improving the understanding of the complicated cellular mechanisms underlying metastasis requires investigations of the signaling pathways. To this end, we developed a METastasis (MET) network visualization and curation tool to assist metastasis researchers retrieve network information of interest while browsing through the large volume of studies in PubMed. MET can recognize relations among genes, cancers, tissues and organs of metastasis mentioned in the literature through text-mining techniques, and then produce a visualization of all mined relations in a metastasis network. To facilitate the curation process, MET is developed as a browser extension that allows curators to review and edit concepts and relations related to metastasis directly in PubMed. PubMed users can also view the metastatic networks integrated from the large collection of research papers directly through MET. For the BioCreative 2015 interactive track (IAT), a curation task was proposed to curate metastatic networks among PubMed abstracts. Six curators participated in the proposed task and a post-IAT task, curating 963 unique metastatic relations from 174 PubMed abstracts using MET.Database URL: http://btm.tmu.edu.tw/metastasisway. © The Author(s) 2016. Published by Oxford University Press.

  1. LncRNAWiki: harnessing community knowledge in collaborative curation of human long non-coding RNAs

    KAUST Repository

    Ma, L.

    2014-11-15

    Long non-coding RNAs (lncRNAs) perform a diversity of functions in numerous important biological processes and are implicated in many human diseases. In this report we present lncRNAWiki (http://lncrna.big.ac.cn), a wiki-based platform that is open-content and publicly editable and aimed at community-based curation and collection of information on human lncRNAs. Current related databases are dependent primarily on curation by experts, making it laborious to annotate the exponentially accumulated information on lncRNAs, which inevitably requires collective efforts in community-based curation of lncRNAs. Unlike existing databases, lncRNAWiki features comprehensive integration of information on human lncRNAs obtained from multiple different resources and allows not only existing lncRNAs to be edited, updated and curated by different users but also the addition of newly identified lncRNAs by any user. It harnesses community collective knowledge in collecting, editing and annotating human lncRNAs and rewards community-curated efforts by providing explicit authorship based on quantified contributions. LncRNAWiki relies on the underling knowledge of scientific community for collective and collaborative curation of human lncRNAs and thus has the potential to serve as an up-to-date and comprehensive knowledgebase for human lncRNAs.

  2. Directly e-mailing authors of newly published papers encourages community curation

    Science.gov (United States)

    Bunt, Stephanie M.; Grumbling, Gary B.; Field, Helen I.; Marygold, Steven J.; Brown, Nicholas H.; Millburn, Gillian H.

    2012-01-01

    Much of the data within Model Organism Databases (MODs) comes from manual curation of the primary research literature. Given limited funding and an increasing density of published material, a significant challenge facing all MODs is how to efficiently and effectively prioritize the most relevant research papers for detailed curation. Here, we report recent improvements to the triaging process used by FlyBase. We describe an automated method to directly e-mail corresponding authors of new papers, requesting that they list the genes studied and indicate (‘flag’) the types of data described in the paper using an online tool. Based on the author-assigned flags, papers are then prioritized for detailed curation and channelled to appropriate curator teams for full data extraction. The overall response rate has been 44% and the flagging of data types by authors is sufficiently accurate for effective prioritization of papers. In summary, we have established a sustainable community curation program, with the result that FlyBase curators now spend less time triaging and can devote more effort to the specialized task of detailed data extraction. Database URL: http://flybase.org/ PMID:22554788

  3. Advanced Curation Activities at NASA: Preparation for Upcoming Missions

    Science.gov (United States)

    Fries, M. D.; Evans, C. A.; McCubbin, F. M.; Harrington, A. D.; Regberg, A. B.; Snead, C. J.; Zeigler, R. A.

    2017-07-01

    NASA Curation cares for NASA's astromaterials and performs advanced curation so as to improve current practices and prepare for future collections. Cold curation, microbial monitoring, contamination control/knowledge and other aspects are reviewed.

  4. Alfred Drury: The Artist as Curator

    Directory of Open Access Journals (Sweden)

    Ben Thomas

    2016-06-01

    Full Text Available This article presents a series of reflections on the experience of curating the exhibition ‘Alfred Drury and the New Sculpture’ in 2013. In particular, it charts the evolution of the design of the exhibition, notably its central tableau based on a photograph of the sculptor Alfred Drury’s studio in 1900. This photograph records a display of Drury’s works for visiting Australian patrons, and could be said to record evidence of the artist curating his own work. The legitimacy of deriving a curatorial approach from this photographic evidence is discussed, along with the broader problem of ‘historicizing’ approaches to curating.

  5. Curating the innate immunity interactome.

    LENUS (Irish Health Repository)

    Lynn, David J

    2010-01-01

    The innate immune response is the first line of defence against invading pathogens and is regulated by complex signalling and transcriptional networks. Systems biology approaches promise to shed new light on the regulation of innate immunity through the analysis and modelling of these networks. A key initial step in this process is the contextual cataloguing of the components of this system and the molecular interactions that comprise these networks. InnateDB (http:\\/\\/www.innatedb.com) is a molecular interaction and pathway database developed to facilitate systems-level analyses of innate immunity.

  6. A hybrid human and machine resource curation pipeline for the Neuroscience Information Framework.

    Science.gov (United States)

    Bandrowski, A E; Cachat, J; Li, Y; Müller, H M; Sternberg, P W; Ciccarese, P; Clark, T; Marenco, L; Wang, R; Astakhov, V; Grethe, J S; Martone, M E

    2012-01-01

    The breadth of information resources available to researchers on the Internet continues to expand, particularly in light of recently implemented data-sharing policies required by funding agencies. However, the nature of dense, multifaceted neuroscience data and the design of contemporary search engine systems makes efficient, reliable and relevant discovery of such information a significant challenge. This challenge is specifically pertinent for online databases, whose dynamic content is 'hidden' from search engines. The Neuroscience Information Framework (NIF; http://www.neuinfo.org) was funded by the NIH Blueprint for Neuroscience Research to address the problem of finding and utilizing neuroscience-relevant resources such as software tools, data sets, experimental animals and antibodies across the Internet. From the outset, NIF sought to provide an accounting of available resources, whereas developing technical solutions to finding, accessing and utilizing them. The curators therefore, are tasked with identifying and registering resources, examining data, writing configuration files to index and display data and keeping the contents current. In the initial phases of the project, all aspects of the registration and curation processes were manual. However, as the number of resources grew, manual curation became impractical. This report describes our experiences and successes with developing automated resource discovery and semiautomated type characterization with text-mining scripts that facilitate curation team efforts to discover, integrate and display new content. We also describe the DISCO framework, a suite of automated web services that significantly reduce manual curation efforts to periodically check for resource updates. Lastly, we discuss DOMEO, a semi-automated annotation tool that improves the discovery and curation of resources that are not necessarily website-based (i.e. reagents, software tools). Although the ultimate goal of automation was to

  7. Curating NASA's Past, Present, and Future Extraterrestrial Sample Collections

    Science.gov (United States)

    McCubbin, F. M.; Allton, J. H.; Evans, C. A.; Fries, M. D.; Nakamura-Messenger, K.; Righter, K.; Zeigler, R. A.; Zolensky, M.; Stansbery, E. K.

    2016-01-01

    The Astromaterials Acquisition and Curation Office (henceforth referred to herein as NASA Curation Office) at NASA Johnson Space Center (JSC) is responsible for curating all of NASA's extraterrestrial samples. Under the governing document, NASA Policy Directive (NPD) 7100.10E "Curation of Extraterrestrial Materials", JSC is charged with "...curation of all extra-terrestrial material under NASA control, including future NASA missions." The Directive goes on to define Curation as including "...documentation, preservation, preparation, and distribution of samples for research, education, and public outreach." Here we describe some of the past, present, and future activities of the NASA Curation Office.

  8. Curation Micro-Services: A Pipeline Metaphor for Repositories

    OpenAIRE

    Abrams, Stephen; Cruse, Patricia; Kunze, John; Minor, David

    2010-01-01

    The effective long-term curation of digital content requires expert analysis, policy setting, and decision making, and a robust technical infrastructure that can effect and enforce curation policies and implement appropriate curation activities. Since the number, size, and diversity of content under curation management will undoubtedly continue to grow over time, and the state of curation understanding and best practices relative to that content will undergo a similar constant evolution, one ...

  9. Mycobacteriophage genome database.

    Science.gov (United States)

    Joseph, Jerrine; Rajendran, Vasanthi; Hassan, Sameer; Kumar, Vanaja

    2011-01-01

    Mycobacteriophage genome database (MGDB) is an exclusive repository of the 64 completely sequenced mycobacteriophages with annotated information. It is a comprehensive compilation of the various gene parameters captured from several databases pooled together to empower mycobacteriophage researchers. The MGDB (Version No.1.0) comprises of 6086 genes from 64 mycobacteriophages classified into 72 families based on ACLAME database. Manual curation was aided by information available from public databases which was enriched further by analysis. Its web interface allows browsing as well as querying the classification. The main objective is to collect and organize the complexity inherent to mycobacteriophage protein classification in a rational way. The other objective is to browse the existing and new genomes and describe their functional annotation. The database is available for free at http://mpgdb.ibioinformatics.org/mpgdb.php.

  10. DataShare: Empowering Researcher Data Curation

    Directory of Open Access Journals (Sweden)

    Stephen Abrams

    2014-07-01

    Full Text Available Researchers are increasingly being asked to ensure that all products of research activity – not just traditional publications – are preserved and made widely available for study and reuse as a precondition for publication or grant funding, or to conform to disciplinary best practices. In order to conform to these requirements, scholars need effective, easy-to-use tools and services for the long-term curation of their research data. The DataShare service, developed at the University of California, is being used by researchers to: (1 prepare for curation by reviewing best practice recommendations for the acquisition or creation of digital research data; (2 select datasets using intuitive file browsing and drag-and-drop interfaces; (3 describe their data for enhanced discoverability in terms of the DataCite metadata schema; (4 preserve their data by uploading to a public access collection in the UC3 Merritt curation repository; (5 cite their data in terms of persistent and globally-resolvable DOI identifiers; (6 expose their data through registration with well-known abstracting and indexing services and major internet search engines; (7 control the dissemination of their data through enforceable data use agreements; and (8 discover and retrieve datasets of interest through a faceted search and browse environment. Since the widespread adoption of effective data management practices is highly dependent on ease of use and integration into existing individual, institutional, and disciplinary workflows, the emphasis throughout the design and implementation of DataShare is to provide the highest level of curation service with the lowest possible technical barriers to entry by individual researchers. By enabling intuitive, self-service access to data curation functions, DataShare helps to contribute to more widespread adoption of good data curation practices that are critical to open scientific inquiry, discourse, and advancement.

  11. Annotation of phenotypic diversity: decoupling data curation and ontology curation using Phenex.

    Science.gov (United States)

    Balhoff, James P; Dahdul, Wasila M; Dececchi, T Alexander; Lapp, Hilmar; Mabee, Paula M; Vision, Todd J

    2014-01-01

    Phenex (http://phenex.phenoscape.org/) is a desktop application for semantically annotating the phenotypic character matrix datasets common in evolutionary biology. Since its initial publication, we have added new features that address several major bottlenecks in the efficiency of the phenotype curation process: allowing curators during the data curation phase to provisionally request terms that are not yet available from a relevant ontology; supporting quality control against annotation guidelines to reduce later manual review and revision; and enabling the sharing of files for collaboration among curators. We decoupled data annotation from ontology development by creating an Ontology Request Broker (ORB) within Phenex. Curators can use the ORB to request a provisional term for use in data annotation; the provisional term can be automatically replaced with a permanent identifier once the term is added to an ontology. We added a set of annotation consistency checks to prevent common curation errors, reducing the need for later correction. We facilitated collaborative editing by improving the reliability of Phenex when used with online folder sharing services, via file change monitoring and continual autosave. With the addition of these new features, and in particular the Ontology Request Broker, Phenex users have been able to focus more effectively on data annotation. Phenoscape curators using Phenex have reported a smoother annotation workflow, with much reduced interruptions from ontology maintenance and file management issues.

  12. Protein-Protein Interaction Databases

    DEFF Research Database (Denmark)

    Szklarczyk, Damian; Jensen, Lars Juhl

    2015-01-01

    Years of meticulous curation of scientific literature and increasingly reliable computational predictions have resulted in creation of vast databases of protein interaction data. Over the years, these repositories have become a basic framework in which experiments are analyzed and new directions...

  13. AtomPy: an open atomic-data curation environment

    Science.gov (United States)

    Bautista, Manuel; Mendoza, Claudio; Boswell, Josiah S; Ajoku, Chukwuemeka

    2014-06-01

    We present a cloud-computing environment for atomic data curation, networking among atomic data providers and users, teaching-and-learning, and interfacing with spectral modeling software. The system is based on Google-Drive Sheets, Pandas (Python Data Analysis Library) DataFrames, and IPython Notebooks for open community-driven curation of atomic data for scientific and technological applications. The atomic model for each ionic species is contained in a multi-sheet Google-Drive workbook, where the atomic parameters from all known public sources are progressively stored. Metadata (provenance, community discussion, etc.) accompanying every entry in the database are stored through Notebooks. Education tools on the physics of atomic processes as well as their relevance to plasma and spectral modeling are based on IPython Notebooks that integrate written material, images, videos, and active computer-tool workflows. Data processing workflows and collaborative software developments are encouraged and managed through the GitHub social network. Relevant issues this platform intends to address are: (i) data quality by allowing open access to both data producers and users in order to attain completeness, accuracy, consistency, provenance and currentness; (ii) comparisons of different datasets to facilitate accuracy assessment; (iii) downloading to local data structures (i.e. Pandas DataFrames) for further manipulation and analysis by prospective users; and (iv) data preservation by avoiding the discard of outdated sets.

  14. A comprehensive curated resource for follicle stimulating hormone signaling

    Directory of Open Access Journals (Sweden)

    Sharma Jyoti

    2011-10-01

    Full Text Available Abstract Background Follicle stimulating hormone (FSH is an important hormone responsible for growth, maturation and function of the human reproductive system. FSH regulates the synthesis of steroid hormones such as estrogen and progesterone, proliferation and maturation of follicles in the ovary and spermatogenesis in the testes. FSH is a glycoprotein heterodimer that binds and acts through the FSH receptor, a G-protein coupled receptor. Although online pathway repositories provide information about G-protein coupled receptor mediated signal transduction, the signaling events initiated specifically by FSH are not cataloged in any public database in a detailed fashion. Findings We performed comprehensive curation of the published literature to identify the components of FSH signaling pathway and the molecular interactions that occur upon FSH receptor activation. Our effort yielded 64 reactions comprising 35 enzyme-substrate reactions, 11 molecular association events, 11 activation events and 7 protein translocation events that occur in response to FSH receptor activation. We also cataloged 265 genes, which were differentially expressed upon FSH stimulation in normal human reproductive tissues. Conclusions We anticipate that the information provided in this resource will provide better insights into the physiological role of FSH in reproductive biology, its signaling mediators and aid in further research in this area. The curated FSH pathway data is freely available through NetPath (http://www.netpath.org, a pathway resource developed previously by our group.

  15. Solubility Study of Curatives in Various Rubbers

    NARCIS (Netherlands)

    Guo, R.; Talma, Auke; Datta, Rabin; Dierkes, Wilma K.; Noordermeer, Jacobus W.M.

    2008-01-01

    The previous works on solubility of curatives in rubbers were mainly carried out in natural rubber. Not too much information available on dissimilar rubbers and this is important because most of the compounds today are blends of dissimilar rubbers. Although solubility can be expected to certain

  16. Curating Media Learning: Towards a Porous Expertise

    Science.gov (United States)

    McDougall, Julian; Potter, John

    2015-01-01

    This article combines research results from a range of projects with two consistent themes. Firstly, we explore the potential for curation to offer a productive metaphor for the convergence of digital media learning across and between home/lifeworld and formal educational/system-world spaces--or between the public and private spheres. Secondly, we…

  17. Curating and Nudging in Virtual CLIL Environments

    Science.gov (United States)

    Nielsen, Helle Lykke

    2014-01-01

    Foreign language teachers can benefit substantially from the notions of curation and nudging when scaffolding CLIL activities on the internet. This article shows how these principles can be integrated into CLILstore, a free multimedia-rich learning tool with seamless access to online dictionaries, and presents feedback from first and second year…

  18. Smart Mobility Stakeholders - Curating Urban Data & Models

    Energy Technology Data Exchange (ETDEWEB)

    Sperling, Joshua [National Renewable Energy Laboratory (NREL), Golden, CO (United States)

    2017-09-01

    This presentation provides an overview of the curation of urban data and models through engaging SMART mobility stakeholders. SMART Mobility Urban Science Efforts are helping to expose key data sets, models, and roles for the U.S. Department of Energy in engaging across stakeholders to ensure useful insights. This will help to support other Urban Science and broader SMART initiatives.

  19. Textpresso Central: a customizable platform for searching, text mining, viewing, and curating biomedical literature.

    Science.gov (United States)

    Müller, H-M; Van Auken, K M; Li, Y; Sternberg, P W

    2018-03-09

    The biomedical literature continues to grow at a rapid pace, making the challenge of knowledge retrieval and extraction ever greater. Tools that provide a means to search and mine the full text of literature thus represent an important way by which the efficiency of these processes can be improved. We describe the next generation of the Textpresso information retrieval system, Textpresso Central (TPC). TPC builds on the strengths of the original system by expanding the full text corpus to include the PubMed Central Open Access Subset (PMC OA), as well as the WormBase C. elegans bibliography. In addition, TPC allows users to create a customized corpus by uploading and processing documents of their choosing. TPC is UIMA compliant, to facilitate compatibility with external processing modules, and takes advantage of Lucene indexing and search technology for efficient handling of millions of full text documents. Like Textpresso, TPC searches can be performed using keywords and/or categories (semantically related groups of terms), but to provide better context for interpreting and validating queries, search results may now be viewed as highlighted passages in the context of full text. To facilitate biocuration efforts, TPC also allows users to select text spans from the full text and annotate them, create customized curation forms for any data type, and send resulting annotations to external curation databases. As an example of such a curation form, we describe integration of TPC with the Noctua curation tool developed by the Gene Ontology (GO) Consortium. Textpresso Central is an online literature search and curation platform that enables biocurators and biomedical researchers to search and mine the full text of literature by integrating keyword and category searches with viewing search results in the context of the full text. It also allows users to create customized curation interfaces, use those interfaces to make annotations linked to supporting evidence statements

  20. Morbidity of curative cancer surgery and suicide risk.

    Science.gov (United States)

    Jayakrishnan, Thejus T; Sekigami, Yurie; Rajeev, Rahul; Gamblin, T Clark; Turaga, Kiran K

    2017-11-01

    Curative cancer operations lead to debility and loss of autonomy in a population vulnerable to suicide death. The extent to which operative intervention impacts suicide risk is not well studied. To examine the effects of morbidity of curative cancer surgeries and prognosis of disease on the risk of suicide in patients with solid tumors. Retrospective cohort study using Surveillance, Epidemiology, and End Results data from 2004 to 2011; multilevel systematic review. General US population. Participants were 482 781 patients diagnosed with malignant neoplasm between 2004 and 2011 who underwent curative cancer surgeries. Death by suicide or self-inflicted injury. Among 482 781 patients that underwent curative cancer surgery, 231 committed suicide (16.58/100 000 person-years [95% confidence interval, CI, 14.54-18.82]). Factors significantly associated with suicide risk included male sex (incidence rate [IR], 27.62; 95% CI, 23.82-31.86) and age >65 years (IR, 22.54; 95% CI, 18.84-26.76). When stratified by 30-day overall postoperative morbidity, a significantly higher incidence of suicide was found for high-morbidity surgeries (IR, 33.30; 95% CI, 26.50-41.33) vs moderate morbidity (IR, 24.27; 95% CI, 18.92-30.69) and low morbidity (IR, 9.81; 95% CI, 7.90-12.04). Unit increase in morbidity was significantly associated with death by suicide (odds ratio, 1.01; 95% CI, 1.00-1.03; P = .02) and decreased suicide-specific survival (hazards ratio, 1.02; 95% CI, 1.00-1.03, P = .01) in prognosis-adjusted models. In this sample of cancer patients in the Surveillance, Epidemiology, and End Results database, patients that undergo high-morbidity surgeries appear most vulnerable to death by suicide. The identification of this high-risk cohort should motivate health care providers and particularly surgeons to adopt screening measures during the postoperative follow-up period for these patients. Copyright © 2016 John Wiley & Sons, Ltd.

  1. Agile Data Curation Case Studies Leading to the Identification and Development of Data Curation Design Patterns

    Science.gov (United States)

    Benedict, K. K.; Lenhardt, W. C.; Young, J. W.; Gordon, L. C.; Hughes, S.; Santhana Vannan, S. K.

    2017-12-01

    The planning for and development of efficient workflows for the creation, reuse, sharing, documentation, publication and preservation of research data is a general challenge that research teams of all sizes face. In response to: requirements from funding agencies for full-lifecycle data management plans that will result in well documented, preserved, and shared research data products increasing requirements from publishers for shared data in conjunction with submitted papers interdisciplinary research team's needs for efficient data sharing within projects, and increasing reuse of research data for replication and new, unanticipated research, policy development, and public use alternative strategies to traditional data life cycle approaches must be developed and shared that enable research teams to meet these requirements while meeting the core science objectives of their projects within the available resources. In support of achieving these goals, the concept of Agile Data Curation has been developed in which there have been parallel activities in support of 1) identifying a set of shared values and principles that underlie the objectives of agile data curation, 2) soliciting case studies from the Earth science and other research communities that illustrate aspects of what the contributors consider agile data curation methods and practices, and 3) identifying or developing design patterns that are high-level abstractions from successful data curation practice that are related to common data curation problems for which common solution strategies may be employed. This paper provides a collection of case studies that have been contributed by the Earth science community, and an initial analysis of those case studies to map them to emerging shared data curation problems and their potential solutions. Following the initial analysis of these problems and potential solutions, existing design patterns from software engineering and related disciplines are identified as a

  2. The baladi curative system of Cairo, Egypt.

    Science.gov (United States)

    Early, E A

    1988-03-01

    The article explores the symbolic structure of the baladi (traditional) cultural system as revealed in everyday narratives, with a focus on baladi curative action. The everyday illness narrative provides a cultural window to the principles of fluidity and restorative balance of baladi curative practices. The body is seen as a dynamic organism through which both foreign objects and physiological entities can move. The body should be in balance, as with any humorally-influenced system, and so baladi cures aim to restore normal balance and functioning of the body. The article examines in detail a narrative on treatment of a sick child, and another on treatment of fertility problems. It traces such cultural oppositions as insider: outsider; authentic:inauthentic; home remedy:cosmopolitan medicine. In the social as well as the medical arena these themes organize social/medical judgements about correct action and explanations of events.

  3. An emerging role: the nurse content curator.

    Science.gov (United States)

    Brooks, Beth A

    2015-01-01

    A new phenomenon, the inverted or "flipped" classroom, assumes that students are no longer acquiring knowledge exclusively through textbooks or lectures. Instead, they are seeking out the vast amount of free information available to them online (the very essence of open source) to supplement learning gleaned in textbooks and lectures. With so much open-source content available to nursing faculty, it benefits the faculty to use readily available, technologically advanced content. The nurse content curator supports nursing faculty in its use of such content. Even more importantly, the highly paid, time-strapped faculty is not spending an inordinate amount of effort surfing for and evaluating content. The nurse content curator does that work, while the faculty uses its time more effectively to help students vet the truth, make meaning of the content, and learn to problem-solve. Brooks. © 2014 Wiley Periodicals, Inc.

  4. Database Description - tRNADB-CE | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available switchLanguage; BLAST Search Image Search Home About Archive Update History Data List Contact us tRNAD...B-CE Database Description General information of database Database name tRNADB-CE Alter...CC BY-SA Detail Background and funding Name: MEXT Integrated Database Project Reference(s) Article title: tRNAD... 2009 Jan;37(Database issue):D163-8. External Links: Article title: tRNADB-CE 2011: tRNA gene database curat...n Download License Update History of This Database Site Policy | Contact Us Database Description - tRNADB-CE | LSDB Archive ...

  5. Curating and nudging in virtual CLIL environments

    Directory of Open Access Journals (Sweden)

    Helle Lykke Nielsen

    2014-03-01

    Full Text Available Foreign language teachers can benefit substantially from the notions of curation and nudging when scaffolding CLIL activities on the internet. This article shows how these principles can be integrated into CLILstore, a free multimedia-rich learning tool with seamless access to online dictionaries, and presents feedback from first and second year university students of Arabic as a second language to inform foreign language teachers about students’ needs and preferences in virtual learning environments.

  6. The Distinction Between Curative and Assistive Technology.

    Science.gov (United States)

    Stramondo, Joseph A

    2018-05-01

    Disability activists have sometimes claimed their disability has actually increased their well-being. Some even say they would reject a cure to keep these gains. Yet, these same activists often simultaneously propose improvements to the quality and accessibility of assistive technology. However, for any argument favoring assistive over curative technology (or vice versa) to work, there must be a coherent distinction between the two. This line is already vague and will become even less clear with the emergence of novel technologies. This paper asks and tries to answer the question: what is it about the paradigmatic examples of curative and assistive technologies that make them paradigmatic and how can these defining features help us clarify the hard cases? This analysis will begin with an argument that, while the common views of this distinction adequately explain the paradigmatic cases, they fail to accurately pick out the relevant features of those technologies that make them paradigmatic and to provide adequate guidance for parsing the hard cases. Instead, it will be claimed that these categories of curative or assistive technologies are defined by the role the technologies play in establishing a person's relational narrative identity as a member of one of two social groups: disabled people or non-disabled people.

  7. [Curative effect of ozone hydrotherapy for pemphigus].

    Science.gov (United States)

    Jiang, Fuqiong; Deng, Danqi; Li, Xiaolan; Wang, Wenfang; Xie, Hong; Wu, Yongzhuo; Luan, Chunyan; Yang, Binbin

    2018-02-28

    To determine clinical curative effects of ozone therapy for pemphigus vulgaris.
 Methods: Ozone hydrotherapy was used as an aid treatment for 32 patients with pemphigus vulgaris. The hydropathic compression of potassium permanganate solution for 34 patients with pemphigus vulgaris served as a control. The main treatment for both groups were glucocorticoids and immune inhibitors. The lesions of patients, bacterial infection, usage of antibiotics, patient's satisfaction, and clinical curative effect were evaluated in the 2 groups.
 Results: There was no significant difference in the curative effect and the average length of staying at hospital between the 2 groups (P>0.05). But rate for the usage of antibiotics was significantly reduced in the group of ozone hydrotherapy (P=0.039). The patients were more satisfied in using ozone hydrotherapy than the potassium permanganate solution after 7-day therapy (P>0.05).
 Conclusion: Ozone hydrotherapy is a safe and effective aid method for pemphigus vulgaris. It can reduce the usage of antibiotics.

  8. Argo: an integrative, interactive, text mining-based workbench supporting curation

    Science.gov (United States)

    Rak, Rafal; Rowley, Andrew; Black, William; Ananiadou, Sophia

    2012-01-01

    Curation of biomedical literature is often supported by the automatic analysis of textual content that generally involves a sequence of individual processing components. Text mining (TM) has been used to enhance the process of manual biocuration, but has been focused on specific databases and tasks rather than an environment integrating TM tools into the curation pipeline, catering for a variety of tasks, types of information and applications. Processing components usually come from different sources and often lack interoperability. The well established Unstructured Information Management Architecture is a framework that addresses interoperability by defining common data structures and interfaces. However, most of the efforts are targeted towards software developers and are not suitable for curators, or are otherwise inconvenient to use on a higher level of abstraction. To overcome these issues we introduce Argo, an interoperable, integrative, interactive and collaborative system for text analysis with a convenient graphic user interface to ease the development of processing workflows and boost productivity in labour-intensive manual curation. Robust, scalable text analytics follow a modular approach, adopting component modules for distinct levels of text analysis. The user interface is available entirely through a web browser that saves the user from going through often complicated and platform-dependent installation procedures. Argo comes with a predefined set of processing components commonly used in text analysis, while giving the users the ability to deposit their own components. The system accommodates various areas and levels of user expertise, from TM and computational linguistics to ontology-based curation. One of the key functionalities of Argo is its ability to seamlessly incorporate user-interactive components, such as manual annotation editors, into otherwise completely automatic pipelines. As a use case, we demonstrate the functionality of an in

  9. Research Problems in Data Curation: Outcomes from the Data Curation Education in Research Centers Program

    Science.gov (United States)

    Palmer, C. L.; Mayernik, M. S.; Weber, N.; Baker, K. S.; Kelly, K.; Marlino, M. R.; Thompson, C. A.

    2013-12-01

    The need for data curation is being recognized in numerous institutional settings as national research funding agencies extend data archiving mandates to cover more types of research grants. Data curation, however, is not only a practical challenge. It presents many conceptual and theoretical challenges that must be investigated to design appropriate technical systems, social practices and institutions, policies, and services. This presentation reports on outcomes from an investigation of research problems in data curation conducted as part of the Data Curation Education in Research Centers (DCERC) program. DCERC is developing a new model for educating data professionals to contribute to scientific research. The program is organized around foundational courses and field experiences in research and data centers for both master's and doctoral students. The initiative is led by the Graduate School of Library and Information Science at the University of Illinois at Urbana-Champaign, in collaboration with the School of Information Sciences at the University of Tennessee, and library and data professionals at the National Center for Atmospheric Research (NCAR). At the doctoral level DCERC is educating future faculty and researchers in data curation and establishing a research agenda to advance the field. The doctoral seminar, Research Problems in Data Curation, was developed and taught in 2012 by the DCERC principal investigator and two doctoral fellows at the University of Illinois. It was designed to define the problem space of data curation, examine relevant concepts and theories related to both technical and social perspectives, and articulate research questions that are either unexplored or under theorized in the current literature. There was a particular emphasis on the Earth and environmental sciences, with guest speakers brought in from NCAR, National Snow and Ice Data Center (NSIDC), and Rensselaer Polytechnic Institute. Through the assignments, students

  10. Correcting Inconsistencies and Errors in Bacterial Genome Metadata Using an Automated Curation Tool in Excel (AutoCurE).

    Science.gov (United States)

    Schmedes, Sarah E; King, Jonathan L; Budowle, Bruce

    2015-01-01

    Whole-genome data are invaluable for large-scale comparative genomic studies. Current sequencing technologies have made it feasible to sequence entire bacterial genomes with relative ease and time with a substantially reduced cost per nucleotide, hence cost per genome. More than 3,000 bacterial genomes have been sequenced and are available at the finished status. Publically available genomes can be readily downloaded; however, there are challenges to verify the specific supporting data contained within the download and to identify errors and inconsistencies that may be present within the organizational data content and metadata. AutoCurE, an automated tool for bacterial genome database curation in Excel, was developed to facilitate local database curation of supporting data that accompany downloaded genomes from the National Center for Biotechnology Information. AutoCurE provides an automated approach to curate local genomic databases by flagging inconsistencies or errors by comparing the downloaded supporting data to the genome reports to verify genome name, RefSeq accession numbers, the presence of archaea, BioProject/UIDs, and sequence file descriptions. Flags are generated for nine metadata fields if there are inconsistencies between the downloaded genomes and genomes reports and if erroneous or missing data are evident. AutoCurE is an easy-to-use tool for local database curation for large-scale genome data prior to downstream analyses.

  11. Competencies for preservation and digital curation

    Directory of Open Access Journals (Sweden)

    Sonia Boeres

    2016-09-01

    Full Text Available Information Science, throughout its existence, has been a multi and interdisciplinary field, and has undergone constant change because of its object of study: information. Seen that this element is not static and is increasingly linked to information technology, we have witnessed a challenge arise: how to ensure the permanence of digital libraries? How to secure the terabytes generated with increasing speed, and in various formats, will be available and fully capable of use over time? This is a challenge that Information Science professionals are being challenged to solve in the process of so-called digital preservation and curation. Thus, this article aims to raise the skills that the information professional must have to carry out the process of preservation and digital curation. The article discusses the emergence of professions (from the perspective of Sociology, the need to work for the realization of the human being (Psychology and proficiencies of exercising the office of Information Science to ensure the preservation of digital information in information units.

  12. Linking the Congenital Heart Surgery Databases of the Society of Thoracic Surgeons and the Congenital Heart Surgeons’ Society: Part 2—Lessons Learned and Implications

    Science.gov (United States)

    Jacobs, Jeffrey P.; Pasquali, Sara K.; Austin, Erle; Gaynor, J. William; Backer, Carl; Hirsch-Romano, Jennifer C.; Williams, William G.; Caldarone, Christopher A.; McCrindle, Brian W.; Graham, Karen E.; Dokholyan, Rachel S.; Shook, Gregory J.; Poteat, Jennifer; Baxi, Maulik V.; Karamlou, Tara; Blackstone, Eugene H.; Mavroudis, Constantine; Mayer, John E.; Jonas, Richard A.; Jacobs, Marshall L.

    2014-01-01

    Purpose A link has been created between the Society of Thoracic Surgeons Congenital Heart Surgery Database (STS-CHSD) and the Congenital Heart Surgeons’ Society Database (CHSS-D). Five matrices have been created that facilitate the automated identification of patients who are potentially eligible for the five active CHSS studies using the STS-CHSD. These matrices are now used to (1) estimate the denominator of patients eligible for CHSS studies and (2) compare “eligible and enrolled patients” to “potentially eligible and not enrolled patients” to assess the generalizability of CHSS studies. Methods The matrices were applied to 40 consenting institutions that participate in both the STS-CHSD and the CHSS to (1) estimate the denominator of patients that are potentially eligible for CHSS studies, (2) estimate the completeness of enrollment of patients eligible for CHSS studies among all CHSS sites, (3) estimate the completeness of enrollment of patients eligible for CHSS studies among those CHSS institutions participating in each CHSS cohort study, and (4) compare “eligible and enrolled patients” to “potentially eligible and not enrolled patients” to assess the generalizability of CHSS studies. The matrices were applied to all participants in the STS-CHSD to identify patients who underwent frequently performed operations and compare “eligible and enrolled patients” to “potentially eligible and not enrolled patients” in following five domains: (1) age at surgery, (2) gender, (3) race, (4) discharge mortality, and (5) postoperative length of stay. Completeness of enrollment was defined as the number of actually enrolled patients divided by the number of patients identified as being potentially eligible for enrollment. Results For the CHSS Critical Left Ventricular Outflow Tract Study (LVOTO) study, for the Norwood procedure, completeness of enrollment at centers actively participating in the LVOTO study was 34%. For the Norwood operation

  13. Curating research data a handbook of current practice

    CERN Document Server

    Johnston, Lisa R

    2017-01-01

    Curating Research Data, Volume Two: A Handbook of Current Practice guides you across the data lifecycle through the practical strategies and techniques for curating research data in a digital repository setting. The data curation steps for receiving, appraising, selecting, ingesting, transforming, describing, contextualizing, disseminating, and preserving digital research data are each explored, and then supplemented with detailed case studies written by more than forty international practitioners from national, disciplinary, and institutional data repositories. The steps in this volume detail the sequential actions that you might take to curate a data set from receiving the data (Step 1) to eventual reuse (Step 8). Data curators, archivists, research data management specialists, subject librarians, institutional repository managers, and digital library staff will benefit from these current and practical approaches to data curation.

  14. Curating research data practical strategies for your digital repository

    CERN Document Server

    Johnston, Lisa R

    2017-01-01

    Volume One of Curating Research Data explores the variety of reasons, motivations, and drivers for why data curation services are needed in the context of academic and disciplinary data repository efforts. Twelve chapters, divided into three parts, take an in-depth look at the complex practice of data curation as it emerges around us. Part I sets the stage for data curation by describing current policies, data sharing cultures, and collaborative efforts currently underway that impact potential services. Part II brings several key issues, such as cost recovery and marketing strategy, into focus for practitioners when considering how to put data curation services in action. Finally, Part III describes the full lifecycle of data by examining the ethical and practical reuse issues that data curation practitioners must consider as we strive to prepare data for the future.

  15. MS_HistoneDB, a manually curated resource for proteomic analysis of human and mouse histones.

    Science.gov (United States)

    El Kennani, Sara; Adrait, Annie; Shaytan, Alexey K; Khochbin, Saadi; Bruley, Christophe; Panchenko, Anna R; Landsman, David; Pflieger, Delphine; Govin, Jérôme

    2017-01-01

    Histones and histone variants are essential components of the nuclear chromatin. While mass spectrometry has opened a large window to their characterization and functional studies, their identification from proteomic data remains challenging. Indeed, the current interpretation of mass spectrometry data relies on public databases which are either not exhaustive (Swiss-Prot) or contain many redundant entries (UniProtKB or NCBI). Currently, no protein database is ideally suited for the analysis of histones and the complex array of mammalian histone variants. We propose two proteomics-oriented manually curated databases for mouse and human histone variants. We manually curated >1700 gene, transcript and protein entries to produce a non-redundant list of 83 mouse and 85 human histones. These entries were annotated in accordance with the current nomenclature and unified with the "HistoneDB2.0 with Variants" database. This resource is provided in a format that can be directly read by programs used for mass spectrometry data interpretation. In addition, it was used to interpret mass spectrometry data acquired on histones extracted from mouse testis. Several histone variants, which had so far only been inferred by homology or detected at the RNA level, were detected by mass spectrometry, confirming the existence of their protein form. Mouse and human histone entries were collected from different databases and subsequently curated to produce a non-redundant protein-centric resource, MS_HistoneDB. It is dedicated to the proteomic study of histones in mouse and human and will hopefully facilitate the identification and functional study of histone variants.

  16. An Emergent Micro-Services Approach to Digital Curation Infrastructure

    OpenAIRE

    Abrams, Stephen; Kunze, John; Loy, David

    2010-01-01

    In order better to meet the needs of its diverse University of California (UC) constituencies, the California Digital Library UC Curation Center is re-envisioning its approach to digital curation infrastructure by devolving function into a set of granular, independent, but interoperable micro-services. Since each of these services is small and self-contained, they are more easily developed, deployed, maintained, and enhanced; at the same time, complex curation function can emerge from the str...

  17. Self-Rerouting and Curative Interconnect Technology (SERCUIT)

    Science.gov (United States)

    2017-12-01

    SPECIAL REPORT RDMR-CS-17-01 SELF-REROUTING AND CURATIVE INTERCONNECT TECHNOLOGY (SERCUIT) Shiv Joshi Concepts to Systems, Inc...Final 4. TITLE AND SUBTITLE Self-Rerouting and Curative Interconnect Technology (SERCUIT) 5. FUNDING NUMBERS 6. AUTHOR(S) Shiv Joshi...concepts2systems.com (p) 434-207-5189 x (f) Click to view full size Title Contract Number SELF-REROUTING AND CURATIVE INTERCONNECT TECHNOLOGY (SERCUIT) W911W6-17-C-0029

  18. Somatic cancer variant curation and harmonization through consensus minimum variant level data

    Directory of Open Access Journals (Sweden)

    Deborah I. Ritter

    2016-11-01

    Full Text Available Abstract Background To truly achieve personalized medicine in oncology, it is critical to catalog and curate cancer sequence variants for their clinical relevance. The Somatic Working Group (WG of the Clinical Genome Resource (ClinGen, in cooperation with ClinVar and multiple cancer variant curation stakeholders, has developed a consensus set of minimal variant level data (MVLD. MVLD is a framework of standardized data elements to curate cancer variants for clinical utility. With implementation of MVLD standards, and in a working partnership with ClinVar, we aim to streamline the somatic variant curation efforts in the community and reduce redundancy and time burden for the interpretation of cancer variants in clinical practice. Methods We developed MVLD through a consensus approach by i reviewing clinical actionability interpretations from institutions participating in the WG, ii conducting extensive literature search of clinical somatic interpretation schemas, and iii survey of cancer variant web portals. A forthcoming guideline on cancer variant interpretation, from the Association of Molecular Pathology (AMP, can be incorporated into MVLD. Results Along with harmonizing standardized terminology for allele interpretive and descriptive fields that are collected by many databases, the MVLD includes unique fields for cancer variants such as Biomarker Class, Therapeutic Context and Effect. In addition, MVLD includes recommendations for controlled semantics and ontologies. The Somatic WG is collaborating with ClinVar to evaluate MVLD use for somatic variant submissions. ClinVar is an open and centralized repository where sequencing laboratories can report summary-level variant data with clinical significance, and ClinVar accepts cancer variant data. Conclusions We expect the use of the MVLD to streamline clinical interpretation of cancer variants, enhance interoperability among multiple redundant curation efforts, and increase submission of

  19. Sample Curation at a Lunar Outpost

    Science.gov (United States)

    Allen, Carlton C.; Lofgren, Gary E.; Treiman, A. H.; Lindstrom, Marilyn L.

    2007-01-01

    The six Apollo surface missions returned 2,196 individual rock and soil samples, with a total mass of 381.6 kg. Samples were collected based on visual examination by the astronauts and consultation with geologists in the science back room in Houston. The samples were photographed during collection, packaged in uniquely-identified containers, and transported to the Lunar Module. All samples collected on the Moon were returned to Earth. NASA's upcoming return to the Moon will be different. Astronauts will have extended stays at an out-post and will collect more samples than they will return. They will need curation and analysis facilities on the Moon in order to carefully select samples for return to Earth.

  20. SABIO-RK: an updated resource for manually curated biochemical reaction kinetics

    Science.gov (United States)

    Rey, Maja; Weidemann, Andreas; Kania, Renate; Müller, Wolfgang

    2018-01-01

    Abstract SABIO-RK (http://sabiork.h-its.org/) is a manually curated database containing data about biochemical reactions and their reaction kinetics. The data are primarily extracted from scientific literature and stored in a relational database. The content comprises both naturally occurring and alternatively measured biochemical reactions and is not restricted to any organism class. The data are made available to the public by a web-based search interface and by web services for programmatic access. In this update we describe major improvements and extensions of SABIO-RK since our last publication in the database issue of Nucleic Acid Research (2012). (i) The website has been completely revised and (ii) allows now also free text search for kinetics data. (iii) Additional interlinkages with other databases in our field have been established; this enables users to gain directly comprehensive knowledge about the properties of enzymes and kinetics beyond SABIO-RK. (iv) Vice versa, direct access to SABIO-RK data has been implemented in several systems biology tools and workflows. (v) On request of our experimental users, the data can be exported now additionally in spreadsheet formats. (vi) The newly established SABIO-RK Curation Service allows to respond to specific data requirements. PMID:29092055

  1. Crowdsourcing and curation: perspectives from biology and natural language processing.

    Science.gov (United States)

    Hirschman, Lynette; Fort, Karën; Boué, Stéphanie; Kyrpides, Nikos; Islamaj Doğan, Rezarta; Cohen, Kevin Bretonnel

    2016-01-01

    Crowdsourcing is increasingly utilized for performing tasks in both natural language processing and biocuration. Although there have been many applications of crowdsourcing in these fields, there have been fewer high-level discussions of the methodology and its applicability to biocuration. This paper explores crowdsourcing for biocuration through several case studies that highlight different ways of leveraging 'the crowd'; these raise issues about the kind(s) of expertise needed, the motivations of participants, and questions related to feasibility, cost and quality. The paper is an outgrowth of a panel session held at BioCreative V (Seville, September 9-11, 2015). The session consisted of four short talks, followed by a discussion. In their talks, the panelists explored the role of expertise and the potential to improve crowd performance by training; the challenge of decomposing tasks to make them amenable to crowdsourcing; and the capture of biological data and metadata through community editing.Database URL: http://www.mitre.org/publications/technical-papers/crowdsourcing-and-curation-perspectives. © The Author(s) 2016. Published by Oxford University Press.

  2. Data Curation in the World Data System: Proposed Framework

    Directory of Open Access Journals (Sweden)

    P Laughton

    2013-09-01

    Full Text Available The value of data in society is increasing rapidly. Organisations that work with data should have standard practices in place to ensure successful curation of data. The World Data System (WDS consists of a number of data centres responsible for curating research data sets for the scientific community. The WDS has no formal data curation framework or model in place to act as a guideline for member data centres. The objective of this research was to develop a framework for the curation of data in the WDS. A multiple-case case study was conducted. Interviews were used to gather qualitative data and analysis of the data, which led to the development of this framework. The proposed framework is largely based on the Open Archival Information System (OAIS functional model and caters for the curation of both analogue and digital data.

  3. Digital Management and Curation of the National Rock and Ore Collections at NMNH, Smithsonian

    Science.gov (United States)

    Cottrell, E.; Andrews, B.; Sorensen, S. S.; Hale, L. J.

    2011-12-01

    The National Museum of Natural History, Smithsonian Institution, is home to the world's largest curated rock collection. The collection houses 160,680 physical rock and ore specimen lots ("samples"), all of which already have a digital record that can be accessed by the public through a searchable web interface (http://collections.mnh.si.edu/search/ms/). In addition, there are 66 accessions pending that when catalogued will add approximately 60,000 specimen lots. NMNH's collections are digitally managed on the KE EMu° platform which has emerged as the premier system for managing collections in natural history museums worldwide. In 2010 the Smithsonian released an ambitious 5 year Digitization Strategic Plan. In Mineral Sciences, new digitization efforts in the next five years will focus on integrating various digital resources for volcanic specimens. EMu sample records will link to the corresponding records for physical eruption information housed within the database of Smithsonian's Global Volcanism Program (GVP). Linkages are also planned between our digital records and geochemical databases (like EarthChem or PetDB) maintained by third parties. We anticipate that these linkages will increase the use of NMNH collections as well as engender new scholarly directions for research. Another large project the museum is currently undertaking involves the integration of the functionality of in-house designed Transaction Management software with the EMu database. This will allow access to the details (borrower, quantity, date, and purpose) of all loans of a given specimen through its catalogue record. We hope this will enable cross-referencing and fertilization of research ideas while avoiding duplicate efforts. While these digitization efforts are critical, we propose that the greatest challenge to sample curation is not posed by digitization and that a global sample registry alone will not ensure that samples are available for reuse. We suggest instead that the ability

  4. Saccharomyces genome database informs human biology

    OpenAIRE

    Skrzypek, Marek S; Nash, Robert S; Wong, Edith D; MacPherson, Kevin A; Hellerstedt, Sage T; Engel, Stacia R; Karra, Kalpana; Weng, Shuai; Sheppard, Travis K; Binkley, Gail; Simison, Matt; Miyasato, Stuart R; Cherry, J Michael

    2017-01-01

    Abstract The Saccharomyces Genome Database (SGD; http://www.yeastgenome.org) is an expertly curated database of literature-derived functional information for the model organism budding yeast, Saccharomyces cerevisiae. SGD constantly strives to synergize new types of experimental data and bioinformatics predictions with existing data, and to organize them into a comprehensive and up-to-date information resource. The primary mission of SGD is to facilitate research into the biology of yeast and...

  5. Investigating Astromaterials Curation Applications for Dexterous Robotic Arms

    Science.gov (United States)

    Snead, C. J.; Jang, J. H.; Cowden, T. R.; McCubbin, F. M.

    2018-01-01

    The Astromaterials Acquisition and Curation office at NASA Johnson Space Center is currently investigating tools and methods that will enable the curation of future astromaterials collections. Size and temperature constraints for astromaterials to be collected by current and future proposed missions will require the development of new robotic sample and tool handling capabilities. NASA Curation has investigated the application of robot arms in the past, and robotic 3-axis micromanipulators are currently in use for small particle curation in the Stardust and Cosmic Dust laboratories. While 3-axis micromanipulators have been extremely successful for activities involving the transfer of isolated particles in the 5-20 micron range (e.g. from microscope slide to epoxy bullet tip, beryllium SEM disk), their limited ranges of motion and lack of yaw, pitch, and roll degrees of freedom restrict their utility in other applications. For instance, curators removing particles from cosmic dust collectors by hand often employ scooping and rotating motions to successfully free trapped particles from the silicone oil coatings. Similar scooping and rotating motions are also employed when isolating a specific particle of interest from an aliquot of crushed meteorite. While cosmic dust curators have been remarkably successful with these kinds of particle manipulations using handheld tools, operator fatigue limits the number of particles that can be removed during a given extraction session. The challenges for curation of small particles will be exacerbated by mission requirements that samples be processed in N2 sample cabinets (i.e. gloveboxes). We have been investigating the use of compact robot arms to facilitate sample handling within gloveboxes. Six-axis robot arms potentially have applications beyond small particle manipulation. For instance, future sample return missions may involve biologically sensitive astromaterials that can be easily compromised by physical interaction with

  6. Curated compendium of human transcriptional biomarker data.

    Science.gov (United States)

    Golightly, Nathan P; Bell, Avery; Bischoff, Anna I; Hollingsworth, Parker D; Piccolo, Stephen R

    2018-04-17

    One important use of genome-wide transcriptional profiles is to identify relationships between transcription levels and patient outcomes. These translational insights can guide the development of biomarkers for clinical application. Data from thousands of translational-biomarker studies have been deposited in public repositories, enabling reuse. However, data-reuse efforts require considerable time and expertise because transcriptional data are generated using heterogeneous profiling technologies, preprocessed using diverse normalization procedures, and annotated in non-standard ways. To address this problem, we curated 45 publicly available, translational-biomarker datasets from a variety of human diseases. To increase the data's utility, we reprocessed the raw expression data using a uniform computational pipeline, addressed quality-control problems, mapped the clinical annotations to a controlled vocabulary, and prepared consistently structured, analysis-ready data files. These data, along with scripts we used to prepare the data, are available in a public repository. We believe these data will be particularly useful to researchers seeking to perform benchmarking studies-for example, to compare and optimize machine-learning algorithms' ability to predict biomedical outcomes.

  7. Trafkintu: seed curators defending food sovereignty

    Directory of Open Access Journals (Sweden)

    Nastassja Nicole Mancilla Ivaca

    2014-10-01

    Full Text Available This paper examines the resurgence of Trafkintu, an ancient Mapuche ritual of seed trade; now as a folk-communication practice of resistance, against neoliberal transformations in farming that threaten food sovereignty of rural communities in southern Chile. Drawing onparticipant observation and semi-structured interviews with peasant and Mapuche women involved in these practices, we show that seed curators women act as agents that revalue the localness [lo local] through a process of resignification of Trafkintu, this time linking it tofood self-sufficiency. In addition, they build networks between indigenous and peasant communities as a resistance strategy. However, this resurgence of Trafkintu becomes ambivalent as its new symbolic expression is being appropriated by local mainstreampoliticians, for electoral purposes, to promote an image of 'concern about popular culture'. That is, a tool of resistance, on the one hand, and a kind of political folk-marketing, on the other.

  8. 3DSwap: Curated knowledgebase of proteins involved in 3D domain swapping

    KAUST Repository

    Shameer, Khader

    2011-09-29

    Three-dimensional domain swapping is a unique protein structural phenomenon where two or more protein chains in a protein oligomer share a common structural segment between individual chains. This phenomenon is observed in an array of protein structures in oligomeric conformation. Protein structures in swapped conformations perform diverse functional roles and are also associated with deposition diseases in humans. We have performed in-depth literature curation and structural bioinformatics analyses to develop an integrated knowledgebase of proteins involved in 3D domain swapping. The hallmark of 3D domain swapping is the presence of distinct structural segments such as the hinge and swapped regions. We have curated the literature to delineate the boundaries of these regions. In addition, we have defined several new concepts like \\'secondary major interface\\' to represent the interface properties arising as a result of 3D domain swapping, and a new quantitative measure for the \\'extent of swapping\\' in structures. The catalog of proteins reported in 3DSwap knowledgebase has been generated using an integrated structural bioinformatics workflow of database searches, literature curation, by structure visualization and sequence-structure-function analyses. The current version of the 3DSwap knowledgebase reports 293 protein structures, the analysis of such a compendium of protein structures will further the understanding molecular factors driving 3D domain swapping. The Author(s) 2011.

  9. Curative effects of small incision cataract surgery versus phacoemulsification: a Meta-analysis

    Directory of Open Access Journals (Sweden)

    Chang-Jian Yang

    2013-08-01

    Full Text Available AIM: To evaluate the curative efficacy of small incision cataract surgery(SICSversus phacoemulsification(Phaco.METHODS: A computerized literature search was carried out in Chinese Biomedical Database(CBM, Wanfang Data, VIP and Chinese National Knowledge Infrastructure(CNKIto collect articles published between 1989-2013 concerning the curative efficacy of SICS versus Phaco. The studies were assessed in terms of clinical case-control criteria. Meta-analysis were performed to assess the visual acuity, the complications rates between SICS and Phaco 90 days after surgery. Treatment effects were measured as risk difference(RDbetween SICS and Phaco. Fixed and random effect models were employed to combine results after a heterogeneity test. RESULTS:A total of 8 studies were included in our Meta-analysis. At 90 days postoperative time, there were no significant differences between the two groups at the visual acuity >0.5(P=0.14; and no significant differences on the complications rates of corneal astigmatism, corneal edema, posterior capsular rupture and anterior iris reaction(P>0.05.CONCLUSION: These results suggest that there is no different on the curative effects of SICS and Phaco for cataract.

  10. Curation and Computational Design of Bioenergy-Related Metabolic Pathways

    Energy Technology Data Exchange (ETDEWEB)

    Karp, Peter D. [SRI International, Menlo Park, CA (United States)

    2014-09-12

    Pathway Tools is a systems-biology software package written by SRI International (SRI) that produces Pathway/Genome Databases (PGDBs) for organisms with a sequenced genome. Pathway Tools also provides a wide range of capabilities for analyzing predicted metabolic networks and user-generated omics data. More than 5,000 academic, industrial, and government groups have licensed Pathway Tools. This user community includes researchers at all three DOE bioenergy centers, as well as academic and industrial metabolic engineering (ME) groups. An integral part of the Pathway Tools software is MetaCyc, a large, multiorganism database of metabolic pathways and enzymes that SRI and its academic collaborators manually curate. This project included two main goals: I. Enhance the MetaCyc content of bioenergy-related enzymes and pathways. II. Develop computational tools for engineering metabolic pathways that satisfy specified design goals, in particular for bioenergy-related pathways. In part I, SRI proposed to significantly expand the coverage of bioenergy-related metabolic information in MetaCyc, followed by the generation of organism-specific PGDBs for all energy-relevant organisms sequenced at the DOE Joint Genome Institute (JGI). Part I objectives included: 1: Expand the content of MetaCyc to include bioenergy-related enzymes and pathways. 2: Enhance the Pathway Tools software to enable display of complex polymer degradation processes. 3: Create new PGDBs for the energy-related organisms sequenced by JGI, update existing PGDBs with new MetaCyc content, and make these data available to JBEI via the BioCyc website. In part II, SRI proposed to develop an efficient computational tool for the engineering of metabolic pathways. Part II objectives included: 4: Develop computational tools for generating metabolic pathways that satisfy specified design goals, enabling users to specify parameters such as starting and ending compounds, and preferred or disallowed intermediate compounds

  11. Bisphosphonate adverse effects, lessons from large databases

    DEFF Research Database (Denmark)

    Abrahamsen, Bo

    2010-01-01

    is on bisphosphonates used for osteoporosis. RECENT FINDINGS: Register studies have so far not confirmed a shift from classical to nonclassical femur fractures with bisphosphonates. However, studies were either small or without X-ray adjudication. Two new studies found no increase in jaw surgery for inflammatory...

  12. WeCurate: Designing for synchronised browsing and social negotiation

    OpenAIRE

    Hazelden, Katina; Yee-King, Matthew; d'Inverno, Mark; Confalonieri, Roberto; De Jonge, Dave; Amgoud, Leila; Osman, Nardine; Prade, Henri; Sierra, Carles

    2012-01-01

    WeCurate is a shared image browser for collaboratively curating a virtual exhibition from a cultural image archive. This paper is concerned with the evaluation and iteration of a prototype UI (User Interface) design to enable this community image browsing. In WeCurate, several remote users work together with autonomic agents to browse the archive and to select, through negotiation and voting, a set of images which are of the greatest interest to the group. The UI allows users to synchronize v...

  13. Automatic vs. manual curation of a multi-source chemical dictionary: the impact on text mining

    Science.gov (United States)

    2010-01-01

    Background Previously, we developed a combined dictionary dubbed Chemlist for the identification of small molecules and drugs in text based on a number of publicly available databases and tested it on an annotated corpus. To achieve an acceptable recall and precision we used a number of automatic and semi-automatic processing steps together with disambiguation rules. However, it remained to be investigated which impact an extensive manual curation of a multi-source chemical dictionary would have on chemical term identification in text. ChemSpider is a chemical database that has undergone extensive manual curation aimed at establishing valid chemical name-to-structure relationships. Results We acquired the component of ChemSpider containing only manually curated names and synonyms. Rule-based term filtering, semi-automatic manual curation, and disambiguation rules were applied. We tested the dictionary from ChemSpider on an annotated corpus and compared the results with those for the Chemlist dictionary. The ChemSpider dictionary of ca. 80 k names was only a 1/3 to a 1/4 the size of Chemlist at around 300 k. The ChemSpider dictionary had a precision of 0.43 and a recall of 0.19 before the application of filtering and disambiguation and a precision of 0.87 and a recall of 0.19 after filtering and disambiguation. The Chemlist dictionary had a precision of 0.20 and a recall of 0.47 before the application of filtering and disambiguation and a precision of 0.67 and a recall of 0.40 after filtering and disambiguation. Conclusions We conclude the following: (1) The ChemSpider dictionary achieved the best precision but the Chemlist dictionary had a higher recall and the best F-score; (2) Rule-based filtering and disambiguation is necessary to achieve a high precision for both the automatically generated and the manually curated dictionary. ChemSpider is available as a web service at http://www.chemspider.com/ and the Chemlist dictionary is freely available as an XML file in

  14. Curative radiotherapy for primary orbital lymphoma

    International Nuclear Information System (INIS)

    Bhatia, Sudershan; Paulino, Arnold C.; Buatti, John M.; Mayr, Nina A.; Wen, B.-C.

    2002-01-01

    Purpose: To review our institutional experience with primary orbital lymphoma and determine the prognostic factors for survival, local control, and distant metastases. In addition, we also analyzed the risk factors for complications in the radiotherapeutic management of this tumor. Methods and Materials: Between 1973 and 1998, 47 patients (29 women [62%] and 18 men [38%], median age 69 years, range 32-89) with Stage IAE orbital lymphoma were treated with curative intent at one department. Five had bilateral orbital involvement. The tumor was located in the eyelid and extraocular muscles in 23 (44%), conjunctiva in 17 (33%), and lacrimal apparatus in 12 (23%). The histologic features according to the World Heath Organization classification of lymphoid neoplasms was follicular lymphoma in 25, extranodal marginal zone B-cell lymphoma of mucosa-associated lymphoid tissue type in 8, diffuse large B-cell lymphoma in 12, mantle cell lymphoma in 6, and peripheral T-cell lymphoma in 1. For the purposes of comparison with the existing literature on orbital lymphomas, the grading system according to the Working Formulation was also recorded. The histologic grade was low in 33 (63%), intermediate in 18 (35%), and high in 1 (2%). All patients were treated with primary radiotherapy alone. The median dose for low-grade tumors was 3000 cGy (range 2000-4020); the median dose for intermediate and high-grade tumors was 4000 cGy (range 3000-5100). A lens-sparing approach was used in 19 patients (37%). Late complications for the lens and cornea were scored according to the subjective, objective, management, and analytic (SOMA) scale of the Late Effects of Normal Tissue (LENT) scoring system. The median follow-up was 55 months (range 6-232). Results: The local control rate was 100% in the 52 orbits treated. The 5-year overall survival and relapse-free survival rate was 73.6% and 65.5%, respectively. Tumor grade and location did not predict for overall survival or relapse-free survival

  15. An Emergent Micro-Services Approach to Digital Curation Infrastructure

    Directory of Open Access Journals (Sweden)

    Stephen Abrams

    2010-07-01

    Full Text Available In order better to meet the needs of its diverse University of California (UC constituencies, the California Digital Library UC Curation Center is re-envisioning its approach to digital curation infrastructure by devolving function into a set of granular, independent, but interoperable micro-services. Since each of these services is small and self-contained, they are more easily developed, deployed, maintained, and enhanced; at the same time, complex curation function can emerge from the strategic combination of atomistic services. The emergent approach emphasizes the persistence of content rather than the systems in which that content is managemed, thus the paradigmatic archival culture is not unduly coupled to any particular technological context. This results in a curation environment that is comprehensive in scope, yet flexible with regard to local policies and practices and sustainable despite the inevitability of disruptive change in technology and user expectation.

  16. Sample Transport for a European Sample Curation Facility

    Science.gov (United States)

    Berthoud, L.; Vrublevskis, J. B.; Bennett, A.; Pottage, T.; Bridges, J. C.; Holt, J. M. C.; Dirri, F.; Longobardo, A.; Palomba, E.; Russell, S.; Smith, C.

    2018-04-01

    This work has looked at the recovery of Mars Sample Return capsule once it arrives on Earth. It covers possible landing sites, planetary protection requirements, and transportation from the landing site to a European Sample Curation Facility.

  17. Judson_Mansouri_Automated_Chemical_Curation_QSAREnvRes_Data

    Data.gov (United States)

    U.S. Environmental Protection Agency — Here we describe the development of an automated KNIME workflow to curate and correct errors in the structure and identity of chemicals using the publically...

  18. ECOTOX Knowledgebase: New tools for data visualization and database interoperability (poster)

    Science.gov (United States)

    The ECOTOXicology knowledgebase (ECOTOX) is a comprehensive, curated database that summarizes toxicology data from single chemical exposure studies to terrestrial and aquatic organisms. The ECOTOX Knowledgebase provides risk assessors and researchers consistent information on tox...

  19. Relational databases

    CERN Document Server

    Bell, D A

    1986-01-01

    Relational Databases explores the major advances in relational databases and provides a balanced analysis of the state of the art in relational databases. Topics covered include capture and analysis of data placement requirements; distributed relational database systems; data dependency manipulation in database schemata; and relational database support for computer graphics and computer aided design. This book is divided into three sections and begins with an overview of the theory and practice of distributed systems, using the example of INGRES from Relational Technology as illustration. The

  20. The BioC-BioGRID corpus: full text articles annotated for curation of protein–protein and genetic interactions

    Science.gov (United States)

    Kim, Sun; Chatr-aryamontri, Andrew; Chang, Christie S.; Oughtred, Rose; Rust, Jennifer; Wilbur, W. John; Comeau, Donald C.; Dolinski, Kara; Tyers, Mike

    2017-01-01

    A great deal of information on the molecular genetics and biochemistry of model organisms has been reported in the scientific literature. However, this data is typically described in free text form and is not readily amenable to computational analyses. To this end, the BioGRID database systematically curates the biomedical literature for genetic and protein interaction data. This data is provided in a standardized computationally tractable format and includes structured annotation of experimental evidence. BioGRID curation necessarily involves substantial human effort by expert curators who must read each publication to extract the relevant information. Computational text-mining methods offer the potential to augment and accelerate manual curation. To facilitate the development of practical text-mining strategies, a new challenge was organized in BioCreative V for the BioC task, the collaborative Biocurator Assistant Task. This was a non-competitive, cooperative task in which the participants worked together to build BioC-compatible modules into an integrated pipeline to assist BioGRID curators. As an integral part of this task, a test collection of full text articles was developed that contained both biological entity annotations (gene/protein and organism/species) and molecular interaction annotations (protein–protein and genetic interactions (PPIs and GIs)). This collection, which we call the BioC-BioGRID corpus, was annotated by four BioGRID curators over three rounds of annotation and contains 120 full text articles curated in a dataset representing two major model organisms, namely budding yeast and human. The BioC-BioGRID corpus contains annotations for 6409 mentions of genes and their Entrez Gene IDs, 186 mentions of organism names and their NCBI Taxonomy IDs, 1867 mentions of PPIs and 701 annotations of PPI experimental evidence statements, 856 mentions of GIs and 399 annotations of GI evidence statements. The purpose, characteristics and possible future

  1. The BioC-BioGRID corpus: full text articles annotated for curation of protein-protein and genetic interactions.

    Science.gov (United States)

    Islamaj Dogan, Rezarta; Kim, Sun; Chatr-Aryamontri, Andrew; Chang, Christie S; Oughtred, Rose; Rust, Jennifer; Wilbur, W John; Comeau, Donald C; Dolinski, Kara; Tyers, Mike

    2017-01-01

    A great deal of information on the molecular genetics and biochemistry of model organisms has been reported in the scientific literature. However, this data is typically described in free text form and is not readily amenable to computational analyses. To this end, the BioGRID database systematically curates the biomedical literature for genetic and protein interaction data. This data is provided in a standardized computationally tractable format and includes structured annotation of experimental evidence. BioGRID curation necessarily involves substantial human effort by expert curators who must read each publication to extract the relevant information. Computational text-mining methods offer the potential to augment and accelerate manual curation. To facilitate the development of practical text-mining strategies, a new challenge was organized in BioCreative V for the BioC task, the collaborative Biocurator Assistant Task. This was a non-competitive, cooperative task in which the participants worked together to build BioC-compatible modules into an integrated pipeline to assist BioGRID curators. As an integral part of this task, a test collection of full text articles was developed that contained both biological entity annotations (gene/protein and organism/species) and molecular interaction annotations (protein-protein and genetic interactions (PPIs and GIs)). This collection, which we call the BioC-BioGRID corpus, was annotated by four BioGRID curators over three rounds of annotation and contains 120 full text articles curated in a dataset representing two major model organisms, namely budding yeast and human. The BioC-BioGRID corpus contains annotations for 6409 mentions of genes and their Entrez Gene IDs, 186 mentions of organism names and their NCBI Taxonomy IDs, 1867 mentions of PPIs and 701 annotations of PPI experimental evidence statements, 856 mentions of GIs and 399 annotations of GI evidence statements. The purpose, characteristics and possible future

  2. Biofuel Database

    Science.gov (United States)

    Biofuel Database (Web, free access)   This database brings together structural, biological, and thermodynamic data for enzymes that are either in current use or are being considered for use in the production of biofuels.

  3. Community Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This excel spreadsheet is the result of merging at the port level of several of the in-house fisheries databases in combination with other demographic databases such...

  4. NeuroRDF: semantic integration of highly curated data to prioritize biomarker candidates in Alzheimer's disease.

    Science.gov (United States)

    Iyappan, Anandhi; Kawalia, Shweta Bagewadi; Raschka, Tamara; Hofmann-Apitius, Martin; Senger, Philipp

    2016-07-08

    Neurodegenerative diseases are incurable and debilitating indications with huge social and economic impact, where much is still to be learnt about the underlying molecular events. Mechanistic disease models could offer a knowledge framework to help decipher the complex interactions that occur at molecular and cellular levels. This motivates the need for the development of an approach integrating highly curated and heterogeneous data into a disease model of different regulatory data layers. Although several disease models exist, they often do not consider the quality of underlying data. Moreover, even with the current advancements in semantic web technology, we still do not have cure for complex diseases like Alzheimer's disease. One of the key reasons accountable for this could be the increasing gap between generated data and the derived knowledge. In this paper, we describe an approach, called as NeuroRDF, to develop an integrative framework for modeling curated knowledge in the area of complex neurodegenerative diseases. The core of this strategy lies in the usage of well curated and context specific data for integration into one single semantic web-based framework, RDF. This increases the probability of the derived knowledge to be novel and reliable in a specific disease context. This infrastructure integrates highly curated data from databases (Bind, IntAct, etc.), literature (PubMed), and gene expression resources (such as GEO and ArrayExpress). We illustrate the effectiveness of our approach by asking real-world biomedical questions that link these resources to prioritize the plausible biomarker candidates. Among the 13 prioritized candidate genes, we identified MIF to be a potential emerging candidate due to its role as a pro-inflammatory cytokine. We additionally report on the effort and challenges faced during generation of such an indication-specific knowledge base comprising of curated and quality-controlled data. Although many alternative approaches

  5. Database Administrator

    Science.gov (United States)

    Moore, Pam

    2010-01-01

    The Internet and electronic commerce (e-commerce) generate lots of data. Data must be stored, organized, and managed. Database administrators, or DBAs, work with database software to find ways to do this. They identify user needs, set up computer databases, and test systems. They ensure that systems perform as they should and add people to the…

  6. Curating NASA's Future Extraterrestrial Sample Collections: How Do We Achieve Maximum Proficiency?

    Science.gov (United States)

    McCubbin, Francis; Evans, Cynthia; Zeigler, Ryan; Allton, Judith; Fries, Marc; Righter, Kevin; Zolensky, Michael

    2016-01-01

    The Astromaterials Acquisition and Curation Office (henceforth referred to herein as NASA Curation Office) at NASA Johnson Space Center (JSC) is responsible for curating all of NASA's extraterrestrial samples. Under the governing document, NASA Policy Directive (NPD) 7100.10E "Curation of Extraterrestrial Materials", JSC is charged with "The curation of all extraterrestrial material under NASA control, including future NASA missions." The Directive goes on to define Curation as including "... documentation, preservation, preparation, and distribution of samples for research, education, and public outreach." Here we describe some of the ongoing efforts to ensure that the future activities of the NASA Curation Office are working towards a state of maximum proficiency.

  7. Safety and Mission Assurance for In-House Design Lessons Learned from Ares I Upper Stage

    Science.gov (United States)

    Anderson, Joel M.

    2011-01-01

    This viewgraph presentation identifies lessons learned in the course of the Ares I Upper Stage design and in-house development effort. The contents include: 1) Constellation Organization; 2) Upper Stage Organization; 3) Presentation Structure; 4) Lesson-Importance of Systems Engineering/Integration; 5) Lesson-Importance of Early S&MA Involvement; 6) Lesson-Importance of Appropriate Staffing Levels; 7) Lesson-Importance S&MA Team Deployment; 8) Lesson-Understanding of S&MA In-Line Engineering versus Assurance; 9) Lesson-Importance of Close Coordination between Supportability and Reliability/Maintainability; 10) Lesson-Importance of Engineering Data Systems; 11) Lesson-Importance of Early Development of Supporting Databases; 12) Lesson-Importance of Coordination with Safety Assessment/Review Panels; 13) Lesson-Implementation of Software Reliability; 14) Lesson-Implementation of S&MA Technical Authority/Chief S&MA Officer; 15) Lesson-Importance of S&MA Evaluation of Project Risks; 16) Lesson-Implementation of Critical Items List and Government Mandatory Inspections; 17) Lesson-Implementation of Critical Items List Mandatory Inspections; 18) Lesson-Implementation of Test Article Safety Analysis; and 19) Lesson-Importance of Procurement Quality.

  8. International survey of academic library data curation practices

    CERN Document Server

    2013-01-01

    This survey looks closely at the data curation practices of a sample of research-oriented universities largely from the USA, the UK, Australia and Scandinavia but also including India, South Africa and other countries. The study looks at how major universities are assisting faculty in developing data curation and management plans for large scale data projects, largely in the sciences and social sciences, often as pre-conditions for major grants. The report looks at which departments of universities are shouldering the data curation burden, the personnel involved in the efforts, the costs involved, types of software used, difficulties in procuring scientific experiment logs and other hard to obtain information, types of training offered to faculty, and other issues in large scale data management.

  9. Protective, curative and eradicative activities of fungicides against grapevine rust

    Directory of Open Access Journals (Sweden)

    Francislene Angelotti

    2014-01-01

    Full Text Available The protective, eradicative and curative activities of the fungicides azoxystrobin, tebuconazole, pyraclostrobin+metiram, and ciproconazole against grapevine rust, were determined in greenhouse. To evaluate the protective activity, leaves of potted ´Niagara´ (Vitis labrusca vines were artificially inoculated with an urediniospore suspension of Phakopsora euvitis four, eight or forteen days after fungicidal spray; and to evaluate the curative and eradicative activities, leaves were sprayed with fungicides two, four or eight days after inoculation. Disease severity was assessed 14 days after each inoculation. All tested fungicides present excellent preventive activity against grapevine rust; however, tebuconazole and ciproconazole provide better curative activity than azoxystrobin and pyraclostrobin+metiram. It was observed also that all tested fungicides significantly reduced the germination of urediniospore produced on sprayed leaves.

  10. Curating Public Art 2.0: The case of Autopoiesis

    DEFF Research Database (Denmark)

    Ajana, Btihaj

    2017-01-01

    This article examines the intersections between public art, curation and Web 2.0 technology. Building on the case study of Autopoiesis, a digital art project focusing on the curation and online exhibition of artworks received from members of the public in the United Arab Emirates, the article...... to facilitate autonomous creative self-expressions and enable greater public participation in culture. By providing a critical reflection on the ‘material’ contexts of this digital project, the article also demonstrates the related tensions between the virtual and the physical, and the wider ‘local’ realities...

  11. Organic Contamination Baseline Study on NASA JSC Astromaterial Curation Gloveboxes

    Science.gov (United States)

    Calaway, Michael J.; Allton, J. H.; Allen, C. C.; Burkett, P. J.

    2013-01-01

    Future planned sample return missions to carbon-rich asteroids and Mars in the next two decades will require strict handling and curation protocols as well as new procedures for reducing organic contamination. After the Apollo program, astromaterial collections have mainly been concerned with inorganic contamination [1-4]. However, future isolation containment systems for astromaterials, possibly nitrogen enriched gloveboxes, must be able to reduce organic and inorganic cross-contamination. In 2012, a baseline study was orchestrated to establish the current state of organic cleanliness in gloveboxes used by NASA JSC astromaterials curation labs that could be used as a benchmark for future mission designs.

  12. Observations on the Curative Effect of Acupuncture on Depressive Neurosis

    Institute of Scientific and Technical Information of China (English)

    FU Wen-bin; WANG Si-you

    2003-01-01

    Purpose To evaluate the curative effect of acupuncture on depressive neurosis. Method Sixty-two patients were randomly divided into a treatment group of 32 cases and a control group of 30 cases. The treatment group and the control group were treated with acupuncture and Fluoxetine, respectively. The curative effects were evaluated by HAMD. Results There was a significant difference between pretreatment and posttreatmentin each group ( P 0.05). But acupuncture had no side effects and was good in compliance. Conclusion Acupuncture is an effective method for treating depressive neurosis.

  13. The prognostic importance of jaundice in surgical resection with curative intent for gallbladder cancer.

    Science.gov (United States)

    Yang, Xin-wei; Yuan, Jian-mao; Chen, Jun-yi; Yang, Jue; Gao, Quan-gen; Yan, Xing-zhou; Zhang, Bao-hua; Feng, Shen; Wu, Meng-chao

    2014-09-03

    Preoperative jaundice is frequent in gallbladder cancer (GBC) and indicates advanced disease. Resection is rarely recommended to treat advanced GBC. An aggressive surgical approach for advanced GBC remains lacking because of the association of this disease with serious postoperative complications and poor prognosis. This study aims to re-assess the prognostic value of jaundice for the morbidity, mortality, and survival of GBC patients who underwent surgical resection with curative intent. GBC patients who underwent surgical resection with curative intent at a single institution between January 2003 and December 2012 were identified from a prospectively maintained database. A total of 192 patients underwent surgical resection with curative intent, of whom 47 had preoperative jaundice and 145 had none. Compared with the non-jaundiced patients, the jaundiced patients had significantly longer operative time (p jaundice was the only independent predictor of postoperative complications. The jaundiced patients had lower survival rates than the non-jaundiced patients (p jaundiced patients. The survival rates of the jaundiced patients with preoperative biliary drainage (PBD) were similar to those of the jaundiced patients without PBD (p = 0.968). No significant differences in the rate of postoperative intra-abdominal abscesses were found between the jaundiced patients with and without PBD (n = 4, 21.1% vs. n = 5, 17.9%, p = 0.787). Preoperative jaundice indicates poor prognosis and high postoperative morbidity but is not a surgical contraindication. Gallbladder neck tumors significantly increase the surgical difficulty and reduce the opportunities for radical resection. Gallbladder neck tumors can independently predict poor outcome. PBD correlates with neither a low rate of postoperative intra-abdominal abscesses nor a high survival rate.

  14. Data Curation Network: How Do We Compare? A Snapshot of Six Academic Library Institutions’ Data Repository and Curation Services

    Directory of Open Access Journals (Sweden)

    Lisa R. Johnston

    2017-02-01

    Full Text Available Objective: Many academic and research institutions are exploring opportunities to better support researchers in sharing their data. As partners in the Data Curation Network project, our six institutions developed a comparison of the current levels of support provided for researchers to meet their data sharing goals through library-based data repository and curation services. Methods: Each institutional lead provided a written summary of their services based on a previously developed structure, followed by group discussion and refinement of descriptions. Service areas assessed include the repository services for data, technologies used, policies, and staffing in place. Conclusions: Through this process we aim to better define the current levels of support offered by our institutions as a first step toward meeting our project's overarching goal to develop a shared staffing model for data curation across multiple institutions.

  15. The Danish Bladder Cancer Database

    Directory of Open Access Journals (Sweden)

    Hansen E

    2016-10-01

    Full Text Available Erik Hansen,1–3 Heidi Larsson,4 Mette Nørgaard,4 Peter Thind,3,5 Jørgen Bjerggaard Jensen1–3 1Department of Urology, Hospital of West Jutland-Holstebro, Holstebro, 2Department of Urology, Aarhus University Hospital, Aarhus, 3The Danish Bladder Cancer Database Group, 4Department of Clinical Epidemiology, Aarhus University Hospital, Aarhus, 5Department of Urology, Copenhagen University Hospital, Copenhagen, Denmark Aim of database: The aim of the Danish Bladder Cancer Database (DaBlaCa-data is to monitor the treatment of all patients diagnosed with invasive bladder cancer (BC in Denmark. Study population: All patients diagnosed with BC in Denmark from 2012 onward were included in the study. Results presented in this paper are predominantly from the 2013 population. Main variables: In 2013, 970 patients were diagnosed with BC in Denmark and were included in a preliminary report from the database. A total of 458 (47% patients were diagnosed with non-muscle-invasive BC (non-MIBC and 512 (53% were diagnosed with muscle-invasive BC (MIBC. A total of 300 (31% patients underwent cystectomy. Among the 135 patients diagnosed with MIBC, who were 75 years of age or younger, 67 (50% received neoadjuvent chemotherapy prior to cystectomy. In 2013, a total of 147 patients were treated with curative-intended radiation therapy. Descriptive data: One-year mortality was 28% (95% confidence interval [CI]: 15–21. One-year cancer-specific mortality was 25% (95% CI: 22–27%. One-year mortality after cystectomy was 14% (95% CI: 10–18. Ninety-day mortality after cystectomy was 3% (95% CI: 1–5 in 2013. One-year mortality following curative-intended radiation therapy was 32% (95% CI: 24–39 and 1-year cancer-specific mortality was 23% (95% CI: 16–31 in 2013. Conclusion: This preliminary DaBlaCa-data report showed that the treatment of MIBC in Denmark overall meet high international academic standards. The database is able to identify Danish BC patients and

  16. Earth System Model Development and Analysis using FRE-Curator and Live Access Servers: On-demand analysis of climate model output with data provenance.

    Science.gov (United States)

    Radhakrishnan, A.; Balaji, V.; Schweitzer, R.; Nikonov, S.; O'Brien, K.; Vahlenkamp, H.; Burger, E. F.

    2016-12-01

    There are distinct phases in the development cycle of an Earth system model. During the model development phase, scientists make changes to code and parameters and require rapid access to results for evaluation. During the production phase, scientists may make an ensemble of runs with different settings, and produce large quantities of output, that must be further analyzed and quality controlled for scientific papers and submission to international projects such as the Climate Model Intercomparison Project (CMIP). During this phase, provenance is a key concern:being able to track back from outputs to inputs. We will discuss one of the paths taken at GFDL in delivering tools across this lifecycle, offering on-demand analysis of data by integrating the use of GFDL's in-house FRE-Curator, Unidata's THREDDS and NOAA PMEL's Live Access Servers (LAS).Experience over this lifecycle suggests that a major difficulty in developing analysis capabilities is only partially the scientific content, but often devoted to answering the questions "where is the data?" and "how do I get to it?". "FRE-Curator" is the name of a database-centric paradigm used at NOAA GFDL to ingest information about the model runs into an RDBMS (Curator database). The components of FRE-Curator are integrated into Flexible Runtime Environment workflow and can be invoked during climate model simulation. The front end to FRE-Curator, known as the Model Development Database Interface (MDBI) provides an in-house web-based access to GFDL experiments: metadata, analysis output and more. In order to provide on-demand visualization, MDBI uses Live Access Servers which is a highly configurable web server designed to provide flexible access to geo-referenced scientific data, that makes use of OPeNDAP. Model output saved in GFDL's tape archive, the size of the database and experiments, continuous model development initiatives with more dynamic configurations add complexity and challenges in providing an on

  17. Bookshelf: a simple curation system for the storage of biomolecular simulation data.

    Science.gov (United States)

    Vohra, Shabana; Hall, Benjamin A; Holdbrook, Daniel A; Khalid, Syma; Biggin, Philip C

    2010-01-01

    Molecular dynamics simulations can now routinely generate data sets of several hundreds of gigabytes in size. The ability to generate this data has become easier over recent years and the rate of data production is likely to increase rapidly in the near future. One major problem associated with this vast amount of data is how to store it in a way that it can be easily retrieved at a later date. The obvious answer to this problem is a database. However, a key issue in the development and maintenance of such a database is its sustainability, which in turn depends on the ease of the deposition and retrieval process. Encouraging users to care about meta-data is difficult and thus the success of any storage system will ultimately depend on how well used by end-users the system is. In this respect we suggest that even a minimal amount of metadata if stored in a sensible fashion is useful, if only at the level of individual research groups. We discuss here, a simple database system which we call 'Bookshelf', that uses python in conjunction with a mysql database to provide an extremely simple system for curating and keeping track of molecular simulation data. It provides a user-friendly, scriptable solution to the common problem amongst biomolecular simulation laboratories; the storage, logging and subsequent retrieval of large numbers of simulations. Download URL: http://sbcb.bioch.ox.ac.uk/bookshelf/

  18. Discovering biomedical semantic relations in PubMed queries for information retrieval and database curation.

    Science.gov (United States)

    Huang, Chung-Chi; Lu, Zhiyong

    2016-01-01

    Identifying relevant papers from the literature is a common task in biocuration. Most current biomedical literature search systems primarily rely on matching user keywords. Semantic search, on the other hand, seeks to improve search accuracy by understanding the entities and contextual relations in user keywords. However, past research has mostly focused on semantically identifying biological entities (e.g. chemicals, diseases and genes) with little effort on discovering semantic relations. In this work, we aim to discover biomedical semantic relations in PubMed queries in an automated and unsupervised fashion. Specifically, we focus on extracting and understanding the contextual information (or context patterns) that is used by PubMed users to represent semantic relations between entities such as 'CHEMICAL-1 compared to CHEMICAL-2' With the advances in automatic named entity recognition, we first tag entities in PubMed queries and then use tagged entities as knowledge to recognize pattern semantics. More specifically, we transform PubMed queries into context patterns involving participating entities, which are subsequently projected to latent topics via latent semantic analysis (LSA) to avoid the data sparseness and specificity issues. Finally, we mine semantically similar contextual patterns or semantic relations based on LSA topic distributions. Our two separate evaluation experiments of chemical-chemical (CC) and chemical-disease (CD) relations show that the proposed approach significantly outperforms a baseline method, which simply measures pattern semantics by similarity in participating entities. The highest performance achieved by our approach is nearly 0.9 and 0.85 respectively for the CC and CD task when compared against the ground truth in terms of normalized discounted cumulative gain (nDCG), a standard measure of ranking quality. These results suggest that our approach can effectively identify and return related semantic patterns in a ranked order covering diverse bio-entity relations. To assess the potential utility of our automated top-ranked patterns of a given relation in semantic search, we performed a pilot study on frequently sought semantic relations in PubMed and observed improved literature retrieval effectiveness based on post-hoc human relevance evaluation. Further investigation in larger tests and in real-world scenarios is warranted. Published by Oxford University Press 2016. This work is written by US Government employees and is in the public domain in the US.

  19. Modern oncologic and operative outcomes for oesophageal cancer treated with curative intent.

    LENUS (Irish Health Repository)

    Reynolds, J V

    2011-09-01

    The curative approach to oesophageal cancer carries significant risks and a cure is achieved in approximately 20 per cent. There has been a recent trend internationally to observe improved operative and oncological outcomes. This report audits modern outcomes from a high volume centre with a prospective database for the period 2004-08. 603 patients were referred and 310 (52%) were treated with curative intent. Adenocarcinoma represented 68% of the cohort, squamous cell cancer 30%. Of the 310 cases, 227 (73%) underwent surgery, 105 (46%) underwent surgery alone, and 122 (54%) had chemotherapy or combination chemotherapy and radiation therapy. The postoperative mortality rate was 1.7%. The median and 5-year survival of the 310 patients based on intention to treat was 36 months and 36%, respectively, and of the 181 patients undergoing R0 resection, 52 months and 42%, respectively. An in-hospital postoperative mortality rate of less than 2 per cent, and 5-year survival of between 35 and 42% is consistent with benchmarks from international series.

  20. Curator's process of meaning-making in National museums

    DEFF Research Database (Denmark)

    Cole, Anne Jodon

    2014-01-01

    The paper aims to understand the meaning-making process curators engage in designing/developing exhibitions of the nations indigenous peoples. How indigenous people are represented can with perpetuate stereotypes or mediate change while strengthening their personal and group identity. Analysis...

  1. Interview with Smithsonian NASM Spacesuit Curator Dr. Cathleen Lewis

    Science.gov (United States)

    Lewis, Cathleen; Wright, Rebecca

    2012-01-01

    Dr. Cathleen Lewis was interviewed by Rebecca Wright during the presentation of an "Interview with Smithsonian NASM Spacesuit Curator Dr. Cathleen Lewis" on May 14, 2012. Topics included the care, size, and history of the spacesuit collection at the Smithsonian and the recent move to the state-of-the-art permanent storage facility at the Udvar-Hazy facility in Virginia.

  2. Learning relationships: Church of England curates and training ...

    African Journals Online (AJOL)

    2017-06-20

    Jun 20, 2017 ... exploring how this affects the dynamic of the relationship with their curates. Scripture is also ... factors, as employed in the models of personality advanced by Costa and .... psychological type preferences of their training incumbents. The data ..... to conceptualising and implementing Christian vocation.

  3. Collecting, curating, and researching writers' libraries a handbook

    CERN Document Server

    Oram, Richard W

    2014-01-01

    Collecting, Curating, and Researching Writers' Libraries: A Handbook is the first book to examine the history, acquisition, cataloging, and scholarly use of writers' personal libraries. This book also includes interviews with several well-known writers, who discuss their relationship with their books.

  4. Curative care through administration of plant-derived medicines in ...

    African Journals Online (AJOL)

    Curative care through administration of plant-derived medicines in Sekhukhune district municipality of Limpopo province, South Africa. ... Sources of medicine were mostly herbs followed by shrubs, trees, creepers and aloe collected from the communal land. The leaves, bark, roots and bulbs were prepared into decoctions ...

  5. Federal databases

    International Nuclear Information System (INIS)

    Welch, M.J.; Welles, B.W.

    1988-01-01

    Accident statistics on all modes of transportation are available as risk assessment analytical tools through several federal agencies. This paper reports on the examination of the accident databases by personal contact with the federal staff responsible for administration of the database programs. This activity, sponsored by the Department of Energy through Sandia National Laboratories, is an overview of the national accident data on highway, rail, air, and marine shipping. For each mode, the definition or reporting requirements of an accident are determined and the method of entering the accident data into the database is established. Availability of the database to others, ease of access, costs, and who to contact were prime questions to each of the database program managers. Additionally, how the agency uses the accident data was of major interest

  6. A computational platform to maintain and migrate manual functional annotations for BioCyc databases.

    Science.gov (United States)

    Walsh, Jesse R; Sen, Taner Z; Dickerson, Julie A

    2014-10-12

    BioCyc databases are an important resource for information on biological pathways and genomic data. Such databases represent the accumulation of biological data, some of which has been manually curated from literature. An essential feature of these databases is the continuing data integration as new knowledge is discovered. As functional annotations are improved, scalable methods are needed for curators to manage annotations without detailed knowledge of the specific design of the BioCyc database. We have developed CycTools, a software tool which allows curators to maintain functional annotations in a model organism database. This tool builds on existing software to improve and simplify annotation data imports of user provided data into BioCyc databases. Additionally, CycTools automatically resolves synonyms and alternate identifiers contained within the database into the appropriate internal identifiers. Automating steps in the manual data entry process can improve curation efforts for major biological databases. The functionality of CycTools is demonstrated by transferring GO term annotations from MaizeCyc to matching proteins in CornCyc, both maize metabolic pathway databases available at MaizeGDB, and by creating strain specific databases for metabolic engineering.

  7. Database Replication

    CERN Document Server

    Kemme, Bettina

    2010-01-01

    Database replication is widely used for fault-tolerance, scalability and performance. The failure of one database replica does not stop the system from working as available replicas can take over the tasks of the failed replica. Scalability can be achieved by distributing the load across all replicas, and adding new replicas should the load increase. Finally, database replication can provide fast local access, even if clients are geographically distributed clients, if data copies are located close to clients. Despite its advantages, replication is not a straightforward technique to apply, and

  8. Refactoring databases evolutionary database design

    CERN Document Server

    Ambler, Scott W

    2006-01-01

    Refactoring has proven its value in a wide range of development projects–helping software professionals improve system designs, maintainability, extensibility, and performance. Now, for the first time, leading agile methodologist Scott Ambler and renowned consultant Pramodkumar Sadalage introduce powerful refactoring techniques specifically designed for database systems. Ambler and Sadalage demonstrate how small changes to table structures, data, stored procedures, and triggers can significantly enhance virtually any database design–without changing semantics. You’ll learn how to evolve database schemas in step with source code–and become far more effective in projects relying on iterative, agile methodologies. This comprehensive guide and reference helps you overcome the practical obstacles to refactoring real-world databases by covering every fundamental concept underlying database refactoring. Using start-to-finish examples, the authors walk you through refactoring simple standalone databas...

  9. Lesson Learning at JPL

    Science.gov (United States)

    Oberhettinger, David

    2011-01-01

    A lessons learned system is a hallmark of a mature engineering organization A formal lessons learned process can help assure that valuable lessons get written and published, that they are well-written, and that the essential information is "infused" into institutional practice. Requires high-level institutional commitment, and everyone's participation in gathering, disseminating, and using the lessons

  10. RDD Databases

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This database was established to oversee documents issued in support of fishery research activities including experimental fishing permits (EFP), letters of...

  11. Snowstorm Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Snowstorm Database is a collection of over 500 snowstorms dating back to 1900 and updated operationally. Only storms having large areas of heavy snowfall (10-20...

  12. Dealer Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The dealer reporting databases contain the primary data reported by federally permitted seafood dealers in the northeast. Electronic reporting was implemented May 1,...

  13. ALFRED: An Allele Frequency Database for Microevolutionary Studies

    Directory of Open Access Journals (Sweden)

    Kenneth K Kidd

    2005-01-01

    Full Text Available Many kinds of microevolutionary studies require data on multiple polymorphisms in multiple populations. Increasingly, and especially for human populations, multiple research groups collect relevant data and those data are dispersed widely in the literature. ALFRED has been designed to hold data from many sources and make them available over the web. Data are assembled from multiple sources, curated, and entered into the database. Multiple links to other resources are also established by the curators. A variety of search options are available and additional geographic based interfaces are being developed. The database can serve the human anthropologic genetic community by identifying what loci are already typed on many populations thereby helping to focus efforts on a common set of markers. The database can also serve as a model for databases handling similar DNA polymorphism data for other species.

  14. National database

    DEFF Research Database (Denmark)

    Kristensen, Helen Grundtvig; Stjernø, Henrik

    1995-01-01

    Artikel om national database for sygeplejeforskning oprettet på Dansk Institut for Sundheds- og Sygeplejeforskning. Det er målet med databasen at samle viden om forsknings- og udviklingsaktiviteter inden for sygeplejen.......Artikel om national database for sygeplejeforskning oprettet på Dansk Institut for Sundheds- og Sygeplejeforskning. Det er målet med databasen at samle viden om forsknings- og udviklingsaktiviteter inden for sygeplejen....

  15. Comprehensive curation and analysis of global interaction networks in Saccharomyces cerevisiae

    Science.gov (United States)

    Reguly, Teresa; Breitkreutz, Ashton; Boucher, Lorrie; Breitkreutz, Bobby-Joe; Hon, Gary C; Myers, Chad L; Parsons, Ainslie; Friesen, Helena; Oughtred, Rose; Tong, Amy; Stark, Chris; Ho, Yuen; Botstein, David; Andrews, Brenda; Boone, Charles; Troyanskya, Olga G; Ideker, Trey; Dolinski, Kara; Batada, Nizar N; Tyers, Mike

    2006-01-01

    Background The study of complex biological networks and prediction of gene function has been enabled by high-throughput (HTP) methods for detection of genetic and protein interactions. Sparse coverage in HTP datasets may, however, distort network properties and confound predictions. Although a vast number of well substantiated interactions are recorded in the scientific literature, these data have not yet been distilled into networks that enable system-level inference. Results We describe here a comprehensive database of genetic and protein interactions, and associated experimental evidence, for the budding yeast Saccharomyces cerevisiae, as manually curated from over 31,793 abstracts and online publications. This literature-curated (LC) dataset contains 33,311 interactions, on the order of all extant HTP datasets combined. Surprisingly, HTP protein-interaction datasets currently achieve only around 14% coverage of the interactions in the literature. The LC network nevertheless shares attributes with HTP networks, including scale-free connectivity and correlations between interactions, abundance, localization, and expression. We find that essential genes or proteins are enriched for interactions with other essential genes or proteins, suggesting that the global network may be functionally unified. This interconnectivity is supported by a substantial overlap of protein and genetic interactions in the LC dataset. We show that the LC dataset considerably improves the predictive power of network-analysis approaches. The full LC dataset is available at the BioGRID () and SGD () databases. Conclusion Comprehensive datasets of biological interactions derived from the primary literature provide critical benchmarks for HTP methods, augment functional prediction, and reveal system-level attributes of biological networks. PMID:16762047

  16. Advanced Curation: Solving Current and Future Sample Return Problems

    Science.gov (United States)

    Fries, M.; Calaway, M.; Evans, C.; McCubbin, F.

    2015-01-01

    Advanced Curation is a wide-ranging and comprehensive research and development effort at NASA Johnson Space Center that identifies and remediates sample related issues. For current collections, Advanced Curation investigates new cleaning, verification, and analytical techniques to assess their suitability for improving curation processes. Specific needs are also assessed for future sample return missions. For each need, a written plan is drawn up to achieve the requirement. The plan draws while upon current Curation practices, input from Curators, the analytical expertise of the Astromaterials Research and Exploration Science (ARES) team, and suitable standards maintained by ISO, IEST, NIST and other institutions. Additionally, new technologies are adopted on the bases of need and availability. Implementation plans are tested using customized trial programs with statistically robust courses of measurement, and are iterated if necessary until an implementable protocol is established. Upcoming and potential NASA missions such as OSIRIS-REx, the Asteroid Retrieval Mission (ARM), sample return missions in the New Frontiers program, and Mars sample return (MSR) all feature new difficulties and specialized sample handling requirements. The Mars 2020 mission in particular poses a suite of challenges since the mission will cache martian samples for possible return to Earth. In anticipation of future MSR, the following problems are among those under investigation: What is the most efficient means to achieve the less than 1.0 ng/sq cm total organic carbon (TOC) cleanliness required for all sample handling hardware? How do we maintain and verify cleanliness at this level? The Mars 2020 Organic Contamination Panel (OCP) predicts that organic carbon, if present, will be present at the "one to tens" of ppb level in martian near-surface samples. The same samples will likely contain wt% perchlorate salts, or approximately 1,000,000x as much perchlorate oxidizer as organic carbon

  17. Advanced Curation Protocols for Mars Returned Sample Handling

    Science.gov (United States)

    Bell, M.; Mickelson, E.; Lindstrom, D.; Allton, J.

    Introduction: Johnson Space Center has over 30 years experience handling precious samples which include Lunar rocks and Antarctic meteorites. However, we recognize that future curation of samples from such missions as Genesis, Stardust, and Mars S mple Return, will require a high degree of biosafety combined witha extremely low levels of inorganic, organic, and biological contamination. To satisfy these requirements, research in the JSC Advanced Curation Lab is currently focused toward two major areas: preliminary examination techniques and cleaning and verification techniques . Preliminary Examination Techniques : In order to minimize the number of paths for contamination we are exploring the synergy between human &robotic sample handling in a controlled environment to help determine the limits of clean curation. Within the Advanced Curation Laboratory is a prototype, next-generation glovebox, which contains a robotic micromanipulator. The remotely operated manipulator has six degrees-of- freedom and can be programmed to perform repetitive sample handling tasks. Protocols are being tested and developed to perform curation tasks such as rock splitting, weighing, imaging, and storing. Techniques for sample transfer enabling more detailed remote examination without compromising the integrity of sample science are also being developed . The glovebox is equipped with a rapid transfer port through which samples can be passed without exposure. The transfer is accomplished by using a unique seal and engagement system which allows passage between containers while maintaining a first seal to the outside environment and a second seal to prevent the outside of the container cover and port door from becoming contaminated by the material being transferred. Cleaning and Verification Techniques: As part of the contamination control effort, innovative cleaning techniques are being identified and evaluated in conjunction with sensitive cleanliness verification methods. Towards this

  18. Content curation en periodismo (y en documentación periodística)

    OpenAIRE

    Guallar, Javier

    2014-01-01

    The last years we have seen the appearance of concepts such as content curation and content curator as, respectively, the activity or system and the professional or specialist. Although the term is originally linked to the world of marketing, -considering marketer's Rohit Bhargava 'Manifesto for the content curator' (2009) as its founding article-, and its features mostly identify with those of the Information Science professional, content curation goes beyond a specific discipline or profess...

  19. Curated routes: the project of developing experiential tracks in sub-urban landscape

    OpenAIRE

    Papathanasiou, Maximi; Uyttenhove, Pieter

    2015-01-01

    The Curated Routes project reflects on the visiting routes’ ability to make apparent the internal characteristics of urban environments. The project’s name allude to the intellectual function of curation and the materiality of routes. Curate deals with the practice of arranging material –tangible or intangible- in a way that a new understanding of an area is revealed. The word routes refers to the linear associations that link places and guide movement. The Curated Routes aim to reinforce the...

  20. The PDS4 Information Model and its Role in Agile Science Data Curation

    Science.gov (United States)

    Hughes, J. S.; Crichton, D.

    2017-12-01

    PDS4 is an information model-driven service architecture supporting the capture, management, distribution and integration of massive planetary science data captured in distributed data archives world-wide. The PDS4 Information Model (IM), the core element of the architecture, was developed using lessons learned from 20 years of archiving Planetary Science Data and best practices for information model development. The foundational principles were adopted from the Open Archival Information System (OAIS) Reference Model (ISO 14721), the Metadata Registry Specification (ISO/IEC 11179), and W3C XML (Extensible Markup Language) specifications. These provided respectively an object oriented model for archive information systems, a comprehensive schema for data dictionaries and hierarchical governance, and rules for rules for encoding documents electronically. The PDS4 Information model is unique in that it drives the PDS4 infrastructure by providing the representation of concepts and their relationships, constraints, rules, and operations; a sharable, stable, and organized set of information requirements; and machine parsable definitions that are suitable for configuring and generating code. This presentation will provide an over of the PDS4 Information Model and how it is being leveraged to develop and evolve the PDS4 infrastructure and enable agile curation of over 30 years of science data collected by the international Planetary Science community.

  1. The Princeton Protein Orthology Database (P-POD): a comparative genomics analysis tool for biologists.

    OpenAIRE

    Sven Heinicke; Michael S Livstone; Charles Lu; Rose Oughtred; Fan Kang; Samuel V Angiuoli; Owen White; David Botstein; Kara Dolinski

    2007-01-01

    Many biological databases that provide comparative genomics information and tools are now available on the internet. While certainly quite useful, to our knowledge none of the existing databases combine results from multiple comparative genomics methods with manually curated information from the literature. Here we describe the Princeton Protein Orthology Database (P-POD, http://ortholog.princeton.edu), a user-friendly database system that allows users to find and visualize the phylogenetic r...

  2. Digital curation: a proposal of a semi-automatic digital object selection-based model for digital curation in Big Data environments

    Directory of Open Access Journals (Sweden)

    Moisés Lima Dutra

    2016-08-01

    Full Text Available Introduction: This work presents a new approach for Digital Curations from a Big Data perspective. Objective: The objective is to propose techniques to digital curations for selecting and evaluating digital objects that take into account volume, velocity, variety, reality, and the value of the data collected from multiple knowledge domains. Methodology: This is an exploratory research of applied nature, which addresses the research problem in a qualitative way. Heuristics allow this semi-automatic process to be done either by human curators or by software agents. Results: As a result, it was proposed a model for searching, processing, evaluating and selecting digital objects to be processed by digital curations. Conclusions: It is possible to use Big Data environments as a source of information resources for Digital Curation; besides, Big Data techniques and tools can support the search and selection process of information resources by Digital Curations.

  3. Use of Ontologies for Data Integration and Curation

    Directory of Open Access Journals (Sweden)

    Judith Gelernter

    2011-03-01

    Full Text Available Data curation includes the goal of facilitating the re-use and combination of datasets, which is often impeded by incompatible data schema. Can we use ontologies to help with data integration? We suggest a semi-automatic process that involves the use of automatic text searching to help identify overlaps in metadata that accompany data schemas, plus human validation of suggested data matches.Problems include different text used to describe the same concept, different forms of data recording and different organizations of data. Ontologies can help by focussing attention on important words, providing synonyms to assist matching, and indicating in what context words are used. Beyond ontologies, data on the statistical behavior of data can be used to decide which data elements appear to be compatible with which other data elements. When curating data which may have hundreds or even thousands of data labels, semi-automatic assistance with data fusion should be of great help.

  4. A Practice and Value Proposal for Doctoral Dissertation Data Curation

    Directory of Open Access Journals (Sweden)

    W. Aaron Collie

    2011-10-01

    Full Text Available The preparation and publication of dissertations can be viewed as a subsystem of scholarly communication, and the treatment of data that support doctoral research can be mapped in a very controlled manner to the data curation lifecycle. Dissertation datasets represent “low-hanging fruit” for universities who are developing institutional data collections. The current workflow for processing electronic theses and dissertations (ETD at a typical American university is presented, and a new practice is proposed that includes datasets in the process of formulating, awarding, and disseminating dissertations in a way that enables them to be linked and curated together. The value proposition and new roles for the university and its student-authors, faculty, graduate programs and librarians are explored.

  5. The latest evidence for possible HIV-1 curative strategies.

    Science.gov (United States)

    Pham, Hanh Thi; Mesplède, Thibault

    2018-01-01

    Human immunodeficiency virus type 1 (HIV-1) infection remains a major health issue worldwide. In developed countries, antiretroviral therapy has extended its reach from treatment of people living with HIV-1 to post-exposure prophylaxis, treatment as prevention, and, more recently, pre-exposure prophylaxis. These healthcare strategies offer the epidemiological tools to curve the epidemic in rich settings and will be concomitantly implemented in developing countries. One of the remaining challenges is to identify an efficacious curative strategy. This review manuscript will focus on some of the current curative strategies aiming at providing a sterilizing or functional cure to HIV-1-positive individuals. These include the following: early treatment initiation in post-treatment controllers as a long-term HIV-1 remission strategy, latency reversal, gene editing with or without stem cell transplantation, and antibodies against either the viral envelope protein or the host integrin α4β7.

  6. Curation of US Martian Meteorites Collected in Antarctica

    Science.gov (United States)

    Lindstrom, M.; Satterwhite, C.; Allton, J.; Stansbury, E.

    1998-01-01

    To date the ANSMET field team has collected five martian meteorites (see below) in Antarctica and returned them for curation at the Johnson Space Center (JSC) Meteorite Processing Laboratory (MPL). ne meteorites were collected with the clean procedures used by ANSMET in collecting all meteorites: They were handled with JSC-cleaned tools, packaged in clean bags, and shipped frozen to JSC. The five martian meteorites vary significantly in size (12-7942 g) and rock type (basalts, lherzolites, and orthopyroxenite). Detailed descriptions are provided in the Mars Meteorite compendium, which describes classification, curation and research results. A table gives the names, classifications and original and curatorial masses of the martian meteorites. The MPL and measures for contamination control are described.

  7. Data Preservation and Curation for the Planetary Science Community

    Science.gov (United States)

    Hughes, J. S.; Crichton, D. J.; Joyner, R.; Hardman, S.; Rye, E.

    2013-12-01

    The Planetary Data System (PDS) has just released PDS4 Version 1.0, its next generation data standards for the planetary science archive. These data standards are the result of a multi-year effort to develop an information model based on accepted standards for data preservation, data curation, metadata management, and model development. The resulting information model is subsequently used to drive information system development from the generation of data standards documentation to the configuration of federated registries and search engines. This paper will provide an overview of the development of the PDS4 Information Model and focus on the application of the Open Archive Information System (OAIS) Reference Model - ISO 14721:2003, the Metadata Registry (MDR) Standard - ISO/IEC 11179, and the E-Business XML Standard to help ensure the long-term preservation and curation of planetary science data. Copyright 2013 California Institute of Technology Government sponsorship acknowledged

  8. Local high voltage radiotherapy with curative intent for prostatic carcinoma

    International Nuclear Information System (INIS)

    Jacobi, G.H.; Kurth, K.H.; Hohenfellner, R.

    1979-01-01

    In a 10-year interval 179 patients with prostatic carcinoma were treated by cobalt-60 teletherapy (7600 R). A selected group of 47 patients with localized disease and irradiated with curative intent had serial prostatic biopsies and were analized after a minimum follow-up of 1 year. Biopsies of half of the patients rendered definitively negative, on an average 14 months after radiotherapy. 8 patients with initial negative biopsy changed to positive secondarily. In one third of the patients histological conversion was missed, considered as radiation persister. Persistent carcinoma were of predominant low grade. 5 patients developed distant metastases 30 months after irradiation on an average. These patients had persistent positive tissue studies. Over all cumulative 5-years survival was 89%. In patients with prostatic carcinoma and local high voltage radiotherapy with curative intent (stage A through C) serial prostatic biopsies to document therapy effect seen mandatory. (orig.) 891 AJ/orig. 892 BRE [de

  9. Experiment Databases

    Science.gov (United States)

    Vanschoren, Joaquin; Blockeel, Hendrik

    Next to running machine learning algorithms based on inductive queries, much can be learned by immediately querying the combined results of many prior studies. Indeed, all around the globe, thousands of machine learning experiments are being executed on a daily basis, generating a constant stream of empirical information on machine learning techniques. While the information contained in these experiments might have many uses beyond their original intent, results are typically described very concisely in papers and discarded afterwards. If we properly store and organize these results in central databases, they can be immediately reused for further analysis, thus boosting future research. In this chapter, we propose the use of experiment databases: databases designed to collect all the necessary details of these experiments, and to intelligently organize them in online repositories to enable fast and thorough analysis of a myriad of collected results. They constitute an additional, queriable source of empirical meta-data based on principled descriptions of algorithm executions, without reimplementing the algorithms in an inductive database. As such, they engender a very dynamic, collaborative approach to experimentation, in which experiments can be freely shared, linked together, and immediately reused by researchers all over the world. They can be set up for personal use, to share results within a lab or to create open, community-wide repositories. Here, we provide a high-level overview of their design, and use an existing experiment database to answer various interesting research questions about machine learning algorithms and to verify a number of recent studies.

  10. User-generated content curation with deep convolutional neural networks

    OpenAIRE

    Tous Liesa, Rubén; Wust, Otto; Gómez, Mauro; Poveda, Jonatan; Elena, Marc; Torres Viñals, Jordi; Makni, Mouna; Ayguadé Parra, Eduard

    2016-01-01

    In this paper, we report a work consisting in using deep convolutional neural networks (CNNs) for curating and filtering photos posted by social media users (Instagram and Twitter). The final goal is to facilitate searching and discovering user-generated content (UGC) with potential value for digital marketing tasks. The images are captured in real time and automatically annotated with multiple CNNs. Some of the CNNs perform generic object recognition tasks while others perform what we call v...

  11. Sharing Responsibility for Data Stewardship Between Scientists and Curators

    Science.gov (United States)

    Hedstrom, M. L.

    2012-12-01

    Data stewardship is becoming increasingly important to support accurate conclusions from new forms of data, integration of and computation across heterogeneous data types, interactions between models and data, replication of results, data governance and long-term archiving. In addition to increasing recognition of the importance of data management, data science, and data curation by US and international scientific agencies, the National Academies of Science Board on Research Data and Information is sponsoring a study on Data Curation Education and Workforce Issues. Effective data stewardship requires a distributed effort among scientists who produce data, IT staff and/or vendors who provide data storage and computational facilities and services, and curators who enhance data quality, manage data governance, provide access to third parties, and assume responsibility for long-term archiving of data. The expertise necessary for scientific data management includes a mix of knowledge of the scientific domain; an understanding of domain data requirements, standards, ontologies and analytical methods; facility with leading edge information technology; and knowledge of data governance, standards, and best practices for long-term preservation and access that rarely are found in a single individual. Rather than developing data science and data curation as new and distinct occupations, this paper examines the set of tasks required for data stewardship. The paper proposes an alternative model that embeds data stewardship in scientific workflows and coordinates hand-offs between instruments, repositories, analytical processing, publishers, distributors, and archives. This model forms the basis for defining knowledge and skill requirements for specific actors in the processes required for data stewardship and the corresponding educational and training needs.

  12. Curative resection of transverse colon cancer via minilaparotomy.

    Science.gov (United States)

    Ishida, Hideyuki; Ishiguro, Tohru; Ishibashi, Keiichiro; Ohsawa, Tomonori; Okada, Norimichi; Kumamoto, Kensuke; Haga, Norihiro

    2011-01-01

    Minilaparotomy has been reported to be a minimally invasive alternative to laparoscopically assisted surgery. We retrospectively evaluated the usefulness of minilaparotomy for the resection of transverse colon cancer, which has generally been considered difficult to resect laparoscopically. Patients for whom curative resection was attempted for transverse colon cancer (n = 21) or sigmoid colon cancer (n = 81) via minilaparotomy (skin incision, transverse colon cancer as well as those with sigmoid colon cancer.

  13. A relevancy algorithm for curating earth science data around phenomenon

    Science.gov (United States)

    Maskey, Manil; Ramachandran, Rahul; Li, Xiang; Weigel, Amanda; Bugbee, Kaylin; Gatlin, Patrick; Miller, J. J.

    2017-09-01

    Earth science data are being collected for various science needs and applications, processed using different algorithms at multiple resolutions and coverages, and then archived at different archiving centers for distribution and stewardship causing difficulty in data discovery. Curation, which typically occurs in museums, art galleries, and libraries, is traditionally defined as the process of collecting and organizing information around a common subject matter or a topic of interest. Curating data sets around topics or areas of interest addresses some of the data discovery needs in the field of Earth science, especially for unanticipated users of data. This paper describes a methodology to automate search and selection of data around specific phenomena. Different components of the methodology including the assumptions, the process, and the relevancy ranking algorithm are described. The paper makes two unique contributions to improving data search and discovery capabilities. First, the paper describes a novel methodology developed for automatically curating data around a topic using Earth science metadata records. Second, the methodology has been implemented as a stand-alone web service that is utilized to augment search and usability of data in a variety of tools.

  14. Gramene database: Navigating plant comparative genomics resources

    Directory of Open Access Journals (Sweden)

    Parul Gupta

    2016-11-01

    Full Text Available Gramene (http://www.gramene.org is an online, open source, curated resource for plant comparative genomics and pathway analysis designed to support researchers working in plant genomics, breeding, evolutionary biology, system biology, and metabolic engineering. It exploits phylogenetic relationships to enrich the annotation of genomic data and provides tools to perform powerful comparative analyses across a wide spectrum of plant species. It consists of an integrated portal for querying, visualizing and analyzing data for 44 plant reference genomes, genetic variation data sets for 12 species, expression data for 16 species, curated rice pathways and orthology-based pathway projections for 66 plant species including various crops. Here we briefly describe the functions and uses of the Gramene database.

  15. The plant phenological online database (PPODB): an online database for long-term phenological data

    Science.gov (United States)

    Dierenbach, Jonas; Badeck, Franz-W.; Schaber, Jörg

    2013-09-01

    We present an online database that provides unrestricted and free access to over 16 million plant phenological observations from over 8,000 stations in Central Europe between the years 1880 and 2009. Unique features are (1) a flexible and unrestricted access to a full-fledged database, allowing for a wide range of individual queries and data retrieval, (2) historical data for Germany before 1951 ranging back to 1880, and (3) more than 480 curated long-term time series covering more than 100 years for individual phenological phases and plants combined over Natural Regions in Germany. Time series for single stations or Natural Regions can be accessed through a user-friendly graphical geo-referenced interface. The joint databases made available with the plant phenological database PPODB render accessible an important data source for further analyses of long-term changes in phenology. The database can be accessed via www.ppodb.de .

  16. Opening Data in the Long Tail for Community Discovery, Curation and Action Using Active and Social Curation

    Science.gov (United States)

    Hedstrom, M. L.; Kumar, P.; Myers, J.; Plale, B. A.

    2012-12-01

    In data science, the most common sequence of steps for data curation are to 1) curate data, 2) enable data discovery, and 3) provide for data reuse. The Sustainable Environments - Actionable Data (SEAD) project, funded through NSF's DataNet program, is creating an environment for sustainability scientists to discover data first, reuse data next, and curate data though an on-going process that we call Active and Social Curation. For active curation we are developing tools and services that support data discovery, data management, and data enhancement for the community while the data is still being used actively for research. We are creating an Active Content Repository, using drop box, semantic web technologies, and a Flickr-like interface for researchers to "drop" data into a repository where it will be replicated and minimally discoverable. For social curation, we are deploying a social networking tool, VIVO, which will allow researchers to discover data-publications-people (e.g. expertise) through a route that can start at any of those entry points. The other dimension of social curation is developing mechanisms to open data for community input, for example, using ranking and commenting mechanisms for data sets and a community-sourcing capability to add tags, clean up and validate data sets. SEAD's strategies and services are aimed at the sustainability science community, which faces numerous challenges including discovery of useful data, cleaning noisy observational data, synthesizing data of different types, defining appropriate models, managing and preserving their research data, and conveying holistic results to colleagues, students, decision makers, and the public. Sustainability researchers make significant use of centrally managed data from satellites and national sensor networks, national scientific and statistical agencies, and data archives. At the same time, locally collected data and custom derived data products that combine observations and

  17. Reducing Organic Contamination in NASA JSC Astromaterial Curation Facility

    Science.gov (United States)

    Calaway, M. J.; Allen, C. C.; Allton, J. H.

    2013-01-01

    Future robotic and human spaceflight missions to the Moon, Mars, asteroids and comets will require handling and storing astromaterial samples with minimal inorganic and organic contamination to preserve the scientific integrity of each sample. Much was learned from the rigorous attempts to minimize and monitor organic contamination during Apollo, but it was not adequate for current analytical requirements; thus [1]. OSIRIS-REx, Hayabusa-2, and future Mars sample return will require better protocols for reducing organic contamination. Future isolation con-tainment systems for astromaterials, possibly nitrogen enriched gloveboxes, must be able to reduce organic and inorganic cross-contamination. In 2012, a baseline study established the current state of organic cleanliness in gloveboxes used by NASA JSC astromaterials curation labs that could be used as a benchmark for future mission designs [2, 3]. After standard ultra-pure water (UPW) cleaning, the majority of organic contaminates found were hydrocarbons, plasticizers, silicones, and solvents. Hydro-carbons loads (> C7) ranged from 1.9 to 11.8 ng/cm2 for TD-GC-MS wafer exposure analyses and 5.0 to 19.5 ng/L for TD-GC-MS adsorbent tube exposure. Plasticizers included peracetic acid sterilization were used in the atmospheric de-contamination (R) cabinets. Later, Lunar curation gloveboxes were degreased with a pressurized Freon 113 wash. Today, UPW has replaced Freon as the standard cleaning procedure, but does not have the degreasing solvency power of Freon. Future Cleaning Studies: Cleaning experiments are cur-rently being orchestrated to study how to degrease and reduce organics in a JSC curation glovebox lower than the established baseline. Several new chemicals in the industry have replaced traditional degreasing solvents such as Freon and others that are now federally restricted. However, these new suites of chemicals remain untested for lowering organics in curation gloveboxes. 3M's HFE-7100DL and Du

  18. The Hayabusa Curation Facility at Johnson Space Center

    Science.gov (United States)

    Zolensky, M.; Bastien, R.; McCann, B.; Frank, D.; Gonzalez, C.; Rodriguez, M.

    2013-01-01

    The Japan Aerospace Exploration Agency (JAXA) Hayabusa spacecraft made contact with the asteroid 25143 Itokawa and collected regolith dust from Muses Sea region of smooth terrain [1]. The spacecraft returned to Earth with more than 10,000 grains ranging in size from just over 300 µm to less than 10 µm [2, 3]. These grains represent the only collection of material returned from an asteroid by a spacecraft. As part of the joint agreement between JAXA and NASA for the mission, 10% of the Hayabusa grains are being transferred to NASA for parallel curation and allocation. In order to properly receive process and curate these samples, a new curation facility was established at Johnson Space Center (JSC). Since the Hayabusa samples within the JAXA curation facility have been stored free from exposure to terrestrial atmosphere and contamination [4], one of the goals of the new NASA curation facility was to continue this treatment. An existing lab space at JSC was transformed into a 120 sq.ft. ISO class 4 (equivalent to the original class 10 standard) clean room. Hayabusa samples are stored, observed, processed, and packaged for allocation inside a stainless steel glove box under dry N2. Construction of the clean laboratory was completed in 2012. Currently, 25 Itokawa particles are lodged in NASA's Hayabusa Lab. Special care has been taken during lab construction to remove or contain materials that may contribute contaminant particles in the same size range as the Hayabusa grains. Several witness plates of various materials are installed around the clean lab and within the glove box to permit characterization of local contaminants at regular intervals by SEM and mass spectrometry, and particle counts of the lab environment are frequently acquired. Of particular interest is anodized aluminum, which contains copious sub-mm grains of a multitude of different materials embedded in its upper surface. Unfortunately the use of anodized aluminum was necessary in the construction

  19. A justification for semantic training in data curation frameworks development

    Science.gov (United States)

    Ma, X.; Branch, B. D.; Wegner, K.

    2013-12-01

    In the complex data curation activities involving proper data access, data use optimization and data rescue, opportunities exist where underlying skills in semantics may play a crucial role in data curation professionals ranging from data scientists, to informaticists, to librarians. Here, We provide a conceptualization of semantics use in the education data curation framework (EDCF) [1] under development by Purdue University and endorsed by the GLOBE program [2] for further development and application. Our work shows that a comprehensive data science training includes both spatial and non-spatial data, where both categories are promoted by standard efforts of organizations such as the Open Geospatial Consortium (OGC) and the World Wide Web Consortium (W3C), as well as organizations such as the Federation of Earth Science Information Partners (ESIP) that share knowledge and propagate best practices in applications. Outside the context of EDCF, semantics training may be same critical to such data scientists, informaticists or librarians in other types of data curation activity. Past works by the authors have suggested that such data science should augment an ontological literacy where data science may become sustainable as a discipline. As more datasets are being published as open data [3] and made linked to each other, i.e., in the Resource Description Framework (RDF) format, or at least their metadata are being published in such a way, vocabularies and ontologies of various domains are being created and used in the data management, such as the AGROVOC [4] for agriculture and the GCMD keywords [5] and CLEAN vocabulary [6] for climate sciences. The new generation of data scientist should be aware of those technologies and receive training where appropriate to incorporate those technologies into their reforming daily works. References [1] Branch, B.D., Fosmire, M., 2012. The role of interdisciplinary GIS and data curation librarians in enhancing authentic scientific

  20. The Importance of Contamination Knowledge in Curation - Insights into Mars Sample Return

    Science.gov (United States)

    Harrington, A. D.; Calaway, M. J.; Regberg, A. B.; Mitchell, J. L.; Fries, M. D.; Zeigler, R. A.; McCubbin, F. M.

    2018-01-01

    The Astromaterials Acquisition and Curation Office at NASA Johnson Space Center (JSC), in Houston, TX (henceforth Curation Office) manages the curation of extraterrestrial samples returned by NASA missions and shared collections from international partners, preserving their integrity for future scientific study while providing the samples to the international community in a fair and unbiased way. The Curation Office also curates flight and non-flight reference materials and other materials from spacecraft assembly (e.g., lubricants, paints and gases) of sample return missions that would have the potential to cross-contaminate a present or future NASA astromaterials collection.

  1. Reactome graph database: Efficient access to complex pathway data.

    Directory of Open Access Journals (Sweden)

    Antonio Fabregat

    2018-01-01

    Full Text Available Reactome is a free, open-source, open-data, curated and peer-reviewed knowledgebase of biomolecular pathways. One of its main priorities is to provide easy and efficient access to its high quality curated data. At present, biological pathway databases typically store their contents in relational databases. This limits access efficiency because there are performance issues associated with queries traversing highly interconnected data. The same data in a graph database can be queried more efficiently. Here we present the rationale behind the adoption of a graph database (Neo4j as well as the new ContentService (REST API that provides access to these data. The Neo4j graph database and its query language, Cypher, provide efficient access to the complex Reactome data model, facilitating easy traversal and knowledge discovery. The adoption of this technology greatly improved query efficiency, reducing the average query time by 93%. The web service built on top of the graph database provides programmatic access to Reactome data by object oriented queries, but also supports more complex queries that take advantage of the new underlying graph-based data storage. By adopting graph database technology we are providing a high performance pathway data resource to the community. The Reactome graph database use case shows the power of NoSQL database engines for complex biological data types.

  2. Reactome graph database: Efficient access to complex pathway data

    Science.gov (United States)

    Korninger, Florian; Viteri, Guilherme; Marin-Garcia, Pablo; Ping, Peipei; Wu, Guanming; Stein, Lincoln; D’Eustachio, Peter

    2018-01-01

    Reactome is a free, open-source, open-data, curated and peer-reviewed knowledgebase of biomolecular pathways. One of its main priorities is to provide easy and efficient access to its high quality curated data. At present, biological pathway databases typically store their contents in relational databases. This limits access efficiency because there are performance issues associated with queries traversing highly interconnected data. The same data in a graph database can be queried more efficiently. Here we present the rationale behind the adoption of a graph database (Neo4j) as well as the new ContentService (REST API) that provides access to these data. The Neo4j graph database and its query language, Cypher, provide efficient access to the complex Reactome data model, facilitating easy traversal and knowledge discovery. The adoption of this technology greatly improved query efficiency, reducing the average query time by 93%. The web service built on top of the graph database provides programmatic access to Reactome data by object oriented queries, but also supports more complex queries that take advantage of the new underlying graph-based data storage. By adopting graph database technology we are providing a high performance pathway data resource to the community. The Reactome graph database use case shows the power of NoSQL database engines for complex biological data types. PMID:29377902

  3. Reactome graph database: Efficient access to complex pathway data.

    Science.gov (United States)

    Fabregat, Antonio; Korninger, Florian; Viteri, Guilherme; Sidiropoulos, Konstantinos; Marin-Garcia, Pablo; Ping, Peipei; Wu, Guanming; Stein, Lincoln; D'Eustachio, Peter; Hermjakob, Henning

    2018-01-01

    Reactome is a free, open-source, open-data, curated and peer-reviewed knowledgebase of biomolecular pathways. One of its main priorities is to provide easy and efficient access to its high quality curated data. At present, biological pathway databases typically store their contents in relational databases. This limits access efficiency because there are performance issues associated with queries traversing highly interconnected data. The same data in a graph database can be queried more efficiently. Here we present the rationale behind the adoption of a graph database (Neo4j) as well as the new ContentService (REST API) that provides access to these data. The Neo4j graph database and its query language, Cypher, provide efficient access to the complex Reactome data model, facilitating easy traversal and knowledge discovery. The adoption of this technology greatly improved query efficiency, reducing the average query time by 93%. The web service built on top of the graph database provides programmatic access to Reactome data by object oriented queries, but also supports more complex queries that take advantage of the new underlying graph-based data storage. By adopting graph database technology we are providing a high performance pathway data resource to the community. The Reactome graph database use case shows the power of NoSQL database engines for complex biological data types.

  4. Thirty years after Alma-Ata: a systematic review of the impact of community health workers delivering curative interventions against malaria, pneumonia and diarrhoea on child mortality and morbidity in sub-Saharan Africa

    Directory of Open Access Journals (Sweden)

    Lewin Simon

    2011-10-01

    Full Text Available Abstract Background Over thirty years have passed since the Alma-Ata Declaration on primary health care in 1978. Many governments in the first decade following the declaration responded by developing national programmes of community health workers (CHWs, but evaluations of these often demonstrated poor outcomes. As many CHW programmes have responded to the HIV/AIDS pandemic, international interest in them has returned and their role in the response to other diseases should be examined carefully so that lessons can be applied to their new roles. Over half of the deaths in African children under five years of age are due to malaria, diarrhoea and pneumonia - a situation which could be addressed through the use of cheap and effective interventions delivered by CHWs. However, to date there is very little evidence from randomised controlled trials of the impacts of CHW programmes on child mortality in Africa. Evidence from non-randomised controlled studies has not previously been reviewed systematically. Methods We searched databases of published and unpublished studies for RCTs and non-randomised studies evaluating CHW programmes delivering curative treatments, with or without preventive components, for malaria, diarrhoea or pneumonia, in children in sub-Saharan Africa from 1987 to 2007. The impact of these programmes on morbidity or mortality in children under six years of age was reviewed. A descriptive analysis of interventional and contextual factors associated with these impacts was attempted. Results The review identified seven studies evaluating CHWs, delivering a range of interventions. Limited descriptive data on programmes, contexts or process outcomes for these CHW programmes were available. CHWs in national programmes achieved large mortality reductions of 63% and 36% respectively, when insecticide-treated nets and anti-malarial chemoprophylaxis were delivered, in addition to curative interventions. Conclusions CHW programmes could

  5. MicroScope—an integrated microbial resource for the curation and comparative analysis of genomic and metabolic data

    Science.gov (United States)

    Vallenet, David; Belda, Eugeni; Calteau, Alexandra; Cruveiller, Stéphane; Engelen, Stefan; Lajus, Aurélie; Le Fèvre, François; Longin, Cyrille; Mornico, Damien; Roche, David; Rouy, Zoé; Salvignol, Gregory; Scarpelli, Claude; Thil Smith, Adam Alexander; Weiman, Marion; Médigue, Claudine

    2013-01-01

    MicroScope is an integrated platform dedicated to both the methodical updating of microbial genome annotation and to comparative analysis. The resource provides data from completed and ongoing genome projects (automatic and expert annotations), together with data sources from post-genomic experiments (i.e. transcriptomics, mutant collections) allowing users to perfect and improve the understanding of gene functions. MicroScope (http://www.genoscope.cns.fr/agc/microscope) combines tools and graphical interfaces to analyse genomes and to perform the manual curation of gene annotations in a comparative context. Since its first publication in January 2006, the system (previously named MaGe for Magnifying Genomes) has been continuously extended both in terms of data content and analysis tools. The last update of MicroScope was published in 2009 in the Database journal. Today, the resource contains data for >1600 microbial genomes, of which ∼300 are manually curated and maintained by biologists (1200 personal accounts today). Expert annotations are continuously gathered in the MicroScope database (∼50 000 a year), contributing to the improvement of the quality of microbial genomes annotations. Improved data browsing and searching tools have been added, original tools useful in the context of expert annotation have been developed and integrated and the website has been significantly redesigned to be more user-friendly. Furthermore, in the context of the European project Microme (Framework Program 7 Collaborative Project), MicroScope is becoming a resource providing for the curation and analysis of both genomic and metabolic data. An increasing number of projects are related to the study of environmental bacterial (meta)genomes that are able to metabolize a large variety of chemical compounds that may be of high industrial interest. PMID:23193269

  6. Hayabusa Asteroidal Sample Preliminary Examination Team (HASPET) and the Astromaterial Curation Facility at JAXA/ISAS

    Science.gov (United States)

    Yano, H.; Fujiwara, A.

    After the successful launch in May 2003, the Hayabusa (MUSES-C) mission of JAXA/ISAS will collect surface materials (e.g., regolith) of several hundred mg to several g in total from the S-type near Earth asteroid (25143) Itokawa in late 2005 and bring them back to ground laboratories in the summer of 2007. The retrieved samples will be given initial analysis at the JAXA/ISAS astromaterial curation facility, which is currently in the preparation for its construction, by the Hayabusa Asteroidal Sample Preliminary Examination Team (HASPET). HASPET is consisted of the ISAS Hayabusa team, the international partners from NASA and Australia and all-Japan meteoritic scientists to be selected as outsourcing parts of the initial analyses. The initial analysis to characterize general aspects of returned samples can consume only 15 % of its total mass and must complete the whole analyses including the database building before international AO for detailed analyses within the maximum of 1 year. Confident exercise of non-destructive, micro-analyses whenever possible are thus vital for the HASPET analysis. In the purpose to survey what kinds and levels of micro-analysis techniques in respective fields, from major elements and mineralogy to trace and isotopic elements and organics, are available in Japan at present, ISAS has conducted the HASPET open competitions in 2000-01 and 2004. The initial evaluation was made by multiple domestic peer reviews. Applicants were then provided two kinds of unknown asteroid sample analogs in order to conduct proposed analysis with self-claimed amount of samples in self-claimed duration. After the completion of multiple, international peer reviews, the Selection Committee compiled evaluations and recommended the finalists of each round. The final members of the HASPET will be appointed about 2 years prior to the Earth return. Then they will conduct a test-run of the whole initial analysis procedures at the ISAS astromaterial curation facility and

  7. Refusal of Curative Radiation Therapy and Surgery Among Patients With Cancer

    Energy Technology Data Exchange (ETDEWEB)

    Aizer, Ayal A., E-mail: aaaizer@partners.org [Harvard Radiation Oncology Program, Boston, Massachusetts (United States); Chen, Ming-Hui [Department of Statistics, University of Connecticut, Storrs, Connecticut (United States); Parekh, Arti [Boston University School of Medicine, Boston, Massachusetts (United States); Choueiri, Toni K. [Lank Center for Genitourinary Oncology, Dana-Farber Cancer Institute/Brigham and Women' s Hospital, Harvard Medical School, Boston, Massachusetts (United States); Hoffman, Karen E. [Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas (United States); Kim, Simon P. [Department of Urology, Mayo Clinic, Rochester, Minnesota (United States); Martin, Neil E. [Lank Center for Genitourinary Oncology, Dana-Farber Cancer Institute/Brigham and Women' s Hospital, Harvard Medical School, Boston, Massachusetts (United States); Hu, Jim C. [Department of Urology, University of California, Los Angeles, California (United States); Trinh, Quoc-Dien [Cancer Prognostics and Health Outcomes Unit, University of Montreal Health Center, Montreal, Quebec (Canada); Nguyen, Paul L. [Lank Center for Genitourinary Oncology, Dana-Farber Cancer Institute/Brigham and Women' s Hospital, Harvard Medical School, Boston, Massachusetts (United States)

    2014-07-15

    Purpose: Surgery and radiation therapy represent the only curative options for many patients with solid malignancies. However, despite the recommendations of their physicians, some patients refuse these therapies. This study characterized factors associated with refusal of surgical or radiation therapy as well as the impact of refusal of recommended therapy on patients with localized malignancies. Methods and Materials: We used the Surveillance, Epidemiology, and End Results program to identify a population-based sample of 925,127 patients who had diagnoses of 1 of 8 common malignancies for which surgery and/or radiation are believed to confer a survival benefit between 1995 and 2008. Refusal of oncologic therapy, as documented in the SEER database, was the primary outcome measure. Multivariable logistic regression was used to investigate factors associated with refusal. The impact of refusal of therapy on cancer-specific mortality was assessed with Fine and Gray's competing risks regression. Results: In total, 2441 of 692,938 patients (0.4%) refused surgery, and 2113 of 232,189 patients (0.9%) refused radiation, despite the recommendations of their physicians. On multivariable analysis, advancing age, decreasing annual income, nonwhite race, and unmarried status were associated with refusal of surgery, whereas advancing age, decreasing annual income, Asian American race, and unmarried status were associated with refusal of radiation (P<.001 in all cases). Refusal of surgery and radiation were associated with increased estimates of cancer-specific mortality for all malignancies evaluated (hazard ratio [HR], 2.80, 95% confidence interval [CI], 2.59-3.03; P<.001 and HR 1.97 [95% CI, 1.78-2.18]; P<.001, respectively). Conclusions: Nonwhite, less affluent, and unmarried patients are more likely to refuse curative surgical and/or radiation-based oncologic therapy, raising concern that socioeconomic factors may drive some patients to forego potentially life

  8. Refusal of Curative Radiation Therapy and Surgery Among Patients With Cancer

    International Nuclear Information System (INIS)

    Aizer, Ayal A.; Chen, Ming-Hui; Parekh, Arti; Choueiri, Toni K.; Hoffman, Karen E.; Kim, Simon P.; Martin, Neil E.; Hu, Jim C.; Trinh, Quoc-Dien; Nguyen, Paul L.

    2014-01-01

    Purpose: Surgery and radiation therapy represent the only curative options for many patients with solid malignancies. However, despite the recommendations of their physicians, some patients refuse these therapies. This study characterized factors associated with refusal of surgical or radiation therapy as well as the impact of refusal of recommended therapy on patients with localized malignancies. Methods and Materials: We used the Surveillance, Epidemiology, and End Results program to identify a population-based sample of 925,127 patients who had diagnoses of 1 of 8 common malignancies for which surgery and/or radiation are believed to confer a survival benefit between 1995 and 2008. Refusal of oncologic therapy, as documented in the SEER database, was the primary outcome measure. Multivariable logistic regression was used to investigate factors associated with refusal. The impact of refusal of therapy on cancer-specific mortality was assessed with Fine and Gray's competing risks regression. Results: In total, 2441 of 692,938 patients (0.4%) refused surgery, and 2113 of 232,189 patients (0.9%) refused radiation, despite the recommendations of their physicians. On multivariable analysis, advancing age, decreasing annual income, nonwhite race, and unmarried status were associated with refusal of surgery, whereas advancing age, decreasing annual income, Asian American race, and unmarried status were associated with refusal of radiation (P<.001 in all cases). Refusal of surgery and radiation were associated with increased estimates of cancer-specific mortality for all malignancies evaluated (hazard ratio [HR], 2.80, 95% confidence interval [CI], 2.59-3.03; P<.001 and HR 1.97 [95% CI, 1.78-2.18]; P<.001, respectively). Conclusions: Nonwhite, less affluent, and unmarried patients are more likely to refuse curative surgical and/or radiation-based oncologic therapy, raising concern that socioeconomic factors may drive some patients to forego potentially life

  9. Refusal of curative radiation therapy and surgery among patients with cancer.

    Science.gov (United States)

    Aizer, Ayal A; Chen, Ming-Hui; Parekh, Arti; Choueiri, Toni K; Hoffman, Karen E; Kim, Simon P; Martin, Neil E; Hu, Jim C; Trinh, Quoc-Dien; Nguyen, Paul L

    2014-07-15

    Surgery and radiation therapy represent the only curative options for many patients with solid malignancies. However, despite the recommendations of their physicians, some patients refuse these therapies. This study characterized factors associated with refusal of surgical or radiation therapy as well as the impact of refusal of recommended therapy on patients with localized malignancies. We used the Surveillance, Epidemiology, and End Results program to identify a population-based sample of 925,127 patients who had diagnoses of 1 of 8 common malignancies for which surgery and/or radiation are believed to confer a survival benefit between 1995 and 2008. Refusal of oncologic therapy, as documented in the SEER database, was the primary outcome measure. Multivariable logistic regression was used to investigate factors associated with refusal. The impact of refusal of therapy on cancer-specific mortality was assessed with Fine and Gray's competing risks regression. In total, 2441 of 692,938 patients (0.4%) refused surgery, and 2113 of 232,189 patients (0.9%) refused radiation, despite the recommendations of their physicians. On multivariable analysis, advancing age, decreasing annual income, nonwhite race, and unmarried status were associated with refusal of surgery, whereas advancing age, decreasing annual income, Asian American race, and unmarried status were associated with refusal of radiation (PRefusal of surgery and radiation were associated with increased estimates of cancer-specific mortality for all malignancies evaluated (hazard ratio [HR], 2.80, 95% confidence interval [CI], 2.59-3.03; Prefuse curative surgical and/or radiation-based oncologic therapy, raising concern that socioeconomic factors may drive some patients to forego potentially life-saving care. Copyright © 2014 Elsevier Inc. All rights reserved.

  10. IAEA Radiation Events Database (RADEV)

    International Nuclear Information System (INIS)

    Wheatley, J.; Ortiz-Lopez, P.

    2001-01-01

    Whilst the use of ionizing radiation continues to bring benefits to many people throughout the world there is increasing concern at the number of reported accidents involving radiation. Such accidents have had an impact on the lives of patients, workers and members of the public, the consequences of which have ranged from trivial health effects to fatalities. In order to reduce the number of accidents and to mitigate their consequences it is, therefore, necessary to raise awareness of the causes of accidents and to note the lessons that can be learned. The IAEA's database on unusual radiation events (RADEV) is intended to provide a world-wide focal point for such information. (author)

  11. The Danish Bladder Cancer Database

    DEFF Research Database (Denmark)

    Hansen, Erik; Larsson, Heidi Jeanet; Nørgaard, Mette

    2016-01-01

    AIM OF DATABASE: The aim of the Danish Bladder Cancer Database (DaBlaCa-data) is to monitor the treatment of all patients diagnosed with invasive bladder cancer (BC) in Denmark. STUDY POPULATION: All patients diagnosed with BC in Denmark from 2012 onward were included in the study. Results......-intended radiation therapy. DESCRIPTIVE DATA: One-year mortality was 28% (95% confidence interval [CI]: 15-21). One-year cancer-specific mortality was 25% (95% CI: 22-27%). One-year mortality after cystectomy was 14% (95% CI: 10-18). Ninety-day mortality after cystectomy was 3% (95% CI: 1-5) in 2013. One......-year mortality following curative-intended radiation therapy was 32% (95% CI: 24-39) and 1-year cancer-specific mortality was 23% (95% CI: 16-31) in 2013. CONCLUSION: This preliminary DaBlaCa-data report showed that the treatment of MIBC in Denmark overall meet high international academic standards. The database...

  12. The STRING database in 2017

    DEFF Research Database (Denmark)

    Szklarczyk, Damian; Morris, John H; Cook, Helen

    2017-01-01

    A system-wide understanding of cellular function requires knowledge of all functional interactions between the expressed proteins. The STRING database aims to collect and integrate this information, by consolidating known and predicted protein-protein association data for a large number of organi......A system-wide understanding of cellular function requires knowledge of all functional interactions between the expressed proteins. The STRING database aims to collect and integrate this information, by consolidating known and predicted protein-protein association data for a large number...... of organisms. The associations in STRING include direct (physical) interactions, as well as indirect (functional) interactions, as long as both are specific and biologically meaningful. Apart from collecting and reassessing available experimental data on protein-protein interactions, and importing known...... pathways and protein complexes from curated databases, interaction predictions are derived from the following sources: (i) systematic co-expression analysis, (ii) detection of shared selective signals across genomes, (iii) automated text-mining of the scientific literature and (iv) computational transfer...

  13. Curating NASA's Astromaterials Collections: Past, Present, and Future

    Science.gov (United States)

    Zeigler, Ryan

    2015-01-01

    Planning for the curation of samples from future sample return missions must begin during the initial planning stages of a mission. Waiting until the samples have been returned to Earth, or even when you begin to physically build the spacecraft is too late. A lack of proper planning could lead to irreversible contamination of the samples, which in turn would compromise the scientific integrity of the mission. For example, even though the Apollo missions first returned samples in 1969, planning for the curation facility began in the early 1960s, and construction of the Lunar Receiving Laboratory was completed in 1967. In addition to designing the receiving facility and laboratory that the samples will be characterized and stored in, there are many aspects of contamination that must be addressed during the planning and building of the spacecraft: planetary protection (both outbound and inbound); cataloging, documenting, and preserving the materials used to build spacecraft (also known as coupons); near real-time monitoring of the environment in which the spacecraft is being built using witness plates for critical aspects of contamination (known as contamination control); and long term monitoring and preservation of the environment in which the spacecraft is being built for most aspects of potential contamination through the use of witness plates (known as contamination knowledge). The OSIRIS REx asteroid sample return mission, currently being built, is dealing with all of these aspects of contamination in order to ensure they return the best preserved sample possible. Coupons and witness plates from OSIRIS REx are currently being studied and stored (for future studies) at the Johnson Space Center. Similarly, planning for the clean room facility at Johnson Space Center to house the OSIRIS-REx samples is well advanced, and construction of the facility should begin in early 2017 (despite a nominal 2023 return date for OSIRIS-REx samples). Similar development is being

  14. Determinants of recurrence after intended curative resection for colorectal cancer

    DEFF Research Database (Denmark)

    Wilhelmsen, Michael; Kring, Thomas; Jorgensen, Lars N

    2014-01-01

    Despite intended curative resection, colorectal cancer will recur in ∼45% of the patients. Results of meta-analyses conclude that frequent follow-up does not lead to early detection of recurrence, but improves overall survival. The present literature shows that several factors play important roles...... in development of recurrence. It is well established that emergency surgery is a major determinant of recurrence. Moreover, anastomotic leakages, postoperative bacterial infections, and blood transfusions increase the recurrence rates although the exact mechanisms still remain obscure. From pathology studies...

  15. Lessons Learned

    Directory of Open Access Journals (Sweden)

    Amanda Phelan BNS, MSc, PhD

    2015-03-01

    Full Text Available The public health nurses’ scope of practice explicitly includes child protection within their role, which places them in a prime position to identify child protection concerns. This role compliments that of other professions and voluntary agenices who work with children. Public health nurses are in a privileged position as they form a relationship with the child’s parent(s/guardian(s and are able to see the child in its own environment, which many professionals cannot. Child protection in Ireland, while influenced by other countries, has progressed through a distinct pathway that streamlined protocols and procedures. However, despite the above serious failures have occurred in the Irish system, and inquiries over the past 20 years persistently present similar contributing factors, namely, the lack of standardized and comprehensive service responses. Moreover, poor practice is compounded by the lack of recognition of the various interactional processes taking place within and between the different agencies of child protection, leading to psychological barriers in communication. This article will explore the lessons learned for public health nurses practice in safeguarding children in the Republic of Ireland.

  16. The Knitting Lesson.

    Science.gov (United States)

    Smith, Pamela

    1987-01-01

    Based on Jean-Francois Millet's 1869 painting, "The Knitting Lesson," this lesson's goal is to introduce students in grades seven through nine to genre (everyday life) painting the nineteenth century. The lesson is also designed to show that some aspects of genre may be timeless. (BSR)

  17. Creating Large Scale Database Servers

    International Nuclear Information System (INIS)

    Becla, Jacek

    2001-01-01

    The BaBar experiment at the Stanford Linear Accelerator Center (SLAC) is designed to perform a high precision investigation of the decays of the B-meson produced from electron-positron interactions. The experiment, started in May 1999, will generate approximately 300TB/year of data for 10 years. All of the data will reside in Objectivity databases accessible via the Advanced Multi-threaded Server (AMS). To date, over 70TB of data have been placed in Objectivity/DB, making it one of the largest databases in the world. Providing access to such a large quantity of data through a database server is a daunting task. A full-scale testbed environment had to be developed to tune various software parameters and a fundamental change had to occur in the AMS architecture to allow it to scale past several hundred terabytes of data. Additionally, several protocol extensions had to be implemented to provide practical access to large quantities of data. This paper will describe the design of the database and the changes that we needed to make in the AMS for scalability reasons and how the lessons we learned would be applicable to virtually any kind of database server seeking to operate in the Petabyte region

  18. Creating Large Scale Database Servers

    Energy Technology Data Exchange (ETDEWEB)

    Becla, Jacek

    2001-12-14

    The BaBar experiment at the Stanford Linear Accelerator Center (SLAC) is designed to perform a high precision investigation of the decays of the B-meson produced from electron-positron interactions. The experiment, started in May 1999, will generate approximately 300TB/year of data for 10 years. All of the data will reside in Objectivity databases accessible via the Advanced Multi-threaded Server (AMS). To date, over 70TB of data have been placed in Objectivity/DB, making it one of the largest databases in the world. Providing access to such a large quantity of data through a database server is a daunting task. A full-scale testbed environment had to be developed to tune various software parameters and a fundamental change had to occur in the AMS architecture to allow it to scale past several hundred terabytes of data. Additionally, several protocol extensions had to be implemented to provide practical access to large quantities of data. This paper will describe the design of the database and the changes that we needed to make in the AMS for scalability reasons and how the lessons we learned would be applicable to virtually any kind of database server seeking to operate in the Petabyte region.

  19. Improved Functionality and Curation Support in the ADS

    Science.gov (United States)

    Accomazzi, Alberto; Kurtz, Michael J.; Henneken, Edwin A.; Grant, Carolyn S.; Thompson, Donna; Chyla, Roman; Holachek, Alexandra; Sudilovsky, Vladimir; Murray, Stephen S.

    2015-01-01

    In this poster we describe the developments of the new ADS platform over the past year, focusing on the functionality which improves its discovery and curation capabilities.The ADS Application Programming Interface (API) is being updated to support authenticated access to the entire suite of ADS services, in addition to the search functionality itself. This allows programmatic access to resources which are specific to a user or class of users.A new interface, built directly on top of the API, now provides a more intuitive search experience and takes into account the best practices in web usability and responsive design. The interface now incorporates in-line views of graphics from the AAS Astroexplorer and the ADS All-Sky Survey image collections.The ADS Private Libraries, first introduced over 10 years ago, are now being enhanced to allow the bookmarking, tagging and annotation of records of interest. In addition, libraries can be shared with one or more ADS users, providing an easy way to collaborate in the curation of lists of papers. A library can also be explicitly made public and shared at large via the publishing of its URL.In collaboration with the AAS, the ADS plans to support the adoption of ORCID identifiers by implementing a plugin which will simplify the import of papers in ORCID via a query to the ADS API. Deeper integration between the two systems will depend on available resources and feedback from the community.

  20. Astrolabe: Curating, Linking, and Computing Astronomy’s Dark Data

    Science.gov (United States)

    Heidorn, P. Bryan; Stahlman, Gretchen R.; Steffen, Julie

    2018-05-01

    Where appropriate repositories are not available to support all relevant astronomical data products, data can fall into darkness: unseen and unavailable for future reference and reuse. Some data in this category are legacy or old data, but newer data sets are also often uncurated and could remain dark. This paper provides a description of the design motivation and development of Astrolabe, a cyberinfrastructure project that addresses a set of community recommendations for locating and ensuring the long-term curation of dark or otherwise at-risk data and integrated computing. This paper also describes the outcomes of the series of community workshops that informed creation of Astrolabe. According to participants in these workshops, much astronomical dark data currently exist that are not curated elsewhere, as well as software that can only be executed by a few individuals and therefore becomes unusable because of changes in computing platforms. Astronomical research questions and challenges would be better addressed with integrated data and computational resources that fall outside the scope of existing observatory and space mission projects. As a solution, the design of the Astrolabe system is aimed at developing new resources for management of astronomical data. The project is based in CyVerse cyberinfrastructure technology and is a collaboration between the University of Arizona and the American Astronomical Society. Overall, the project aims to support open access to research data by leveraging existing cyberinfrastructure resources and promoting scientific discovery by making potentially useful data available to the astronomical community, in a computable format.

  1. Virtual Collections: An Earth Science Data Curation Service

    Science.gov (United States)

    Bugbee, Kaylin; Ramachandran, Rahul; Maskey, Manil; Gatlin, Patrick

    2016-01-01

    The role of Earth science data centers has traditionally been to maintain central archives that serve openly available Earth observation data. However, in order to ensure data are as useful as possible to a diverse user community, Earth science data centers must move beyond simply serving as an archive to offering innovative data services to user communities. A virtual collection, the end product of a curation activity that searches, selects, and synthesizes diffuse data and information resources around a specific topic or event, is a data curation service that improves the discoverability, accessibility, and usability of Earth science data and also supports the needs of unanticipated users. Virtual collections minimize the amount of the time and effort needed to begin research by maximizing certainty of reward and by providing a trustworthy source of data for unanticipated users. This presentation will define a virtual collection in the context of an Earth science data center and will highlight a virtual collection case study created at the Global Hydrology Resource Center data center.

  2. On Curating Multimodal Sensory Data for Health and Wellness Platforms

    Directory of Open Access Journals (Sweden)

    Muhammad Bilal Amin

    2016-06-01

    Full Text Available In recent years, the focus of healthcare and wellness technologies has shown a significant shift towards personal vital signs devices. The technology has evolved from smartphone-based wellness applications to fitness bands and smartwatches. The novelty of these devices is the accumulation of activity data as their users go about their daily life routine. However, these implementations are device specific and lack the ability to incorporate multimodal data sources. Data accumulated in their usage does not offer rich contextual information that is adequate for providing a holistic view of a user’s lifelog. As a result, making decisions and generating recommendations based on this data are single dimensional. In this paper, we present our Data Curation Framework (DCF which is device independent and accumulates a user’s sensory data from multimodal data sources in real time. DCF curates the context of this accumulated data over the user’s lifelog. DCF provides rule-based anomaly detection over this context-rich lifelog in real time. To provide computation and persistence over the large volume of sensory data, DCF utilizes the distributed and ubiquitous environment of the cloud platform. DCF has been evaluated for its performance, correctness, ability to detect complex anomalies, and management support for a large volume of sensory data.

  3. Curation Pedagogy for Media Studies: (Further Towards the Inexpert

    Directory of Open Access Journals (Sweden)

    Ben Andrews

    2012-12-01

    Full Text Available An educational ‘model’ for participatory learning and media literacy in the contemporary context of digitally mediated lifeworlds is emerging (Jenkins, 2010; Gauntlett, 2011; Fraser and Wardle, 2011. The critical problem, we argue, for its ‘adequacy’ is the privilege granted to curriculum content and skills over pedagogy. Elsewhere, we offered a ‘pedagogy of the inexpert’ as such a model for textconscious disciplines such as, but not restricted to, Media Studies. This strategy removes ‘the media’ from our gaze and looks awry to develop new ways of working with students – to ‘show and tell’ textual agency and more reflexive deconstruction of what it is to read and make media, and to ‘be’ in mediation. In this article, we develop this approach further towards a model of ‘curation’. Understood in this way, students ‘show’ media literacy in new spaces – not by recourse to skills, competences or analytical unmasking of the properties of a (contained text, but by exhibiting, by curating a moment in time of textual meaningmaking and meaning-taking, but with a mindfulness – a critical acceptance of such an attempt to hold and curate the flow of meaning as an artifice.

  4. On Curating Multimodal Sensory Data for Health and Wellness Platforms

    Science.gov (United States)

    Amin, Muhammad Bilal; Banos, Oresti; Khan, Wajahat Ali; Muhammad Bilal, Hafiz Syed; Gong, Jinhyuk; Bui, Dinh-Mao; Cho, Soung Ho; Hussain, Shujaat; Ali, Taqdir; Akhtar, Usman; Chung, Tae Choong; Lee, Sungyoung

    2016-01-01

    In recent years, the focus of healthcare and wellness technologies has shown a significant shift towards personal vital signs devices. The technology has evolved from smartphone-based wellness applications to fitness bands and smartwatches. The novelty of these devices is the accumulation of activity data as their users go about their daily life routine. However, these implementations are device specific and lack the ability to incorporate multimodal data sources. Data accumulated in their usage does not offer rich contextual information that is adequate for providing a holistic view of a user’s lifelog. As a result, making decisions and generating recommendations based on this data are single dimensional. In this paper, we present our Data Curation Framework (DCF) which is device independent and accumulates a user’s sensory data from multimodal data sources in real time. DCF curates the context of this accumulated data over the user’s lifelog. DCF provides rule-based anomaly detection over this context-rich lifelog in real time. To provide computation and persistence over the large volume of sensory data, DCF utilizes the distributed and ubiquitous environment of the cloud platform. DCF has been evaluated for its performance, correctness, ability to detect complex anomalies, and management support for a large volume of sensory data. PMID:27355955

  5. Strategies for the Curation of CAD Engineering Models

    Directory of Open Access Journals (Sweden)

    Manjula Patel

    2009-06-01

    Full Text Available Normal 0 Product Lifecycle Management (PLM has become increasingly important in the engineering community over the last decade or so, due to the globalisation of markets and the rising popularity of products provided as services. It demands the efficient capture, representation, organisation, retrieval and reuse of product data over its entire life. Simultaneously, there is now a much greater reliance on CAD models for communicating designs to manufacturers, builders, maintenance crews and regulators, and for definitively expressing designs. Creating the engineering record digitally, however, presents problems not only for its long-term maintenance and accessibility - due in part to the rapid obsolescence of the hardware, software and file formats involved - but also for recording the evolution of designs, artefacts and products. We examine the curation and preservation requirements in PLM and suggest ways of alleviating the problems of sustaining CAD engineering models through the use of lightweight formats, layered annotation and the collection of Representation Information as defined in the Open Archival Information System (OAIS Reference Model.  We describe two tools which have been specifically developed to aid in the curation of CAD engineering models in the context of PLM: Lightweight Models with Multilayered Annotation (LiMMA and a Registry/Repository of Representation Information for Engineering (RRoRIfE.

  6. Virtual Collections: An Earth Science Data Curation Service

    Science.gov (United States)

    Bugbee, K.; Ramachandran, R.; Maskey, M.; Gatlin, P. N.

    2016-12-01

    The role of Earth science data centers has traditionally been to maintain central archives that serve openly available Earth observation data. However, in order to ensure data are as useful as possible to a diverse user community, Earth science data centers must move beyond simply serving as an archive to offering innovative data services to user communities. A virtual collection, the end product of a curation activity that searches, selects, and synthesizes diffuse data and information resources around a specific topic or event, is a data curation service that improves the discoverability, accessibility and usability of Earth science data and also supports the needs of unanticipated users. Virtual collections minimize the amount of time and effort needed to begin research by maximizing certainty of reward and by providing a trustworthy source of data for unanticipated users. This presentation will define a virtual collection in the context of an Earth science data center and will highlight a virtual collection case study created at the Global Hydrology Resource Center data center.

  7. REDfly: a Regulatory Element Database for Drosophila.

    Science.gov (United States)

    Gallo, Steven M; Li, Long; Hu, Zihua; Halfon, Marc S

    2006-02-01

    Bioinformatics studies of transcriptional regulation in the metazoa are significantly hindered by the absence of readily available data on large numbers of transcriptional cis-regulatory modules (CRMs). Even the richly annotated Drosophila melanogaster genome lacks extensive CRM information. We therefore present here a database of Drosophila CRMs curated from the literature complete with both DNA sequence and a searchable description of the gene expression pattern regulated by each CRM. This resource should greatly facilitate the development of computational approaches to CRM discovery as well as bioinformatics analyses of regulatory sequence properties and evolution.

  8. Advanced Curation Activities at NASA: Implications for Astrobiological Studies of Future Sample Collections

    Science.gov (United States)

    McCubbin, F. M.; Evans, C. A.; Fries, M. D.; Harrington, A. D.; Regberg, A. B.; Snead, C. J.; Zeigler, R. A.

    2017-01-01

    The Astromaterials Acquisition and Curation Office (henceforth referred to herein as NASA Curation Office) at NASA Johnson Space Center (JSC) is responsible for curating all of NASA's extraterrestrial samples. Under the governing document, NASA Policy Directive (NPD) 7100.10F JSC is charged with curation of all extraterrestrial material under NASA control, including future NASA missions. The Directive goes on to define Curation as including documentation, preservation, preparation, and distribution of samples for re-search, education, and public outreach. Here we briefly describe NASA's astromaterials collections and our ongoing efforts related to enhancing the utility of our current collections as well as our efforts to prepare for future sample return missions. We collectively refer to these efforts as advanced curation.

  9. A systematic review of randomized controlled trials on curative and health enhancement effects of forest therapy

    Directory of Open Access Journals (Sweden)

    Kamioka H

    2012-07-01

    Full Text Available Hiroharu Kamioka,1 Kiichiro Tsutani,2 Yoshiteru Mutoh,3 Takuya Honda,4 Nobuyoshi Shiozawa,5 Shinpei Okada,6 Sang-Jun Park,6 Jun Kitayuguchi,7 Masamitsu Kamada,8 Hiroyasu Okuizumi,9 Shuichi Handa91Faculty of Regional Environment Science, Tokyo University of Agriculture, Tokyo, 2Department of Drug Policy and Management, Graduate School of Pharmaceutical Sciences, The University of Tokyo, Tokyo, 3Todai Policy Alternatives Research Institute, The University of Tokyo, Tokyo, 4Japanese Society for the Promotion of Science, Tokyo, 5Food Labeling Division, Consumer Affairs Agency, Cabinet Office, Government of Japan, Tokyo, 6Physical Education and Medicine Research Foundation, Nagano, 7Physical Education and Medicine Research Center Unnan, Shimane, 8Department of Environmental and Preventive Medicine, Shimane University School of Medicine, Shimane, 9Mimaki Onsen (Spa Clinic, Tomi City, Nagano, JapanObjective: To summarize the evidence for curative and health enhancement effects through forest therapy and to assess the quality of studies based on a review of randomized controlled trials (RCTs.Study design: A systematic review based on RCTs.Methods: Studies were eligible if they were RCTs. Studies included one treatment group in which forest therapy was applied. The following databases – from 1990 to November 9, 2010 – were searched: MEDLINE via PubMed, CINAHL, Web of Science, and Ichushi-Web. All Cochrane databases and Campbell Systematic Reviews were also searched up to November 9, 2010.Results: Two trials met all inclusion criteria. No specific diseases were evaluated, and both studies reported significant effectiveness in one or more outcomes for health enhancement. However, the results of evaluations with the CONSORT (Consolidated Standards of Reporting Trials 2010 and CLEAR NPT (A Checklist to Evaluate a Report of a Nonpharmacological Trial checklists generally showed a remarkable lack of description in the studies. Furthermore, there was a

  10. Stackfile Database

    Science.gov (United States)

    deVarvalho, Robert; Desai, Shailen D.; Haines, Bruce J.; Kruizinga, Gerhard L.; Gilmer, Christopher

    2013-01-01

    This software provides storage retrieval and analysis functionality for managing satellite altimetry data. It improves the efficiency and analysis capabilities of existing database software with improved flexibility and documentation. It offers flexibility in the type of data that can be stored. There is efficient retrieval either across the spatial domain or the time domain. Built-in analysis tools are provided for frequently performed altimetry tasks. This software package is used for storing and manipulating satellite measurement data. It was developed with a focus on handling the requirements of repeat-track altimetry missions such as Topex and Jason. It was, however, designed to work with a wide variety of satellite measurement data [e.g., Gravity Recovery And Climate Experiment -- GRACE). The software consists of several command-line tools for importing, retrieving, and analyzing satellite measurement data.

  11. Science Support: The Building Blocks of Active Data Curation

    Science.gov (United States)

    Guillory, A.

    2013-12-01

    While the scientific method is built on reproducibility and transparency, and results are published in peer reviewed literature, we have come to the digital age of very large datasets (now of the order of petabytes and soon exabytes) which cannot be published in the traditional way. To preserve reproducibility and transparency, active curation is necessary to keep and protect the information in the long term, and 'science support' activities provide the building blocks for active data curation. With the explosive growth of data in all fields in recent years, there is a pressing urge for data centres to now provide adequate services to ensure long-term preservation and digital curation of project data outputs, however complex those may be. Science support provides advice and support to science projects on data and information management, from file formats through to general data management awareness. Another purpose of science support is to raise awareness in the science community of data and metadata standards and best practice, engendering a culture where data outputs are seen as valued assets. At the heart of Science support is the Data Management Plan (DMP) which sets out a coherent approach to data issues pertaining to the data generating project. It provides an agreed record of the data management needs and issues within the project. The DMP is agreed upon with project investigators to ensure that a high quality documented data archive is created. It includes conditions of use and deposit to clearly express the ownership, responsibilities and rights associated with the data. Project specific needs are also identified for data processing, visualization tools and data sharing services. As part of the National Centre for Atmospheric Science (NCAS) and National Centre for Earth Observation (NCEO), the Centre for Environmental Data Archival (CEDA) fulfills this science support role of facilitating atmospheric and Earth observation data generating projects to ensure

  12. The curative management of synchronous rectal and prostate cancer

    Science.gov (United States)

    Kavanagh, Dara O; Martin, Joseph; Small, Cormac; Joyce, Myles R; Faul, Clare M; Kelly, Paul J; O'Riordain, Michael; Gillham, Charles M; Armstrong, John G; Salib, Osama; McNamara, Deborah A; McVey, Gerard; O'Neill, Brian D P

    2016-01-01

    Objective: Neoadjuvant “long-course” chemoradiation is considered a standard of care in locally advanced rectal cancer. In addition to prostatectomy, external beam radiotherapy and brachytherapy with or without androgen suppression (AS) are well established in prostate cancer management. A retrospective review of ten cases was completed to explore the feasibility and safety of applying these standards in patients with dual pathology. To our knowledge, this is the largest case series of synchronous rectal and prostate cancers treated with curative intent. Methods: Eligible patients had synchronous histologically proven locally advanced rectal cancer (defined as cT3-4Nx; cTxN1-2) and non-metastatic prostate cancer (pelvic nodal disease permissible). Curative treatment was delivered to both sites simultaneously. Follow-up was as per institutional guidelines. Acute and late toxicities were reviewed, and a literature search performed. Results: Pelvic external beam radiotherapy (RT) 45–50.4 Gy was delivered concurrent with 5-fluorouracil (5FU). Prostate total dose ranged from 70.0 to 79.2 Gy. No acute toxicities occurred, excluding AS-induced erectile dysfunction. Nine patients proceeded to surgery, and one was managed expectantly. Three relapsed with metastatic colorectal cancer, two with metastatic prostate cancer. Five patients have no evidence of recurrence, and four remain alive with metastatic disease. With a median follow-up of 2.2 years (range 1.2–6.3 years), two significant late toxicities occurred; G3 proctitis in a patient receiving palliative bevacizumab and a G3 anastomotic stricture precluding stoma reversal. Conclusion: Patients proceeding to synchronous radical treatment of both primary sites should receive 45–50.4 Gy pelvic RT with infusional 5FU. Prostate dose escalation should be given with due consideration to the potential impact of prostate cancer on patient survival, as increasing dose may result in significant late morbidity

  13. Curative Radiation Therapy for T2N0M0

    International Nuclear Information System (INIS)

    Park, In Kyu; Kim, Jae Choel

    1995-01-01

    Purpose : Surgery is the treatment of choice for resectable non-small cell lung cancer. For patients who are medically unable to tolerate a surgical resection or who refuse surgery, radiation therapy is an acceptable alternative. A retrospective analysis of patients with stage I non-small cell lung cancer treated with curative radiation therapy was performed to determine the results of curative radiation therapy and patterns of failure, and to identify factors that may influence survival. Materials and Methods : From 1986 through 1993, 39 patients with T2N0M0 non-small cell lung cancer were treated with curative radiation therapy at department of radiation oncology, Kyungpook national university hospital All patients were not candidates for surgical resection because of either patient refusal (16 patients), poor pulmonary function (12 patients), old age (7 patients), poor performance (2 patients) or coexisting medical disease (2 patients). Median age of patients was 67 years. Histologic cell type was squamous cell carcinoma in 1. All patients were treated with megavoltage irradiation and radiation dose raged from 5000cGy to 6150 cGy with a median dose of 600cGy. The median follow-up was 17 months with a range of 4 to 82 months. Survival was measured from the date therapy initiated. Results : The overall survival rate for entire patients was 40.6% at 2 years and 27.7% at 3 years, with a median survival time of 21 months he disease-free survival at 2 and 3 years was 51.7% and 25.8%, respectively. Of evaluable 20 Patients with complete response, 15 Patients were considered to have failed. Of these, 13 patients showed local failure and 2 patients failed distantly. Response to treatment (p=0.0001), tumor size (p=0.0019) and age p=0.0247) were favorably associated with overall survival. Only age was predictive for disease-free survival (p=0.0452). Conclusion : Radiation therapy is an effective treatment for small (less than 3 cm) tumors, and should be offered as an

  14. Curating Performance on the Edge of the Art Museum

    DEFF Research Database (Denmark)

    Groth, Sanne Krogh; Schwarzbart, Judith

    two-day festival offers a format that vary considerable from the exhibition series the museum puts on most of the time. The performance program includes artists such as composers usually working with contemporary music, electronic music composers, as well as performance artists working from......Since the Intermedia and Fluxus movements a variety of timebased artforms have been contained within visual art contexts. The performative works draw often as much on the tradition of theatre, music, dance, and poetry reading as fine art. Although the institutional context plays a role...... art institution. Our research relates specifically to a festival for performative art, ACTS 2014, which we co-curate for the Museum of Contemporary Art in Roskilde. Having grown out of a Fluxus spirit, the museum is not foreign to time-based practices like many museums are. Nevertheless, the intensive...

  15. Orthology prediction methods: a quality assessment using curated protein families.

    Science.gov (United States)

    Trachana, Kalliopi; Larsson, Tomas A; Powell, Sean; Chen, Wei-Hua; Doerks, Tobias; Muller, Jean; Bork, Peer

    2011-10-01

    The increasing number of sequenced genomes has prompted the development of several automated orthology prediction methods. Tests to evaluate the accuracy of predictions and to explore biases caused by biological and technical factors are therefore required. We used 70 manually curated families to analyze the performance of five public methods in Metazoa. We analyzed the strengths and weaknesses of the methods and quantified the impact of biological and technical challenges. From the latter part of the analysis, genome annotation emerged as the largest single influencer, affecting up to 30% of the performance. Generally, most methods did well in assigning orthologous group but they failed to assign the exact number of genes for half of the groups. The publicly available benchmark set (http://eggnog.embl.de/orthobench/) should facilitate the improvement of current orthology assignment protocols, which is of utmost importance for many fields of biology and should be tackled by a broad scientific community. Copyright © 2011 WILEY Periodicals, Inc.

  16. The role of endobroncial irradiation as a curative therapy

    International Nuclear Information System (INIS)

    Fuwa, Nobukazu; Ito, Yoshiyuki; Kato, Eriko; Kamata, Minoru; Morita, Kozo; Kikuchi, Yuzo.

    1997-01-01

    Endobronchial irradiation for lung cancer has primarily been used in cases of local progression or recurrence. Although its use for palliation of symptoms has been well evaluated, its role in treatment for cure is still unknown. We would like to report on the role of endobronchial irradiation as a curative therapy based on our clinical experience (long time survivors). Forty-one patients treated with endobronchial irradiation using low dose rate 192Iridium between February 1987 and December 1993 were made available for study. Of these, 17 were chest X-P negative cancer, 13 were post operative recurrent cancer, 7 were advanced cancer and 4 were tracheal cancer, respectively. The dose of endobronchial irradiation using an applicator with spacer was 5 to 7 Gy per session, administered either once or twice a week. External irradiation was administered except one case. Local recurrence was observed in two cases of chest X-P negative cancer, three cases of post operative cancer and five cases of advanced cancer. More than three years survivors were observed in 6 cases of chest X-P negative cancer, 5 cases of post operative cancer and one case of tracheal cancer. Complications due to endobronchial irradiation were seen in 2 cases, one case was pulmonary hemorrage and the other was shallow ulceration of the bronchus. It was shown that chest X-P negative lung cancer and part of post operative recurrent cancer could be cured by endobronchial irradiation. This technique is considered to be useful for not only palliative therapy but curative therapy as well. (author)

  17. Characteristics of knowledge content in a curated online evidence library.

    Science.gov (United States)

    Varada, Sowmya; Lacson, Ronilda; Raja, Ali S; Ip, Ivan K; Schneider, Louise; Osterbur, David; Bain, Paul; Vetrano, Nicole; Cellini, Jacqueline; Mita, Carol; Coletti, Margaret; Whelan, Julia; Khorasani, Ramin

    2018-05-01

    To describe types of recommendations represented in a curated online evidence library, report on the quality of evidence-based recommendations pertaining to diagnostic imaging exams, and assess underlying knowledge representation. The evidence library is populated with clinical decision rules, professional society guidelines, and locally developed best practice guidelines. Individual recommendations were graded based on a standard methodology and compared using chi-square test. Strength of evidence ranged from grade 1 (systematic review) through grade 5 (recommendations based on expert opinion). Finally, variations in the underlying representation of these recommendations were identified. The library contains 546 individual imaging-related recommendations. Only 15% (16/106) of recommendations from clinical decision rules were grade 5 vs 83% (526/636) from professional society practice guidelines and local best practice guidelines that cited grade 5 studies (P < .0001). Minor head trauma, pulmonary embolism, and appendicitis were topic areas supported by the highest quality of evidence. Three main variations in underlying representations of recommendations were "single-decision," "branching," and "score-based." Most recommendations were grade 5, largely because studies to test and validate many recommendations were absent. Recommendation types vary in amount and complexity and, accordingly, the structure and syntax of statements they generate. However, they can be represented in single-decision, branching, and score-based representations. In a curated evidence library with graded imaging-based recommendations, evidence quality varied widely, with decision rules providing the highest-quality recommendations. The library may be helpful in highlighting evidence gaps, comparing recommendations from varied sources on similar clinical topics, and prioritizing imaging recommendations to inform clinical decision support implementation.

  18. Health promotion, preventive and curative aspects of diseases in astrology.

    Science.gov (United States)

    Sharma, Bhuvnesh Kumar; Subhakta, P K J P; Narayana, A

    2007-01-01

    The whole universe is intermingling into a unit in the period of globalization. Different cultures, life-styles and sciences are co-operating with each other in this situation. World Health Organization is working towards collaborating all prevalent medical sciences for attainment of good health and family welfare for each and every individual by 2020. Astrology is a part of Indian heritage. Astrology means the art of predicting or determining the influence of the planets and stars on human affairs. The origin of this word is from Greek word astron, star + logos (discourse). The account of deeds of good and bad during the present life and previous lives, their consequences of health or ill health during this life i.e. what, when and how the things takes place will be clearly known through Astrology. Highly advanced knowledge related to Astrology on medicine is preserved in Indian scriptures and the knowledge was transmitted from generation to generation. It is also a good source for health promotion, preventive, curative and other medical aspects. Brief direction related to astrological medical aspects is also available in Ayurvedic literature (Carakasamhită, Suśrutasamhhită, Aşţăngasangraha, Aşţăngahŗdaya, Sărngadharasamhită , Băvaprakăśa etc.) Some Ayurvedic practitioners, scholars and scientists realize the need of astrological knowledge related to medicine in the present time. In ancient times physician, astrologer and purŏhita (Hindu priest) simultaneously looked after the health and family welfare of individual, families and country. Astrologer guides medication and suitable time for the better cure of ailments. Even the medicinal herbs were collected and treated at appropriate time for their efficacy. Astrology and Ayurvĕda are inseparable sciences of life. Hence, in this article, a concise astrological evaluation related to health promotion, preventive and curative aspects of Astrology is being presented.

  19. Patterns of failure following curative resection of gastric carcinoma

    Energy Technology Data Exchange (ETDEWEB)

    Landry, J.; Tepper, J.E.; Wood, W.C.; Moulton, E.O.; Koerner, F.; Sullinger, J. (Massachusetts General Hospital Cancer Center, Boston (USA))

    1990-12-01

    To identify patterns of failure following curative resection of gastric carcinoma, the records of 130 patients undergoing resection with curative intent at the Massachusetts General Hospital were reviewed. The total local-regional failure rate was 38% (49/130 patients), with 21 patients having local-regional failure alone and 28 patients having local-regional failure and distant metastases. The incidence of local failure rose with the more advanced stages of disease. Tumors staged B2, B3, C2, and C3 had local-regional failure rates in excess of 35%. This group of patients might benefit from adjuvant radiation therapy to the tumor bed and regional lymphatics. Local-regional failure rate was highest in the anastomosis or stump 33/130 (25%), followed by the stomach bed 27/130 (21%). The overall incidence of distant metastases was 52% (67/130 patients) and rose in the more advanced disease stages. Tumors staged B2, B3, C2, and C3 had rates of distant metastases greater than 50%. Sixty-one patients (77%) had failure in the abdomen (liver, peritoneal surface, adrenal, kidney, and spleen, but excluding tumor bed, anastomosis, or regional nodes). Patients with Stage B2, B3, C2, and C3 tumors had total abdominal failure rates greater than 40%. The highest failure rates in the liver were in Stages B3 and C3, in which the subsequent development of liver metastasis was 40% and 47%, respectively. Peritoneal seeding occurred in 30/130 (23%) of patients and was highest in Stages C2 and C3, with rates of 27% and 41%, respectively.

  20. Patterns of failure following curative resection of gastric carcinoma

    International Nuclear Information System (INIS)

    Landry, J.; Tepper, J.E.; Wood, W.C.; Moulton, E.O.; Koerner, F.; Sullinger, J.

    1990-01-01

    To identify patterns of failure following curative resection of gastric carcinoma, the records of 130 patients undergoing resection with curative intent at the Massachusetts General Hospital were reviewed. The total local-regional failure rate was 38% (49/130 patients), with 21 patients having local-regional failure alone and 28 patients having local-regional failure and distant metastases. The incidence of local failure rose with the more advanced stages of disease. Tumors staged B2, B3, C2, and C3 had local-regional failure rates in excess of 35%. This group of patients might benefit from adjuvant radiation therapy to the tumor bed and regional lymphatics. Local-regional failure rate was highest in the anastomosis or stump 33/130 (25%), followed by the stomach bed 27/130 (21%). The overall incidence of distant metastases was 52% (67/130 patients) and rose in the more advanced disease stages. Tumors staged B2, B3, C2, and C3 had rates of distant metastases greater than 50%. Sixty-one patients (77%) had failure in the abdomen (liver, peritoneal surface, adrenal, kidney, and spleen, but excluding tumor bed, anastomosis, or regional nodes). Patients with Stage B2, B3, C2, and C3 tumors had total abdominal failure rates greater than 40%. The highest failure rates in the liver were in Stages B3 and C3, in which the subsequent development of liver metastasis was 40% and 47%, respectively. Peritoneal seeding occurred in 30/130 (23%) of patients and was highest in Stages C2 and C3, with rates of 27% and 41%, respectively

  1. Using the Pathogen-Host Interactions database (PHI-base to investigate plant pathogen genomes and genes implicated in virulence

    Directory of Open Access Journals (Sweden)

    Martin eUrban

    2015-08-01

    Full Text Available New pathogen-host interaction mechanisms can be revealed by integrating mutant phenotype data with genetic information. PHI-base is a multi-species manually curated database combining peer-reviewed published phenotype data from plant and animal pathogens and gene/protein information in a single database.

  2. Extending Database Integration Technology

    National Research Council Canada - National Science Library

    Buneman, Peter

    1999-01-01

    Formal approaches to the semantics of databases and database languages can have immediate and practical consequences in extending database integration technologies to include a vastly greater range...

  3. An Intelligent Web Digital Image Metadata Service Platform for Social Curation Commerce Environment

    Directory of Open Access Journals (Sweden)

    Seong-Yong Hong

    2015-01-01

    Full Text Available Information management includes multimedia data management, knowledge management, collaboration, and agents, all of which are supporting technologies for XML. XML technologies have an impact on multimedia databases as well as collaborative technologies and knowledge management. That is, e-commerce documents are encoded in XML and are gaining much popularity for business-to-business or business-to-consumer transactions. Recently, the internet sites, such as e-commerce sites and shopping mall sites, deal with a lot of image and multimedia information. This paper proposes an intelligent web digital image information retrieval platform, which adopts XML technology for social curation commerce environment. To support object-based content retrieval on product catalog images containing multiple objects, we describe multilevel metadata structures representing the local features, global features, and semantics of image data. To enable semantic-based and content-based retrieval on such image data, we design an XML-Schema for the proposed metadata. We also describe how to automatically transform the retrieval results into the forms suitable for the various user environments, such as web browser or mobile device, using XSLT. The proposed scheme can be utilized to enable efficient e-catalog metadata sharing between systems, and it will contribute to the improvement of the retrieval correctness and the user’s satisfaction on semantic-based web digital image information retrieval.

  4. AtomPy: An Open Atomic Data Curation Environment for Astrophysical Applications

    Directory of Open Access Journals (Sweden)

    Claudio Mendoza

    2014-05-01

    Full Text Available We present a cloud-computing environment, referred to as AtomPy, based on Google-Drive Sheets and Pandas (Python Data Analysis Library DataFrames to promote community-driven curation of atomic data for astrophysical applications, a stage beyond database development. The atomic model for each ionic species is contained in a multi-sheet workbook, tabulating representative sets of energy levels, A-values and electron impact effective collision strengths from different sources. The relevant issues that AtomPy intends to address are: (i data quality by allowing open access to both data producers and users; (ii comparisons of different datasets to facilitate accuracy assessments; (iii downloading to local data structures (i.e., Pandas DataFrames for further manipulation and analysis by prospective users; and (iv data preservation by avoiding the discard of outdated sets. Data processing workflows are implemented by means of IPython Notebooks, and collaborative software developments are encouraged and managed within the GitHub social network. The facilities of AtomPy are illustrated with the critical assessment of the transition probabilities for ions in the hydrogen and helium isoelectronic sequences with atomic number Z ≤ 10.

  5. A High School Student's Bill of Rights. Teaching Resources in the ERIC Database (TRIED) Series.

    Science.gov (United States)

    Gottlieb, Stephen S.

    Designed to tap the rich collection of instructional techniques in the ERIC database, this compilation of lesson plans focuses on teaching high school students their Constitutional rights and responsibilities. The 40 lesson plans in the book cover the courts and basic rights, the rights of criminal suspects, the rights of minors and education law,…

  6. Design database for quantitative trait loci (QTL) data warehouse, data mining, and meta-analysis.

    Science.gov (United States)

    Hu, Zhi-Liang; Reecy, James M; Wu, Xiao-Lin

    2012-01-01

    A database can be used to warehouse quantitative trait loci (QTL) data from multiple sources for comparison, genomic data mining, and meta-analysis. A robust database design involves sound data structure logistics, meaningful data transformations, normalization, and proper user interface designs. This chapter starts with a brief review of relational database basics and concentrates on issues associated with curation of QTL data into a relational database, with emphasis on the principles of data normalization and structure optimization. In addition, some simple examples of QTL data mining and meta-analysis are included. These examples are provided to help readers better understand the potential and importance of sound database design.

  7. A Phase Blending Study on Rubber Blends Based on the Solubility Preference of Curatives

    NARCIS (Netherlands)

    Guo, R.; Talma, Auke; Datta, Rabin; Dierkes, Wilma K.; Noordermeer, Jacobus W.M.

    2009-01-01

    Using previously obtained data on the solubilities of curatives in SBR, EPDM and in NBR, different mixing procedures were performed on 50/50 SBR/EPDM and NBR/EPDM blends. In contrast to a previous phase-mixing study, the curatives were added to separate phases before final blending, in an attempt to

  8. Digital Curation as a Core Competency in Current Learning and Literacy: A Higher Education Perspective

    Science.gov (United States)

    Ungerer, Leona M.

    2016-01-01

    Digital curation may be regarded as a core competency in higher education since it contributes to establishing a sense of metaliteracy (an essential requirement for optimally functioning in a modern media environment) among students. Digital curation is gradually finding its way into higher education curricula aimed at fostering social media…

  9. Reducing data friction through site-based data curation

    Science.gov (United States)

    Thomer, A.; Palmer, C. L.

    2017-12-01

    Much of geoscience research takes place at "scientifically significant sites": localities which have attracted a critical mass of scientific interest, and thereby merit protection by government bodies, as well as the preservation of specimen and data collections and the development of site-specific permitting requirements for access to the site and its associated collections. However, many data standards and knowledge organization schemas do not adequately describe key characteristics of the sites, despite their centrality to research projects. Through work conducted as part of the IMLS-funded Site-Based Data Curation (SBDC) project, we developed a Minimum Information Framework (MIF) for site-based science, in which "information about a site's structure" is considered a core class of information. Here we present our empirically-derived information framework, as well as the methods used to create it. We believe these approaches will lead to the development of more effective data repositories and tools, and thereby will reduce "data friction" in interdisciplinary, yet site-based, geoscience workflows. The Minimum Information Framework for Site-based Research was developed through work at two scientifically significant sites: the hot springs at Yellowstone National Park, which are key to geobiology research; and the La Brea Tar Pits, an important paleontology locality in Southern California. We employed diverse methods of participatory engagement, in which key stakeholders at our sites (e.g. curators, collections managers, researchers, permit officers) were consulted through workshops, focus groups, interviews, action research methods, and collaborative information modeling and systems analysis. These participatory approaches were highly effective in fostering on-going partnership among a diverse team of domain scientists, information scientists, and software developers. The MIF developed in this work may be viewed as a "proto-standard" that can inform future

  10. Genesis Contingency Planning and Mishap Recovery: The Sample Curation View

    Science.gov (United States)

    Stansbery, E. K.; Allton, J. H.; Allen, C. C.; McNamara, K. M.; Calaway, M.; Rodriques, M. C.

    2007-01-01

    Planning for sample preservation and curation was part of mission design from the beginning. One of the scientific objectives for Genesis included collecting samples of three regimes of the solar wind in addition to collecting bulk solar wind during the mission. Collectors were fabricated in different thicknesses for each regime of the solar wind and attached to separate frames exposed to the solar wind during specific periods of solar activity associated with each regime. The original plan to determine the solar regime sampled for specific collectors was to identify to which frame the collector was attached. However, the collectors were dislodged during the hard landing making identification by frame attachment impossible. Because regimes were also identified by thickness of the collector, the regime sampled is identified by measuring fragment thickness. A variety of collector materials and thin films applied to substrates were selected and qualified for flight. This diversity provided elemental measurement in more than one material, mitigating effects of diffusion rates and/or radiation damage. It also mitigated against different material and substrate strengths resulting in differing effects of the hard landing. For example, silicon crystal substrates broke into smaller fragments than sapphire-based substrates and diamond surfaces were more resilient to flying debris damage than gold. The primary responsibility of the curation team for recovery was process documentation. Contingency planning for the recovery phase expanded this responsibility to include not only equipment to document, but also gather, contain and identify samples from the landing area and the recovered spacecraft. The team developed contingency plans for various scenarios as part of mission planning that included topographic maps to aid in site recovery and identification of different modes of transport and purge capability depending on damage. A clean tent, set-up at Utah Test & Training Range

  11. HCVpro: Hepatitis C virus protein interaction database

    KAUST Repository

    Kwofie, Samuel K.

    2011-12-01

    It is essential to catalog characterized hepatitis C virus (HCV) protein-protein interaction (PPI) data and the associated plethora of vital functional information to augment the search for therapies, vaccines and diagnostic biomarkers. In furtherance of these goals, we have developed the hepatitis C virus protein interaction database (HCVpro) by integrating manually verified hepatitis C virus-virus and virus-human protein interactions curated from literature and databases. HCVpro is a comprehensive and integrated HCV-specific knowledgebase housing consolidated information on PPIs, functional genomics and molecular data obtained from a variety of virus databases (VirHostNet, VirusMint, HCVdb and euHCVdb), and from BIND and other relevant biology repositories. HCVpro is further populated with information on hepatocellular carcinoma (HCC) related genes that are mapped onto their encoded cellular proteins. Incorporated proteins have been mapped onto Gene Ontologies, canonical pathways, Online Mendelian Inheritance in Man (OMIM) and extensively cross-referenced to other essential annotations. The database is enriched with exhaustive reviews on structure and functions of HCV proteins, current state of drug and vaccine development and links to recommended journal articles. Users can query the database using specific protein identifiers (IDs), chromosomal locations of a gene, interaction detection methods, indexed PubMed sources as well as HCVpro, BIND and VirusMint IDs. The use of HCVpro is free and the resource can be accessed via http://apps.sanbi.ac.za/hcvpro/ or http://cbrc.kaust.edu.sa/hcvpro/. © 2011 Elsevier B.V.

  12. Neural systems language: a formal modeling language for the systematic description, unambiguous communication, and automated digital curation of neural connectivity.

    Science.gov (United States)

    Brown, Ramsay A; Swanson, Larry W

    2013-09-01

    Systematic description and the unambiguous communication of findings and models remain among the unresolved fundamental challenges in systems neuroscience. No common descriptive frameworks exist to describe systematically the connective architecture of the nervous system, even at the grossest level of observation. Furthermore, the accelerating volume of novel data generated on neural connectivity outpaces the rate at which this data is curated into neuroinformatics databases to synthesize digitally systems-level insights from disjointed reports and observations. To help address these challenges, we propose the Neural Systems Language (NSyL). NSyL is a modeling language to be used by investigators to encode and communicate systematically reports of neural connectivity from neuroanatomy and brain imaging. NSyL engenders systematic description and communication of connectivity irrespective of the animal taxon described, experimental or observational technique implemented, or nomenclature referenced. As a language, NSyL is internally consistent, concise, and comprehensible to both humans and computers. NSyL is a promising development for systematizing the representation of neural architecture, effectively managing the increasing volume of data on neural connectivity and streamlining systems neuroscience research. Here we present similar precedent systems, how NSyL extends existing frameworks, and the reasoning behind NSyL's development. We explore NSyL's potential for balancing robustness and consistency in representation by encoding previously reported assertions of connectivity from the literature as examples. Finally, we propose and discuss the implications of a framework for how NSyL will be digitally implemented in the future to streamline curation of experimental results and bridge the gaps among anatomists, imagers, and neuroinformatics databases. Copyright © 2013 Wiley Periodicals, Inc.

  13. Lesson study i Danmark?

    DEFF Research Database (Denmark)

    Mogensen, Arne

    2009-01-01

    Der beskrives et japansk lesson study forløb, og det diskuteres i hvilket omfang, de gode japanske erfaringer kan overføres til dansk matematikundervisning.......Der beskrives et japansk lesson study forløb, og det diskuteres i hvilket omfang, de gode japanske erfaringer kan overføres til dansk matematikundervisning....

  14. "Frankenstein." [Lesson Plan].

    Science.gov (United States)

    Simon, Melanie

    Based on Mary Shelley's novel "Frankenstein," this lesson plan presents activities designed to help students understand that active readers interpret a novel (its characters, plot, setting, and theme) in different ways; and the great literature can be and has been adapted in many ways over time. The main activity of the lesson involves students…

  15. Agile Data Curation at a State Geological Survey

    Science.gov (United States)

    Hills, D. J.

    2015-12-01

    State agencies, including geological surveys, are often the gatekeepers for myriad data products essential for scientific research and economic development. For example, the Geological Survey of Alabama (GSA) is mandated to explore for, characterize, and report Alabama's mineral, energy, water, and biological resources in support of economic development, conservation, management, and public policy for the betterment of Alabama's citizens, communities, and businesses. As part of that mandate, the GSA has increasingly been called upon to make our data more accessible to stakeholders. Even as demand for greater data accessibility grows, budgets for such efforts are often small, meaning that agencies must do more for less. Agile software development has yielded efficient, effective products, most often at lower cost and in shorter time. Taking guidance from the agile software development model, the GSA is working towards more agile data management and curation. To date, the GSA's work has been focused primarily on data rescue. By using workflows that maximize clear communication while encouraging simplicity (e.g., maximizing the amount of work not done or that can be automated), the GSA is bringing decades of dark data into the light. Regular checks by the data rescuer with the data provider (or their proxy) provides quality control without adding an overt burden on either party. Moving forward, these workflows will also allow for more efficient and effective data management.

  16. The curator/patron: Foundations and contemporary art

    Directory of Open Access Journals (Sweden)

    Rebecca Coates

    2008-12-01

    Full Text Available This article addresses the role of private foundations in commissioning site-specific ephemeral art works: contemporary art projects of a temporary nature that are realised outside of public institutions. Though small in number, I argue that the private individuals creating and managing private foundations of this nature demonstrate a new form of patronage, creating in the process a new role of ‘curator/patron’. Equally, this process of realisation reflects the changing needs of contemporary art practice. Work of this scale and ambition would increasingly not be possible without the vision, perseverance and funding of these kinds of foundation. In Australia, this trend is demonstrated by two foundations: Kaldor Art Projects, and their commissioning of works by artists such as Christo and Jeanne-Claude, Gilbert & George and Jeff Koons; and the more recently formed Sherman Contemporary Art Foundation, whose first project was with Chinese artist Ai Weiwei. In this article, these examples are placed within the broader international context of foundation models such as Artangel, UK, Fondazione Nicola Trussardi, Milan, and The Public Art Fund, New York.

  17. Changing the focus of care: from curative to palliative care

    Directory of Open Access Journals (Sweden)

    Silvia Soffritti

    2014-06-01

    Full Text Available The improvements in the obstetrical and neonatal diagnosis and therapies have resulted into an increase in the survival rate of infants previously considered as non-viable. Debate is focusing on professionals’ behaviour about withdrawal or withholding of life sustaining treatment (LST and administration of palliative care for newborns whose conditions are incompatible with a prolonged life. Decisions about treatment should be made jointly by the professionals’ team and the family, placing the interest of the baby at the very heart of the decision process. It is very important that the environment in which the family has to make the decision is characterized by openness, dialogue and frankness. A proper and effective communication with parents is always necessary and can resolve any conflict caused by disagreement. Furthermore, parents need time in the decision making process. Other supports, which could help the family in the final decision are the possibility to ask for a specialist’s second opinion and the involvement of religious leaders and of an indipendent clinical ethics committee. Withholding or withdrawal of LST does not mean cessation of care for the baby, it means to change the focus of care from curative to palliative care. Proceedings of the 10th International Workshop on Neonatology · Cagliari (Italy · October 22nd-25th, 2014 · The last ten years, the next ten years in Neonatology Guest Editors: Vassilios Fanos, Michele Mussap, Gavino Faa, Apostolos Papageorgiou 

  18. Imitating manual curation of text-mined facts in biomedicine.

    Directory of Open Access Journals (Sweden)

    Raul Rodriguez-Esteban

    2006-09-01

    Full Text Available Text-mining algorithms make mistakes in extracting facts from natural-language texts. In biomedical applications, which rely on use of text-mined data, it is critical to assess the quality (the probability that the message is correctly extracted of individual facts--to resolve data conflicts and inconsistencies. Using a large set of almost 100,000 manually produced evaluations (most facts were independently reviewed more than once, producing independent evaluations, we implemented and tested a collection of algorithms that mimic human evaluation of facts provided by an automated information-extraction system. The performance of our best automated classifiers closely approached that of our human evaluators (ROC score close to 0.95. Our hypothesis is that, were we to use a larger number of human experts to evaluate any given sentence, we could implement an artificial-intelligence curator that would perform the classification job at least as accurately as an average individual human evaluator. We illustrated our analysis by visualizing the predicted accuracy of the text-mined relations involving the term cocaine.

  19. Database development and management

    CERN Document Server

    Chao, Lee

    2006-01-01

    Introduction to Database Systems Functions of a DatabaseDatabase Management SystemDatabase ComponentsDatabase Development ProcessConceptual Design and Data Modeling Introduction to Database Design Process Understanding Business ProcessEntity-Relationship Data Model Representing Business Process with Entity-RelationshipModelTable Structure and NormalizationIntroduction to TablesTable NormalizationTransforming Data Models to Relational Databases .DBMS Selection Transforming Data Models to Relational DatabasesEnforcing ConstraintsCreating Database for Business ProcessPhysical Design and Database

  20. Data Curation Program Development in U.S. Universities: The Georgia Institute of Technology Example

    Directory of Open Access Journals (Sweden)

    Tyler O. Walters

    2009-12-01

    Full Text Available The curation of scientific research data at U.S. universities is a story of enterprising individuals and of incremental progress. A small number of libraries and data centers who see the possibilities of becoming “digital information management centers” are taking entrepreneurial steps to extend beyond their traditional information assets and include managing scientific and scholarly research data. The Georgia Institute of Technology (GT has had a similar development path toward a data curation program based in its library. This paper will articulate GT’s program development, which the author offers as an experience common in U.S. universities. The main characteristic is a program devoid of top-level mandates and incentives, but rich with independent, “bottom-up” action. The paper will address program antecedents and context, inter-institutional partnerships that advance the library’s curation program, library organizational developments, partnerships with campus research communities, and a proposed model for curation program development. It concludes that despite the clear need for data curation put forth by researchers such as the groups of neuroscientists and bioscientists referenced in this paper, the university experience examined suggests that gathering resources for developing data curation programs at the institutional level is proving to be a quite onerous. However, and in spite of the challenges, some U.S. research universities are beginning to establish perceptible data curation programs.

  1. Supervised Learning for Detection of Duplicates in Genomic Sequence Databases.

    Directory of Open Access Journals (Sweden)

    Qingyu Chen

    Full Text Available First identified as an issue in 1996, duplication in biological databases introduces redundancy and even leads to inconsistency when contradictory information appears. The amount of data makes purely manual de-duplication impractical, and existing automatic systems cannot detect duplicates as precisely as can experts. Supervised learning has the potential to address such problems by building automatic systems that learn from expert curation to detect duplicates precisely and efficiently. While machine learning is a mature approach in other duplicate detection contexts, it has seen only preliminary application in genomic sequence databases.We developed and evaluated a supervised duplicate detection method based on an expert curated dataset of duplicates, containing over one million pairs across five organisms derived from genomic sequence databases. We selected 22 features to represent distinct attributes of the database records, and developed a binary model and a multi-class model. Both models achieve promising performance; under cross-validation, the binary model had over 90% accuracy in each of the five organisms, while the multi-class model maintains high accuracy and is more robust in generalisation. We performed an ablation study to quantify the impact of different sequence record features, finding that features derived from meta-data, sequence identity, and alignment quality impact performance most strongly. The study demonstrates machine learning can be an effective additional tool for de-duplication of genomic sequence databases. All Data are available as described in the supplementary material.

  2. Supervised Learning for Detection of Duplicates in Genomic Sequence Databases.

    Science.gov (United States)

    Chen, Qingyu; Zobel, Justin; Zhang, Xiuzhen; Verspoor, Karin

    2016-01-01

    First identified as an issue in 1996, duplication in biological databases introduces redundancy and even leads to inconsistency when contradictory information appears. The amount of data makes purely manual de-duplication impractical, and existing automatic systems cannot detect duplicates as precisely as can experts. Supervised learning has the potential to address such problems by building automatic systems that learn from expert curation to detect duplicates precisely and efficiently. While machine learning is a mature approach in other duplicate detection contexts, it has seen only preliminary application in genomic sequence databases. We developed and evaluated a supervised duplicate detection method based on an expert curated dataset of duplicates, containing over one million pairs across five organisms derived from genomic sequence databases. We selected 22 features to represent distinct attributes of the database records, and developed a binary model and a multi-class model. Both models achieve promising performance; under cross-validation, the binary model had over 90% accuracy in each of the five organisms, while the multi-class model maintains high accuracy and is more robust in generalisation. We performed an ablation study to quantify the impact of different sequence record features, finding that features derived from meta-data, sequence identity, and alignment quality impact performance most strongly. The study demonstrates machine learning can be an effective additional tool for de-duplication of genomic sequence databases. All Data are available as described in the supplementary material.

  3. Lessons learned from external hazards

    Energy Technology Data Exchange (ETDEWEB)

    Peinador, Miguel; Zerger, Benoit [European Commisison Joint Research Centre, Petten (Netherlands). Inst. for Energy and Transport; Ramos, Manuel Martin [European Commission Joint Research Centre, Brussels (Belgium). Nuclear Safety and Security Coordination; Wattrelos, Didier [Institut de Radioprotection et de Surete Nucleaire (IRSN), Fontenay-aux-Roses (France); Maqua, Michael [Gesellschaft fuer Anlagen- und Reaktorsicherheit (GRS) mbH, Koeln (Germany)

    2014-01-15

    This paper presents a study performed by the European Clearinghouse of the Joint Research Centre on Operational Experience for nuclear power plants in cooperation with IRSN and GRS covering events reported by nuclear power plants in relation to external hazards. It summarizes the review of 235 event reports from 3 different databases. The events were grouped in 9 categories according to the nature of the external hazard involved, and the specific lessons learned and recommendations that can be derived from each of these categories are presented. Additional 'cross-cutting' recommendations covering several or all the external hazards considered are also discussed. These recommendations can be useful in preventing this type of events from happening again or in limiting their consequences. The study was launched in 2010 and therefore it does not cover the Fukushima event. This paper presents the main findings and recommendations raised by this study. (orig.)

  4. The MAR databases: development and implementation of databases specific for marine metagenomics.

    Science.gov (United States)

    Klemetsen, Terje; Raknes, Inge A; Fu, Juan; Agafonov, Alexander; Balasundaram, Sudhagar V; Tartari, Giacomo; Robertsen, Espen; Willassen, Nils P

    2018-01-04

    We introduce the marine databases; MarRef, MarDB and MarCat (https://mmp.sfb.uit.no/databases/), which are publicly available resources that promote marine research and innovation. These data resources, which have been implemented in the Marine Metagenomics Portal (MMP) (https://mmp.sfb.uit.no/), are collections of richly annotated and manually curated contextual (metadata) and sequence databases representing three tiers of accuracy. While MarRef is a database for completely sequenced marine prokaryotic genomes, which represent a marine prokaryote reference genome database, MarDB includes all incomplete sequenced prokaryotic genomes regardless level of completeness. The last database, MarCat, represents a gene (protein) catalog of uncultivable (and cultivable) marine genes and proteins derived from marine metagenomics samples. The first versions of MarRef and MarDB contain 612 and 3726 records, respectively. Each record is built up of 106 metadata fields including attributes for sampling, sequencing, assembly and annotation in addition to the organism and taxonomic information. Currently, MarCat contains 1227 records with 55 metadata fields. Ontologies and controlled vocabularies are used in the contextual databases to enhance consistency. The user-friendly web interface lets the visitors browse, filter and search in the contextual databases and perform BLAST searches against the corresponding sequence databases. All contextual and sequence databases are freely accessible and downloadable from https://s1.sfb.uit.no/public/mar/. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  5. The history of a lesson

    DEFF Research Database (Denmark)

    Rasmussen, Mikkel Vedby

    2003-01-01

    and emphasises the need to study the history of lessons rather than the lessons of history. This approach shows that Munich is the end point of a constitutive history that begins in the failure of the Versailles treaty to create a durable European order following the First World War. The Munich lesson is thus......The article investigates the concept of lessons in IR. By means of a constructivist critique of the 'lessons literature', the article analyses one of the most important of IR lessons: that of Munich. Examining how the Munich lesson came about, the article shows the praxeological nature of lessons...... one element of the lesson of Versailles, which is a praxeology that defines how the West is to make peace, and against whom peace must be defended. The lesson of Versailles has been, at least in part, constitutive of the outbreak of the Cold War, and it continues to define the Western conception...

  6. Human Ageing Genomic Resources: new and updated databases

    Science.gov (United States)

    Tacutu, Robi; Thornton, Daniel; Johnson, Emily; Budovsky, Arie; Barardo, Diogo; Craig, Thomas; Diana, Eugene; Lehmann, Gilad; Toren, Dmitri; Wang, Jingwei; Fraifeld, Vadim E

    2018-01-01

    Abstract In spite of a growing body of research and data, human ageing remains a poorly understood process. Over 10 years ago we developed the Human Ageing Genomic Resources (HAGR), a collection of databases and tools for studying the biology and genetics of ageing. Here, we present HAGR’s main functionalities, highlighting new additions and improvements. HAGR consists of six core databases: (i) the GenAge database of ageing-related genes, in turn composed of a dataset of >300 human ageing-related genes and a dataset with >2000 genes associated with ageing or longevity in model organisms; (ii) the AnAge database of animal ageing and longevity, featuring >4000 species; (iii) the GenDR database with >200 genes associated with the life-extending effects of dietary restriction; (iv) the LongevityMap database of human genetic association studies of longevity with >500 entries; (v) the DrugAge database with >400 ageing or longevity-associated drugs or compounds; (vi) the CellAge database with >200 genes associated with cell senescence. All our databases are manually curated by experts and regularly updated to ensure a high quality data. Cross-links across our databases and to external resources help researchers locate and integrate relevant information. HAGR is freely available online (http://genomics.senescence.info/). PMID:29121237

  7. Organic Contamination Baseline Study: In NASA JSC Astromaterials Curation Laboratories. Summary Report

    Science.gov (United States)

    Calaway, Michael J.

    2013-01-01

    In preparation for OSIRIS-REx and other future sample return missions concerned with analyzing organics, we conducted an Organic Contamination Baseline Study for JSC Curation Labsoratories in FY12. For FY12 testing, organic baseline study focused only on molecular organic contamination in JSC curation gloveboxes: presumably future collections (i.e. Lunar, Mars, asteroid missions) would use isolation containment systems over only cleanrooms for primary sample storage. This decision was made due to limit historical data on curation gloveboxes, limited IR&D funds and Genesis routinely monitors organics in their ISO class 4 cleanrooms.

  8. The Annotometer; Encouraging uptake and use of freely available curation tools.

    OpenAIRE

    Zhe, Xiao Si; Edmunds, Scott; Li, Peter; Goodman, Laurie; Hunter, Christopher

    2016-01-01

    In recent years it has become clear that the amount of data being generated worldwide cannot be curated and annotated by any individual or small group. Currently, there is recognition that one of the best ways to provide ongoing curation is to employ the power of the community. To achieve this, the first hurdle to overcome was the development of user-friendly tools and apps that non-expert curators would be comfortable and capable of using. Such tools are now in place, inclu...

  9. Mathematics for Databases

    NARCIS (Netherlands)

    ir. Sander van Laar

    2007-01-01

    A formal description of a database consists of the description of the relations (tables) of the database together with the constraints that must hold on the database. Furthermore the contents of a database can be retrieved using queries. These constraints and queries for databases can very well be

  10. Databases and their application

    NARCIS (Netherlands)

    Grimm, E.C.; Bradshaw, R.H.W; Brewer, S.; Flantua, S.; Giesecke, T.; Lézine, A.M.; Takahara, H.; Williams, J.W.,Jr; Elias, S.A.; Mock, C.J.

    2013-01-01

    During the past 20 years, several pollen database cooperatives have been established. These databases are now constituent databases of the Neotoma Paleoecology Database, a public domain, multiproxy, relational database designed for Quaternary-Pliocene fossil data and modern surface samples. The

  11. DOT Online Database

    Science.gov (United States)

    Page Home Table of Contents Contents Search Database Search Login Login Databases Advisory Circulars accessed by clicking below: Full-Text WebSearch Databases Database Records Date Advisory Circulars 2092 5 data collection and distribution policies. Document Database Website provided by MicroSearch

  12. Information Models, Data Requirements, and Agile Data Curation

    Science.gov (United States)

    Hughes, John S.; Crichton, Dan; Ritschel, Bernd; Hardman, Sean; Joyner, Ron

    2015-04-01

    The Planetary Data System's next generation system, PDS4, is an example of the successful use of an ontology-based Information Model (IM) to drive the development and operations of a data system. In traditional systems engineering, requirements or statements about what is necessary for the system are collected and analyzed for input into the design stage of systems development. With the advent of big data the requirements associated with data have begun to dominate and an ontology-based information model can be used to provide a formalized and rigorous set of data requirements. These requirements address not only the usual issues of data quantity, quality, and disposition but also data representation, integrity, provenance, context, and semantics. In addition the use of these data requirements during system's development has many characteristics of Agile Curation as proposed by Young et al. [Taking Another Look at the Data Management Life Cycle: Deconstruction, Agile, and Community, AGU 2014], namely adaptive planning, evolutionary development, early delivery, continuous improvement, and rapid and flexible response to change. For example customers can be satisfied through early and continuous delivery of system software and services that are configured directly from the information model. This presentation will describe the PDS4 architecture and its three principle parts: the ontology-based Information Model (IM), the federated registries and repositories, and the REST-based service layer for search, retrieval, and distribution. The development of the IM will be highlighted with special emphasis on knowledge acquisition, the impact of the IM on development and operations, and the use of shared ontologies at multiple governance levels to promote system interoperability and data correlation.

  13. Curating NASA's future extraterrestrial sample collections: How do we achieve maximum proficiency?

    Science.gov (United States)

    McCubbin, Francis; Evans, Cynthia; Allton, Judith; Fries, Marc; Righter, Kevin; Zolensky, Michael; Zeigler, Ryan

    2016-07-01

    Introduction: The Astromaterials Acquisition and Curation Office (henceforth referred to herein as NASA Curation Office) at NASA Johnson Space Center (JSC) is responsible for curating all of NASA's extraterrestrial samples. Under the governing document, NASA Policy Directive (NPD) 7100.10E "Curation of Extraterrestrial Materials", JSC is charged with "The curation of all extraterrestrial material under NASA control, including future NASA missions." The Directive goes on to define Curation as including "…documentation, preservation, preparation, and distribution of samples for research, education, and public outreach." Here we describe some of the ongoing efforts to ensure that the future activities of the NASA Curation Office are working to-wards a state of maximum proficiency. Founding Principle: Curatorial activities began at JSC (Manned Spacecraft Center before 1973) as soon as design and construction planning for the Lunar Receiving Laboratory (LRL) began in 1964 [1], not with the return of the Apollo samples in 1969, nor with the completion of the LRL in 1967. This practice has since proven that curation begins as soon as a sample return mission is conceived, and this founding principle continues to return dividends today [e.g., 2]. The Next Decade: Part of the curation process is planning for the future, and we refer to these planning efforts as "advanced curation" [3]. Advanced Curation is tasked with developing procedures, technology, and data sets necessary for curating new types of collections as envisioned by NASA exploration goals. We are (and have been) planning for future curation, including cold curation, extended curation of ices and volatiles, curation of samples with special chemical considerations such as perchlorate-rich samples, curation of organically- and biologically-sensitive samples, and the use of minimally invasive analytical techniques (e.g., micro-CT, [4]) to characterize samples. These efforts will be useful for Mars Sample Return

  14. BGDB: a database of bivalent genes.

    Science.gov (United States)

    Li, Qingyan; Lian, Shuabin; Dai, Zhiming; Xiang, Qian; Dai, Xianhua

    2013-01-01

    Bivalent gene is a gene marked with both H3K4me3 and H3K27me3 epigenetic modification in the same area, and is proposed to play a pivotal role related to pluripotency in embryonic stem (ES) cells. Identification of these bivalent genes and understanding their functions are important for further research of lineage specification and embryo development. So far, lots of genome-wide histone modification data were generated in mouse and human ES cells. These valuable data make it possible to identify bivalent genes, but no comprehensive data repositories or analysis tools are available for bivalent genes currently. In this work, we develop BGDB, the database of bivalent genes. The database contains 6897 bivalent genes in human and mouse ES cells, which are manually collected from scientific literature. Each entry contains curated information, including genomic context, sequences, gene ontology and other relevant information. The web services of BGDB database were implemented with PHP + MySQL + JavaScript, and provide diverse query functions. Database URL: http://dailab.sysu.edu.cn/bgdb/

  15. Clinical outcomes of Y90 radioembolization for recurrent hepatocellular carcinoma following curative resection

    Energy Technology Data Exchange (ETDEWEB)

    Ali, Rehan; Riaz, Ahsun; Gabr, Ahmed; Abouchaleh, Nadine; Mora, Ronald; Al Asadi, Ali [Northwestern University, Department of Radiology, Section of Interventional Radiology, Chicago, IL (United States); Caicedo, Juan Carlos; Abecassis, Michael; Katariya, Nitin [Northwestern University, Department of Surgery, Division of Transplant Surgery, Chicago, IL (United States); Maddur, Haripriya; Kulik, Laura [Northwestern University, Department of Medicine, Division of Hepatology, Chicago, IL (United States); Lewandowski, Robert J. [Northwestern University, Department of Radiology, Section of Interventional Radiology, Chicago, IL (United States); Northwestern University, Department of Medicine, Division of Hematology and Oncology, Chicago, IL (United States); Salem, Riad [Northwestern University, Department of Radiology, Section of Interventional Radiology, Chicago, IL (United States); Northwestern University, Department of Surgery, Division of Transplant Surgery, Chicago, IL (United States); Northwestern University, Department of Medicine, Division of Hematology and Oncology, Chicago, IL (United States)

    2017-12-15

    To assess safety/efficacy of yttrium-90 radioembolization (Y90) in patients with recurrent hepatocellular carcinoma (HCC) following curative surgical resection. With IRB approval, we searched our prospectively acquired database for patients that were treated with Y90 for recurrent disease following resection. Baseline characteristics and bilirubin toxicities following Y90 were evaluated. Intention-to-treat overall survival (OS) and time-to-progression (TTP) from Y90 were assessed. Forty-one patients met study inclusion criteria. Twenty-six (63%) patients had undergone minor (≤3 hepatic segments) resection while 15 (37%) patients underwent major (>3 hepatic segments) resections. Two patients (5%) had biliary-enteric anastomoses created during surgical resection. The median time from HCC resection to the first radioembolization was 17 months (95% CI: 13-37). The median number of Y90 treatment sessions was 1 (range: 1-5). Ten patients received (entire remnant) lobar Y90 treatment while 31 patients received selective (≤2 hepatic segments) treatment. Grades 1/2/3/4 bilirubin toxicity were seen in nine (22%), four (10%), four (10%), and zero (0%) patients following Y90. No differences in bilirubin toxicities were identified when comparing lobar with selective approaches (P = 0.20). No post-Y90 infectious complications were identified. Median TTP and OS were 11.3 (CI: 6.5-15.5) and 22.1 months (CI: 10.3-31.3), respectively. Radioembolization is a safe and effective method for treating recurrent HCC following surgical resection, with prolonged TTP and promising survival outcomes. (orig.)

  16. Dietary Supplement Ingredient Database

    Science.gov (United States)

    ... and US Department of Agriculture Dietary Supplement Ingredient Database Toggle navigation Menu Home About DSID Mission Current ... values can be saved to build a small database or add to an existing database for national, ...

  17. Energy Consumption Database

    Science.gov (United States)

    Consumption Database The California Energy Commission has created this on-line database for informal reporting ) classifications. The database also provides easy downloading of energy consumption data into Microsoft Excel (XLSX

  18. 3DSwap: Curated knowledgebase of proteins involved in 3D domain swapping

    KAUST Repository

    Shameer, Khader; Shingate, Prashant N.; Manjunath, S. C. P.; Karthika, M.; Pugalenthi, Ganesan; Sowdhamini, Ramanathan

    2011-01-01

    structures in oligomeric conformation. Protein structures in swapped conformations perform diverse functional roles and are also associated with deposition diseases in humans. We have performed in-depth literature curation and structural bioinformatics

  19. DCC DIFFUSE Standards Frameworks: A Standards Path through the Curation Lifecycle

    Directory of Open Access Journals (Sweden)

    Sarah Higgins

    2009-10-01

    Full Text Available Normal 0 false false false MicrosoftInternetExplorer4 DCC DIFFUSE Standards Frameworks aims to offer domain specific advice on standards relevant to digital preservation and curation, to help curators identify which standards they should be using and where they can be appropriately implemented, to ensure authoritative digital material. The Project uses the DCC Curation Lifecycle Model and Web 2.0 technology, to visually present standards frameworks for a number of disciplines. The Digital Curation Centre (DCC is actively working with a different relevant organisations to present searchable frameworks of standards, for a number of domains. These include digital repositories, records management, the geo-information sector, archives and the museum sector. Other domains, such as e-science, will shortly be investigated.

  20. Planning Related to the Curation and Processing of Returned Martian Samples

    Science.gov (United States)

    McCubbin, F. M.; Harrington, A. D.

    2018-04-01

    Many of the planning activities in the NASA Astromaterials Acquisition and Curation Office at JSC are centered around Mars Sample Return. The importance of contamination knowledge and the benefits of a mobile/modular receiving facility are discussed.

  1. EURO-CARES as Roadmap for a European Sample Curation Facility

    Science.gov (United States)

    Brucato, J. R.; Russell, S.; Smith, C.; Hutzler, A.; Meneghin, A.; Aléon, J.; Bennett, A.; Berthoud, L.; Bridges, J.; Debaille, V.; Ferrière, L.; Folco, L.; Foucher, F.; Franchi, I.; Gounelle, M.; Grady, M.; Leuko, S.; Longobardo, A.; Palomba, E.; Pottage, T.; Rettberg, P.; Vrublevskis, J.; Westall, F.; Zipfel, J.; Euro-Cares Team

    2018-04-01

    EURO-CARES is a three-year multinational project funded under the European Commission Horizon2020 research program to develop a roadmap for a European Extraterrestrial Sample Curation Facility for samples returned from solar system missions.

  2. Surveillance Patterns After Curative-Intent Colorectal Cancer Surgery in Ontario

    Directory of Open Access Journals (Sweden)

    Jensen Tan

    2014-01-01

    Full Text Available BACKGROUND: Postoperative surveillance following curative-intent resection of colorectal cancer (CRC is variably performed due to existing guideline differences and to the limited data supporting different strategies.

  3. Palliative or curative treatment intent affects communication in radiation therapy consultations.

    NARCIS (Netherlands)

    Timmermans, L.; Maazen, R.W.M. van der; Leer, J.W.H.; Kraaimaat, F.W.

    2006-01-01

    OBJECTIVE: To assess whether communication in radiotherapy consultations is affected by palliative or curative treatment intent. SUBJECTS AND METHODS: The study involved 160 patients and 8 radiation oncologists. Eighty patients visited the radiation oncologist (RO) for palliative treatment and 80

  4. Curative and eradicant action of fungicides to control Phakopsora pachyrhizi in soybean plants

    Directory of Open Access Journals (Sweden)

    Erlei Melo Reis

    Full Text Available ABSTRACT Experiments were carried out in a growth chamber and laboratory to quantify the curative and eradicant actions of fungicides in Asian soybean rust control. The experiments were conducted with the CD 214 RR cultivar, assessing the following fungicides, separately or in association, chlorothalonil, flutriafol, cyproconazole + trifloxystrobin, epoxiconazole + pyraclostrobin, cyproconazole + azoxystrobin, and cyproconazole + picoxystrobin. The fungicides were applied at four (curative and nine days after inoculation (eradicant treatment. Treatments were evaluated according to the density of lesions and uredia/cm2, and the eradicant treatment was assessed based on the necrosis of lesions/uredia and on uredospore viability. Except for the fungicide chlorothalonil, there was curative action of latent/virtual infections by the fungicides. Penetrant fungicides that are absorbed have curative and eradicant action to soybean rust.

  5. Biocuration at the Saccharomyces genome database.

    Science.gov (United States)

    Skrzypek, Marek S; Nash, Robert S

    2015-08-01

    Saccharomyces Genome Database is an online resource dedicated to managing information about the biology and genetics of the model organism, yeast (Saccharomyces cerevisiae). This information is derived primarily from scientific publications through a process of human curation that involves manual extraction of data and their organization into a comprehensive system of knowledge. This system provides a foundation for further analysis of experimental data coming from research on yeast as well as other organisms. In this review we will demonstrate how biocuration and biocurators add a key component, the biological context, to our understanding of how genes, proteins, genomes and cells function and interact. We will explain the role biocurators play in sifting through the wealth of biological data to incorporate and connect key information. We will also discuss the many ways we assist researchers with their various research needs. We hope to convince the reader that manual curation is vital in converting the flood of data into organized and interconnected knowledge, and that biocurators play an essential role in the integration of scientific information into a coherent model of the cell. © 2015 Wiley Periodicals, Inc.

  6. The Structure-Function Linkage Database.

    Science.gov (United States)

    Akiva, Eyal; Brown, Shoshana; Almonacid, Daniel E; Barber, Alan E; Custer, Ashley F; Hicks, Michael A; Huang, Conrad C; Lauck, Florian; Mashiyama, Susan T; Meng, Elaine C; Mischel, David; Morris, John H; Ojha, Sunil; Schnoes, Alexandra M; Stryke, Doug; Yunes, Jeffrey M; Ferrin, Thomas E; Holliday, Gemma L; Babbitt, Patricia C

    2014-01-01

    The Structure-Function Linkage Database (SFLD, http://sfld.rbvi.ucsf.edu/) is a manually curated classification resource describing structure-function relationships for functionally diverse enzyme superfamilies. Members of such superfamilies are diverse in their overall reactions yet share a common ancestor and some conserved active site features associated with conserved functional attributes such as a partial reaction. Thus, despite their different functions, members of these superfamilies 'look alike', making them easy to misannotate. To address this complexity and enable rational transfer of functional features to unknowns only for those members for which we have sufficient functional information, we subdivide superfamily members into subgroups using sequence information, and lastly into families, sets of enzymes known to catalyze the same reaction using the same mechanistic strategy. Browsing and searching options in the SFLD provide access to all of these levels. The SFLD offers manually curated as well as automatically classified superfamily sets, both accompanied by search and download options for all hierarchical levels. Additional information includes multiple sequence alignments, tab-separated files of functional and other attributes, and sequence similarity networks. The latter provide a new and intuitively powerful way to visualize functional trends mapped to the context of sequence similarity.

  7. The Structure–Function Linkage Database

    Science.gov (United States)

    Akiva, Eyal; Brown, Shoshana; Almonacid, Daniel E.; Barber, Alan E.; Custer, Ashley F.; Hicks, Michael A.; Huang, Conrad C.; Lauck, Florian; Mashiyama, Susan T.; Meng, Elaine C.; Mischel, David; Morris, John H.; Ojha, Sunil; Schnoes, Alexandra M.; Stryke, Doug; Yunes, Jeffrey M.; Ferrin, Thomas E.; Holliday, Gemma L.; Babbitt, Patricia C.

    2014-01-01

    The Structure–Function Linkage Database (SFLD, http://sfld.rbvi.ucsf.edu/) is a manually curated classification resource describing structure–function relationships for functionally diverse enzyme superfamilies. Members of such superfamilies are diverse in their overall reactions yet share a common ancestor and some conserved active site features associated with conserved functional attributes such as a partial reaction. Thus, despite their different functions, members of these superfamilies ‘look alike’, making them easy to misannotate. To address this complexity and enable rational transfer of functional features to unknowns only for those members for which we have sufficient functional information, we subdivide superfamily members into subgroups using sequence information, and lastly into families, sets of enzymes known to catalyze the same reaction using the same mechanistic strategy. Browsing and searching options in the SFLD provide access to all of these levels. The SFLD offers manually curated as well as automatically classified superfamily sets, both accompanied by search and download options for all hierarchical levels. Additional information includes multiple sequence alignments, tab-separated files of functional and other attributes, and sequence similarity networks. The latter provide a new and intuitively powerful way to visualize functional trends mapped to the context of sequence similarity. PMID:24271399

  8. Curating for engagement: Identifying the nature and impact of organizational marketing strategies on Pinterest

    OpenAIRE

    Saxton, Gregory D.; Ghosh, Amanda

    2016-01-01

    In an increasingly overloaded information environment sparked by the explosion of digital media, the ability to curate content has taken on greater importance. This study begins with the supposition that businesses that are able to adopt a content curation role and help consumers sort through the daily influx of information may be more successful in attracting, engaging, and retaining customers while fostering brand awareness and word of mouth. Accordingly, this study investigates organizatio...

  9. The value of imaging examinations in diagnosis and curative effect evaluation of breast cancer

    International Nuclear Information System (INIS)

    Xia Xiaotian; Zhang Yongxue

    2010-01-01

    Breast cancer is a serious impact on women's physical and mental health and a life-threatening common disease. Imaging examinations have great significances in diagnosing and evaluating curative effect on breast cancer. This article aims to introduce and comprehensive the value of diagnosis and curative effect evaluation of breast cancer in the context of imaging examinations (ultrasonography, mammography, breast CT, breast MRI, breast 99 Tc m -MIBI imaging, PET, PET-CT, etc). (authors)

  10. Collecting Taxes Database

    Data.gov (United States)

    US Agency for International Development — The Collecting Taxes Database contains performance and structural indicators about national tax systems. The database contains quantitative revenue performance...

  11. USAID Anticorruption Projects Database

    Data.gov (United States)

    US Agency for International Development — The Anticorruption Projects Database (Database) includes information about USAID projects with anticorruption interventions implemented worldwide between 2007 and...

  12. NoSQL databases

    OpenAIRE

    Mrozek, Jakub

    2012-01-01

    This thesis deals with database systems referred to as NoSQL databases. In the second chapter, I explain basic terms and the theory of database systems. A short explanation is dedicated to database systems based on the relational data model and the SQL standardized query language. Chapter Three explains the concept and history of the NoSQL databases, and also presents database models, major features and the use of NoSQL databases in comparison with traditional database systems. In the fourth ...

  13. Information Assurance Intrusion Detection Sensor Database Design: Lessons Learned

    National Research Council Canada - National Science Library

    Spink, Brian

    2001-01-01

    ... that could lead to compromise or denial-of-service. The obvious benefits of using a relational data model and SQL to help solve this problem continue to be observed at Air Force Research Laboratory (AFRL...

  14. Data Curation Education Grounded in Earth Sciences and the Science of Data

    Science.gov (United States)

    Palmer, C. L.

    2015-12-01

    This presentation looks back over ten years of experience advancing data curation education at two Information Schools, highlighting the vital role of earth science case studies, expertise, and collaborations in development of curriculum and internships. We also consider current data curation practices and workforce demand in data centers in the geosciences, drawing on studies conducted in the Data Curation Education in Research Centers (DCERC) initiative and the Site-Based Data Curation project. Outcomes from this decade of data curation research and education has reinforced the importance of key areas of information science in preparing data professionals to respond to the needs of user communities, provide services across disciplines, invest in standards and interoperability, and promote open data practices. However, a serious void remains in principles to guide education and practice that are distinct to the development of data systems and services that meet both local and global aims. We identify principles emerging from recent empirical studies on the reuse value of data in the earth sciences and propose an approach for advancing data curation education that depends on systematic coordination with data intensive research and propagation of current best practices from data centers into curriculum. This collaborative model can increase both domain-based and cross-disciplinary expertise among data professionals, ultimately improving data systems and services in our universities and data centers while building the new base of knowledge needed for a foundational science of data.

  15. JSC Advanced Curation: Research and Development for Current Collections and Future Sample Return Mission Demands

    Science.gov (United States)

    Fries, M. D.; Allen, C. C.; Calaway, M. J.; Evans, C. A.; Stansbery, E. K.

    2015-01-01

    Curation of NASA's astromaterials sample collections is a demanding and evolving activity that supports valuable science from NASA missions for generations, long after the samples are returned to Earth. For example, NASA continues to loan hundreds of Apollo program samples to investigators every year and those samples are often analyzed using instruments that did not exist at the time of the Apollo missions themselves. The samples are curated in a manner that minimizes overall contamination, enabling clean, new high-sensitivity measurements and new science results over 40 years after their return to Earth. As our exploration of the Solar System progresses, upcoming and future NASA sample return missions will return new samples with stringent contamination control, sample environmental control, and Planetary Protection requirements. Therefore, an essential element of a healthy astromaterials curation program is a research and development (R&D) effort that characterizes and employs new technologies to maintain current collections and enable new missions - an Advanced Curation effort. JSC's Astromaterials Acquisition & Curation Office is continually performing Advanced Curation research, identifying and defining knowledge gaps about research, development, and validation/verification topics that are critical to support current and future NASA astromaterials sample collections. The following are highlighted knowledge gaps and research opportunities.

  16. Practices of research data curation in institutional repositories: A qualitative view from repository staff.

    Science.gov (United States)

    Lee, Dong Joon; Stvilia, Besiki

    2017-01-01

    The importance of managing research data has been emphasized by the government, funding agencies, and scholarly communities. Increased access to research data increases the impact and efficiency of scientific activities and funding. Thus, many research institutions have established or plan to establish research data curation services as part of their Institutional Repositories (IRs). However, in order to design effective research data curation services in IRs, and to build active research data providers and user communities around those IRs, it is essential to study current data curation practices and provide rich descriptions of the sociotechnical factors and relationships shaping those practices. Based on 13 interviews with 15 IR staff members from 13 large research universities in the United States, this paper provides a rich, qualitative description of research data curation and use practices in IRs. In particular, the paper identifies data curation and use activities in IRs, as well as their structures, roles played, skills needed, contradictions and problems present, solutions sought, and workarounds applied. The paper can inform the development of best practice guides, infrastructure and service templates, as well as education in research data curation in Library and Information Science (LIS) schools.

  17. Our journey to digital curation of the Jeghers Medical Index.

    Science.gov (United States)

    Gawdyda, Lori; Carter, Kimbroe; Willson, Mark; Bedford, Denise

    2017-07-01

    Harold Jeghers, a well-known medical educator of the twentieth century, maintained a print collection of about one million medical articles from the late 1800s to the 1990s. This case study discusses how a print collection of these articles was transformed to a digital database. Staff in the Jeghers Medical Index, St. Elizabeth Youngstown Hospital, converted paper articles to Adobe portable document format (PDF)/A-1a files. Optical character recognition was used to obtain searchable text. The data were then incorporated into a specialized database. Lastly, articles were matched to PubMed bibliographic metadata through automation and human review. An online database of the collection was ultimately created. The collection was made part of a discovery search service, and semantic technologies have been explored as a method of creating access points. This case study shows how a small medical library made medical writings of the nineteenth and twentieth centuries available in electronic format for historic or semantic research, highlighting the efficiencies of contemporary information technology.

  18. OzFlux data: network integration from collection to curation

    Directory of Open Access Journals (Sweden)

    P. Isaac

    2017-06-01

    Full Text Available Measurement of the exchange of energy and mass between the surface and the atmospheric boundary-layer by the eddy covariance technique has undergone great change in the last 2 decades. Early studies of these exchanges were confined to brief field campaigns in carefully controlled conditions followed by months of data analysis. Current practice is to run tower-based eddy covariance systems continuously over several years due to the need for continuous monitoring as part of a global effort to develop local-, regional-, continental- and global-scale budgets of carbon, water and energy. Efficient methods of processing the increased quantities of data are needed to maximise the time available for analysis and interpretation. Standardised methods are needed to remove differences in data processing as possible contributors to observed spatial variability. Furthermore, public availability of these data sets assists with undertaking global research efforts. The OzFlux data path has been developed (i to provide a standard set of quality control and post-processing tools across the network, thereby facilitating inter-site integration and spatial comparisons; (ii to increase the time available to researchers for analysis and interpretation by reducing the time spent collecting and processing data; (iii to propagate both data and metadata to the final product; and (iv to facilitate the use of the OzFlux data by adopting a standard file format and making the data available from web-based portals. Discovery of the OzFlux data set is facilitated through incorporation in FLUXNET data syntheses and the publication of collection metadata via the RIF-CS format. This paper serves two purposes. The first is to describe the data sets, along with their quality control and post-processing, for the other papers of this Special Issue. The second is to provide an example of one solution to the data collection and curation challenges that are encountered by similar flux

  19. Pain palliation therapy of bone metastases: palliative or curative?

    International Nuclear Information System (INIS)

    Fischer, M.

    2007-01-01

    about 6 months, for 153 Sm and 186 Re 6-10 weeks. A response rate of about 70-80% is reported, the onset can be expected after 7-day after administration of 153 Sm and 186 Re, after 2 weeks using 89 Sr. This radionuclide therapy can be performed on outpatient basis, depending on national regulations. Radionuclide therapy alone is no curative treatment of bone metastases, but combination therapy with chemo- and/or radiation therapy have shown strong synergistic effect, improving the response on pain syndrome and disease control as well as prolonging mean survival as shown in several clinical trials

  20. Sample Curation in Support of the OSIRIS-REx Asteroid Sample Return Mission

    Science.gov (United States)

    Righter, Kevin; Nakamura-Messenger, Keiko

    2017-01-01

    The OSIRIS-REx asteroid sample return mission launched to asteroid Bennu Sept. 8, 2016. The spacecraft will arrive at Bennu in late 2019, orbit and map the asteroid, and perform a touch and go (TAG) sampling maneuver in July 2020. After sample is stowed and confirmed the spacecraft will return to Earth, and the sample return capsule (SRC) will land in Utah in September 2023. Samples will be recovered from Utah [2] and then transported and stored in a new sample cleanroom at NASA Johnson Space Center in Houston [3]. The materials curated for the mission are described here. a) Materials Archive and Witness Plate Collection: The SRC and TAGSAM were built between March 2014 and Summer of 2015, and instruments (OTES,OVIRS, OLA, OCAMS, REXIS) were integrated from Summer 2015 until May 2016. A total of 395 items were received for the materials archive at NASA-JSC, with archiving finishing 30 days after launch (with the final archived items being related to launch operations)[4]. The materials fall into several general categories including metals (stainless steel, aluminum, titanium alloys, brass and BeCu alloy), epoxies, paints, polymers, lubricants, non-volatile-residue samples (NVR), sapphire, and various miscellaneous materials. All through the ATLO process (from March 2015 until late August 2016) contamination knowledge witness plates (Si wafer and Al foil) were deployed in the various cleanrooms in Denver and KSC to provide an additional record of particle counts and volatiles that is archived for current and future scientific studies. These plates were deployed in roughly monthly increments with each unit containing 4 Si wafers and 4 Al foils. We archived 128 individual witness plates (64 Si wafers and 64 Al foils); one of each witness plate (Si and Al) was analyzed immediately by the science team after archiving, while the remaining 3 of each are archived indefinitely. Information about each material archived is stored in an extensive database at NASA-JSC, and key

  1. Development of Database and Lecture Book for Nuclear Water Chemistry

    International Nuclear Information System (INIS)

    Maeng, Wan Young; Kim, U. C.; Na, J. W.; Choi, B. S.; Lee, E. H.; Kim, K. H.; Kim, K. M.; Kim, S. H.; Im, K. S.

    2010-02-01

    In order to establish a systematic and synthetic knowledge system of nuclear water chemistry, we held nuclear water chemistry experts group meetings. We discussed the way of buildup and propagation of nuclear water chemistry knowledge with domestic experts. We obtained a lot of various opinions that made the good use of this research project. The results will be applied to continuous buildup of domestic nuclear water chemistry knowledge database. Lessons in water chemistry of nuclear power plants (NPPs) have been opened in Nuclear Training and education Center, KAERI to educate the new generation who are working and will be working at the department of water chemistry of NPPs. The lessons were 17 and lesson period was from 12th May through 5th November. In order to progress the programs, many water chemistry experts were invited. They gave lectures to the younger generation once a week for 2 h about their experiences obtained during working on water chemistry of NPPs. The number of attendance was 290. The lessons were very effective and the lesson data will be used to make database for continuous use

  2. Structural and sequence variants in patients with Silver-Russell syndrome or similar features-Curation of a disease database

    DEFF Research Database (Denmark)

    Tümer, Zeynep; López-Hernández, Julia Angélica; Netchine, Irène

    2018-01-01

    data of these patients. The clinical features are scored according to the Netchine-Harbison clinical scoring system (NH-CSS), which has recently been accepted as standard by consensus. The structural and sequence variations are reviewed and where necessary redescribed according to recent...

  3. AllergenOnline: A peer-reviewed, curated allergen database to assess novel food proteins for potential cross-reactivity

    NARCIS (Netherlands)

    Goodman, Richard E.; Ebisawa, Motohiro; Ferreira, Fatima; Sampson, Hugh A.; van Ree, Ronald; Vieths, Stefan; Baumert, Joseph L.; Bohle, Barbara; Lalithambika, Sreedevi; Wise, John; Taylor, Steve L.

    2016-01-01

    Increasingly regulators are demanding evaluation of potential allergenicity of foods prior to marketing. Primary risks are the transfer of allergens or potentially cross-reactive proteins into new foods. AllergenOnline was developed in 2005 as a peer-reviewed bioinformatics platform to evaluate

  4. PrimateLit Database

    Science.gov (United States)

    Primate Info Net Related Databases NCRR PrimateLit: A bibliographic database for primatology Top of any problems with this service. We welcome your feedback. The PrimateLit database is no longer being Resources, National Institutes of Health. The database is a collaborative project of the Wisconsin Primate

  5. Antarctic Meteorite Classification and Petrographic Database

    Science.gov (United States)

    Todd, Nancy S.; Satterwhite, C. E.; Righter, Kevin

    2011-01-01

    The Antarctic Meteorite collection, which is comprised of over 18,700 meteorites, is one of the largest collections of meteorites in the world. These meteorites have been collected since the late 1970's as part of a three-agency agreement between NASA, the National Science Foundation, and the Smithsonian Institution [1]. Samples collected each season are analyzed at NASA s Meteorite Lab and the Smithsonian Institution and results are published twice a year in the Antarctic Meteorite Newsletter, which has been in publication since 1978. Each newsletter lists the samples collected and processed and provides more in-depth details on selected samples of importance to the scientific community. Data about these meteorites is also published on the NASA Curation website [2] and made available through the Meteorite Classification Database allowing scientists to search by a variety of parameters

  6. KALIMER database development

    Energy Technology Data Exchange (ETDEWEB)

    Jeong, Kwan Seong; Lee, Yong Bum; Jeong, Hae Yong; Ha, Kwi Seok

    2003-03-01

    KALIMER database is an advanced database to utilize the integration management for liquid metal reactor design technology development using Web applications. KALIMER design database is composed of results database, Inter-Office Communication (IOC), 3D CAD database, and reserved documents database. Results database is a research results database during all phase for liquid metal reactor design technology development of mid-term and long-term nuclear R and D. IOC is a linkage control system inter sub project to share and integrate the research results for KALIMER. 3D CAD database is a schematic overview for KALIMER design structure. And reserved documents database is developed to manage several documents and reports since project accomplishment.

  7. KALIMER database development

    International Nuclear Information System (INIS)

    Jeong, Kwan Seong; Lee, Yong Bum; Jeong, Hae Yong; Ha, Kwi Seok

    2003-03-01

    KALIMER database is an advanced database to utilize the integration management for liquid metal reactor design technology development using Web applications. KALIMER design database is composed of results database, Inter-Office Communication (IOC), 3D CAD database, and reserved documents database. Results database is a research results database during all phase for liquid metal reactor design technology development of mid-term and long-term nuclear R and D. IOC is a linkage control system inter sub project to share and integrate the research results for KALIMER. 3D CAD database is a schematic overview for KALIMER design structure. And reserved documents database is developed to manage several documents and reports since project accomplishment

  8. WikiPathways: a multifaceted pathway database bridging metabolomics to other omics research.

    Science.gov (United States)

    Slenter, Denise N; Kutmon, Martina; Hanspers, Kristina; Riutta, Anders; Windsor, Jacob; Nunes, Nuno; Mélius, Jonathan; Cirillo, Elisa; Coort, Susan L; Digles, Daniela; Ehrhart, Friederike; Giesbertz, Pieter; Kalafati, Marianthi; Martens, Marvin; Miller, Ryan; Nishida, Kozo; Rieswijk, Linda; Waagmeester, Andra; Eijssen, Lars M T; Evelo, Chris T; Pico, Alexander R; Willighagen, Egon L

    2018-01-04

    WikiPathways (wikipathways.org) captures the collective knowledge represented in biological pathways. By providing a database in a curated, machine readable way, omics data analysis and visualization is enabled. WikiPathways and other pathway databases are used to analyze experimental data by research groups in many fields. Due to the open and collaborative nature of the WikiPathways platform, our content keeps growing and is getting more accurate, making WikiPathways a reliable and rich pathway database. Previously, however, the focus was primarily on genes and proteins, leaving many metabolites with only limited annotation. Recent curation efforts focused on improving the annotation of metabolism and metabolic pathways by associating unmapped metabolites with database identifiers and providing more detailed interaction knowledge. Here, we report the outcomes of the continued growth and curation efforts, such as a doubling of the number of annotated metabolite nodes in WikiPathways. Furthermore, we introduce an OpenAPI documentation of our web services and the FAIR (Findable, Accessible, Interoperable and Reusable) annotation of resources to increase the interoperability of the knowledge encoded in these pathways and experimental omics data. New search options, monthly downloads, more links to metabolite databases, and new portals make pathway knowledge more effortlessly accessible to individual researchers and research communities. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  9. The Danish Microbiology Database (MiBa) 2010 to 2013.

    Science.gov (United States)

    Voldstedlund, M; Haarh, M; Mølbak, K

    2014-01-09

    The Danish Microbiology Database (MiBa) is a national database that receives copies of reports from all Danish departments of clinical microbiology. The database was launched in order to provide healthcare personnel with nationwide access to microbiology reports and to enable real-time surveillance of communicable diseases and microorganisms. The establishment and management of MiBa has been a collaborative process among stakeholders, and the present paper summarises lessons learned from this nationwide endeavour which may be relevant to similar projects in the rapidly changing landscape of health informatics.

  10. Logical database design principles

    CERN Document Server

    Garmany, John; Clark, Terry

    2005-01-01

    INTRODUCTION TO LOGICAL DATABASE DESIGNUnderstanding a Database Database Architectures Relational Databases Creating the Database System Development Life Cycle (SDLC)Systems Planning: Assessment and Feasibility System Analysis: RequirementsSystem Analysis: Requirements Checklist Models Tracking and Schedules Design Modeling Functional Decomposition DiagramData Flow Diagrams Data Dictionary Logical Structures and Decision Trees System Design: LogicalSYSTEM DESIGN AND IMPLEMENTATION The ER ApproachEntities and Entity Types Attribute Domains AttributesSet-Valued AttributesWeak Entities Constraint

  11. An Interoperable Cartographic Database

    OpenAIRE

    Slobodanka Ključanin; Zdravko Galić

    2007-01-01

    The concept of producing a prototype of interoperable cartographic database is explored in this paper, including the possibilities of integration of different geospatial data into the database management system and their visualization on the Internet. The implementation includes vectorization of the concept of a single map page, creation of the cartographic database in an object-relation database, spatial analysis, definition and visualization of the database content in the form of a map on t...

  12. Software listing: CHEMTOX database

    International Nuclear Information System (INIS)

    Moskowitz, P.D.

    1993-01-01

    Initially launched in 1983, the CHEMTOX Database was among the first microcomputer databases containing hazardous chemical information. The database is used in many industries and government agencies in more than 17 countries. Updated quarterly, the CHEMTOX Database provides detailed environmental and safety information on 7500-plus hazardous substances covered by dozens of regulatory and advisory sources. This brief listing describes the method of accessing data and provides ordering information for those wishing to obtain the CHEMTOX Database

  13. Human Variome Project Quality Assessment Criteria for Variation Databases.

    Science.gov (United States)

    Vihinen, Mauno; Hancock, John M; Maglott, Donna R; Landrum, Melissa J; Schaafsma, Gerard C P; Taschner, Peter

    2016-06-01

    Numerous databases containing information about DNA, RNA, and protein variations are available. Gene-specific variant databases (locus-specific variation databases, LSDBs) are typically curated and maintained for single genes or groups of genes for a certain disease(s). These databases are widely considered as the most reliable information source for a particular gene/protein/disease, but it should also be made clear they may have widely varying contents, infrastructure, and quality. Quality is very important to evaluate because these databases may affect health decision-making, research, and clinical practice. The Human Variome Project (HVP) established a Working Group for Variant Database Quality Assessment. The basic principle was to develop a simple system that nevertheless provides a good overview of the quality of a database. The HVP quality evaluation criteria that resulted are divided into four main components: data quality, technical quality, accessibility, and timeliness. This report elaborates on the developed quality criteria and how implementation of the quality scheme can be achieved. Examples are provided for the current status of the quality items in two different databases, BTKbase, an LSDB, and ClinVar, a central archive of submissions about variants and their clinical significance. © 2016 WILEY PERIODICALS, INC.

  14. Breathing Life into Engineering: A Lesson Study Life Science Lesson

    Science.gov (United States)

    Lawrence, Maria; Yang, Li-Ling; Briggs, May; Hession, Alicia; Koussa, Anita; Wagoner, Lisa

    2016-01-01

    A fifth grade life science lesson was implemented through a lesson study approach in two fifth grade classrooms. The research lesson was designed by a team of four elementary school teachers with the goal of emphasizing engineering practices consistent with the "Next Generation Science Standards" (NGSS) (Achieve Inc. 2013). The fifth…

  15. Data integration to prioritize drugs using genomics and curated data.

    Science.gov (United States)

    Louhimo, Riku; Laakso, Marko; Belitskin, Denis; Klefström, Juha; Lehtonen, Rainer; Hautaniemi, Sampsa

    2016-01-01

    Genomic alterations affecting drug target proteins occur in several tumor types and are prime candidates for patient-specific tailored treatments. Increasingly, patients likely to benefit from targeted cancer therapy are selected based on molecular alterations. The selection of a precision therapy benefiting most patients is challenging but can be enhanced with integration of multiple types of molecular data. Data integration approaches for drug prioritization have successfully integrated diverse molecular data but do not take full advantage of existing data and literature. We have built a knowledge-base which connects data from public databases with molecular results from over 2200 tumors, signaling pathways and drug-target databases. Moreover, we have developed a data mining algorithm to effectively utilize this heterogeneous knowledge-base. Our algorithm is designed to facilitate retargeting of existing drugs by stratifying samples and prioritizing drug targets. We analyzed 797 primary tumors from The Cancer Genome Atlas breast and ovarian cancer cohorts using our framework. FGFR, CDK and HER2 inhibitors were prioritized in breast and ovarian data sets. Estrogen receptor positive breast tumors showed potential sensitivity to targeted inhibitors of FGFR due to activation of FGFR3. Our results suggest that computational sample stratification selects potentially sensitive samples for targeted therapies and can aid in precision medicine drug repositioning. Source code is available from http://csblcanges.fimm.fi/GOPredict/.

  16. Curative efficacy and safety of traditional Chinese medicine xuebijing injections combined with ulinastatin for treating sepsis in the Chinese population: A meta-analysis.

    Science.gov (United States)

    Xiao, Shi-Hui; Luo, Liang; Liu, Xiang-Hong; Zhou, Yu-Ming; Liu, Hong-Ming; Huang, Zhen-Fei

    2018-06-01

    Sepsis is a clinically critical disease. However, it is still controversial whether the combined use of traditional Chinese medicine Xuebijing injections (XBJI) and western medicine can enhance curative efficacy and ensure safety compared with western medicine alone. Thus, this research consisted of a systematic review of the curative efficacy and safety of traditional Chinese medicine XBJI combined with ulinastatin for treating sepsis in the Chinese population. A total of 8 databases were retrieved: 4 foreign databases, namely, PubMed, The Cochrane Library, Embase, and Web of Science; and 4 Chinese databases, namely, Sino Med, China National Knowledge Infrastructure (CNKI), VIP, and Wangfang Data. The time span of retrieval began from the establishment of each database and ended on August 1, 2017. Published randomized controlled trials about the combined use of traditional Chinese medicine XBJI and western medicine were included, regardless of language. Stata12.0 software was used for statistical analysis. Finally, 16 papers involving 1335 cases were included. The result of meta-analysis showed that compared with the single use of ulinastatin, traditional Chinese medicine XBJI combined with ulinastatin could reduce the time of mechanical ventilation, shorten the length of intensive care unit (ICU) stay, improve the 28-day survival rate, and decrease the occurrence rate of multiple organ dysfunction syndrome, case fatality rate, procalcitonin (PCT) content, APACKEII score, tumor necrosis factor (TNF)-α level, and interleukin (IL)-6 level. On the basis of the common basic therapeutic regimen, the combined use of traditional Chinese medicine XBJI and ulinastatin was compared with the use of ulinastatin alone for treating sepsis in the Chinese population. It was found that the number of adverse events of combination therapy is not significantly increased, and its clinical safety is well within the permitted range. However, considering the limitations of this

  17. Lessons for Teaching Art Criticism.

    Science.gov (United States)

    Barrett, Terry, Ed.; Clark, Gilbert, Ed.

    This collection of lessons is meant to be a practical guide to help teachers engage children in art criticism. The lessons generally follow a similar format. Most suggest an age group but may be modified for use with younger or older students. Several authors suggest variations and extensions for lessons that include studio activities. A broad…

  18. Lesson Planning the Kodaly Way.

    Science.gov (United States)

    Boshkoff, Ruth

    1991-01-01

    Discusses the contribution of Zoltan Kodaly to music lesson planning. Emphasizes preparation, presentation, and practice as the three important strategies in teaching concepts and skills to be included in a lesson plan. Includes a sample lesson plan covering a semester and advice on choosing song material. (DK)

  19. Building a genome database using an object-oriented approach.

    Science.gov (United States)

    Barbasiewicz, Anna; Liu, Lin; Lang, B Franz; Burger, Gertraud

    2002-01-01

    GOBASE is a relational database that integrates data associated with mitochondria and chloroplasts. The most important data in GOBASE, i. e., molecular sequences and taxonomic information, are obtained from the public sequence data repository at the National Center for Biotechnology Information (NCBI), and are validated by our experts. Maintaining a curated genomic database comes with a towering labor cost, due to the shear volume of available genomic sequences and the plethora of annotation errors and omissions in records retrieved from public repositories. Here we describe our approach to increase automation of the database population process, thereby reducing manual intervention. As a first step, we used Unified Modeling Language (UML) to construct a list of potential errors. Each case was evaluated independently, and an expert solution was devised, and represented as a diagram. Subsequently, the UML diagrams were used as templates for writing object-oriented automation programs in the Java programming language.

  20. Preventive but Not Curative Efficacy of Celecoxib on Bladder Carcinogenesis in a Rat Model

    Directory of Open Access Journals (Sweden)

    José Sereno

    2010-01-01

    Full Text Available To evaluate the effect of a cyclooxygenase 2 inhibitor, celecoxib (CEL, on bladder cancer inhibition in a rat model, when used as preventive versus as curative treatment. The study comprised 52 male Wistar rats, divided in 5 groups, during a 20-week protocol: control: vehicle, carcinogen: 0.05% of N-butyl-N-(4-hydroxybutyl nitrosamine (BBN, CEL: 10 mg/kg/day of the selective COX-2 inhibitor Celebrex, preventive CEL (CEL+BBN-P, and curative CEL (BBN+CEL-C groups. Although tumor growth was markedly inhibited by the preventive application of CEL, it was even aggravated by the curative treatment. The incidence of gross bladder carcinoma was: control 0/8(0%, BBN 13/20(65%, CEL 0/8(0%, CEL+BBN-P 1/8(12.5%, and BBN+CEL-C 6/8(75%. The number and volume of carcinomas were significantly lower in the CEL+BBN-P versus BBN, accompanied by an ample reduction in hyperplasia, dysplasia, and papillary tumors as well as COX-2 immunostaining. In spite of the reduction of tumor volumes in the curative BBN+CEL-C group, tumor malignancy was augmented. An anti-inflammatory and antioxidant profile was encountered only in the group under preventive treatment. In conclusion, preventive, but not curative, celecoxib treatment promoted a striking inhibitory effect on bladder cancer development, reinforcing the potential role of chemopreventive strategies based on cyclooxygenase 2 inhibition.

  1. Determination of protective properties of Bardejovske Kuple spa curative waters by rotational viscometry and ABTS assay

    Directory of Open Access Journals (Sweden)

    TOPOLSKA Dominika

    2014-02-01

    Full Text Available Mineral waters from Bardejovske Kupele spa are natural, strongly mineralized, with healing effects. They are classified as hydrocarbonic - containing chloride and sodium - carbonic, cold, hypotonic, with a relatively high boric acid content. Potential anti-oxidative effects of curative waters from Bardejovske Kupele were investigated against the hyaluronan (HA degradation. High-molar-mass HA was exposed to the action of ascorbate and cupric ions, which initiate free-radical chain degradation. Time-dependent changes of dynamic viscosity (h of the HA solutions were monitored by rotational viscometry. The radical scavenging capacity of curative waters was determined by the ABTS assay. Despite a significantly high content of transition metal ions, especially iron, remarkable protective effects of the two curative spa waters were found, namely Alzbeta and Klara. Even though “Alzbeta´s“ iron content was 3.5-fold higher than “Klara´s“, “Alzbeta“ was shown to have better protective properties against the HA degradation compared to “Klara“. Bolus addition of ferric ions to the reaction system instead of the natural iron-containing curative water caused a significant HA degradation. The ABTS decolorization assay revealed that the curative spa waters were proven as poorly effective scavengers of the ABTS·+ cation radical.

  2. Content curation en periodismo (y en documentación periodística

    Directory of Open Access Journals (Sweden)

    Javier Guallar

    2014-05-01

    Full Text Available Desde hace unos pocos años están con nosotros los conceptos de content curation y de content curator como actividad o sistema y como profesional o especialista, respectivamente. Aunque el origen del término se vincule al mundo del marketing –se considera como artículo fundacional “Manifesto for the content curator” del marketer Rohit Bhargava (2009-, y sus características se identifiquen en buena medida con las del profesional de la Información y la Documentación, la content curation va más allá de una disciplina concreta o de un rol profesional determinado. Ambos perfiles citados, marketers y bibliotecarios-documentalistas, pueden hacer content curation por supuesto, pero también otros, como educadores, por ejemplo. Y, como es el caso que tratamos aquí, periodistas (y documentalistas de prensa. Porque en estas breves líneas nos vamos a centrar en presentar el papel de la content curation en el ámbito del periodismo.

  3. Organic Contamination Baseline Study in NASA Johnson Space Center Astromaterials Curation Laboratories

    Science.gov (United States)

    Calaway, Michael J.; Allen, Carlton C.; Allton, Judith H.

    2014-01-01

    Future robotic and human spaceflight missions to the Moon, Mars, asteroids, and comets will require curating astromaterial samples with minimal inorganic and organic contamination to preserve the scientific integrity of each sample. 21st century sample return missions will focus on strict protocols for reducing organic contamination that have not been seen since the Apollo manned lunar landing program. To properly curate these materials, the Astromaterials Acquisition and Curation Office under the Astromaterial Research and Exploration Science Directorate at NASA Johnson Space Center houses and protects all extraterrestrial materials brought back to Earth that are controlled by the United States government. During fiscal year 2012, we conducted a year-long project to compile historical documentation and laboratory tests involving organic investigations at these facilities. In addition, we developed a plan to determine the current state of organic cleanliness in curation laboratories housing astromaterials. This was accomplished by focusing on current procedures and protocols for cleaning, sample handling, and storage. While the intention of this report is to give a comprehensive overview of the current state of organic cleanliness in JSC curation laboratories, it also provides a baseline for determining whether our cleaning procedures and sample handling protocols need to be adapted and/or augmented to meet the new requirements for future human spaceflight and robotic sample return missions.

  4. The DigCurV Curriculum Framework for Digital Curation in the Cultural Heritage Sector

    Directory of Open Access Journals (Sweden)

    Laura Molloy

    2014-07-01

    Full Text Available In 2013, the DigCurV collaborative network completed development of a Curriculum Framework for digital curation skills in the European cultural heritage sector. DigCurV synthesised a variety of established skills and competence models in the digital curation and LIS sectors with expertise from digital curation professionals, in order to develop a new Curriculum Framework. The resulting Framework provides a common language and helps define the skills, knowledge and abilities that are necessary for the development of digital curation training; for benchmarking existing programmes; and for promoting the continuing production, improvement and refinement of digital curation training programmes. This paper describes the salient points of this work, including how the project team conducted the research necessary to develop the Framework, the structure of the Framework, the processes used to validate the Framework, and three ‘lenses’ onto the Framework. The paper also provides suggestions as to how the Framework might be used, including a description of potential audiences and purposes.

  5. Toward an interactive article: integrating journals and biological databases

    Directory of Open Access Journals (Sweden)

    Marygold Steven J

    2011-05-01

    Full Text Available Abstract Background Journal articles and databases are two major modes of communication in the biological sciences, and thus integrating these critical resources is of urgent importance to increase the pace of discovery. Projects focused on bridging the gap between journals and databases have been on the rise over the last five years and have resulted in the development of automated tools that can recognize entities within a document and link those entities to a relevant database. Unfortunately, automated tools cannot resolve ambiguities that arise from one term being used to signify entities that are quite distinct from one another. Instead, resolving these ambiguities requires some manual oversight. Finding the right balance between the speed and portability of automation and the accuracy and flexibility of manual effort is a crucial goal to making text markup a successful venture. Results We have established a journal article mark-up pipeline that links GENETICS journal articles and the model organism database (MOD WormBase. This pipeline uses a lexicon built with entities from the database as a first step. The entity markup pipeline results in links from over nine classes of objects including genes, proteins, alleles, phenotypes and anatomical terms. New entities and ambiguities are discovered and resolved by a database curator through a manual quality control (QC step, along with help from authors via a web form that is provided to them by the journal. New entities discovered through this pipeline are immediately sent to an appropriate curator at the database. Ambiguous entities that do not automatically resolve to one link are resolved by hand ensuring an accurate link. This pipeline has been extended to other databases, namely Saccharomyces Genome Database (SGD and FlyBase, and has been implemented in marking up a paper with links to multiple databases. Conclusions Our semi-automated pipeline hyperlinks articles published in GENETICS to

  6. Locus-Specific Databases and Recommendations to Strengthen Their Contribution to the Classification of Variants in Cancer Susceptibility Genes

    NARCIS (Netherlands)

    Greenblatt, Marc S.; Brody, Lawrence C.; Foulkes, William D.; Genuardi, Maurizio; Hofstra, Robert M. W.; Olivier, Magali; Plon, Sharon E.; Sijmons, Rolf H.; Sinilnikova, Olga; Spurdle, Amanda B.

    2008-01-01

    Locus-specific databases (LSDBs) are curated collections of sequence variants in genes associated with disease. LSDBs of cancer-related genes often serve as a critical resource to researchers, diagnostic laboratories, clinicians, and others in the cancer genetics community. LSDBs are poised to play

  7. Database Description - PSCDB | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available abase Description General information of database Database name PSCDB Alternative n...rial Science and Technology (AIST) Takayuki Amemiya E-mail: Database classification Structure Databases - Protein structure Database...554-D558. External Links: Original website information Database maintenance site Graduate School of Informat...available URL of Web services - Need for user registration Not available About This Database Database Descri...ption Download License Update History of This Database Site Policy | Contact Us Database Description - PSCDB | LSDB Archive ...

  8. Directory of IAEA databases

    International Nuclear Information System (INIS)

    1991-11-01

    The first edition of the Directory of IAEA Databases is intended to describe the computerized information sources available to IAEA staff members. It contains a listing of all databases produced at the IAEA, together with information on their availability

  9. Native Health Research Database

    Science.gov (United States)

    ... Indian Health Board) Welcome to the Native Health Database. Please enter your search terms. Basic Search Advanced ... To learn more about searching the Native Health Database, click here. Tutorial Video The NHD has made ...

  10. Cell Centred Database (CCDB)

    Data.gov (United States)

    U.S. Department of Health & Human Services — The Cell Centered Database (CCDB) is a web accessible database for high resolution 2D, 3D and 4D data from light and electron microscopy, including correlated imaging.

  11. E3 Staff Database

    Data.gov (United States)

    US Agency for International Development — E3 Staff database is maintained by E3 PDMS (Professional Development & Management Services) office. The database is Mysql. It is manually updated by E3 staff as...

  12. Recovery, Transportation and Acceptance to the Curation Facility of the Hayabusa Re-Entry Capsule

    Science.gov (United States)

    Abe, M.; Fujimura, A.; Yano, H.; Okamoto, C.; Okada, T.; Yada, T.; Ishibashi, Y.; Shirai, K.; Nakamura, T.; Noguchi, T.; hide

    2011-01-01

    The "Hayabusa" re-entry capsule was safely carried into the clean room of Sagamihara Planetary Sample Curation Facility in JAXA on June 18, 2010. After executing computed tomographic (CT) scanning, removal of heat shield, and surface cleaning of sample container, the sample container was enclosed into the clean chamber. After opening the sample container and residual gas sampling in the clean chamber, optical observation, sample recovery, sample separation for initial analysis will be performed. This curation work is continuing for several manths with some selected member of Hayabusa Asteroidal Sample Preliminary Examination Team (HASPET). We report here on the 'Hayabusa' capsule recovery operation, and transportation and acceptance at the curation facility of the Hayabusa re-entry capsule.

  13. State of the Art of Cost and Benefit Models for Digital Curation

    DEFF Research Database (Denmark)

    Kejser, Ulla Bøgvad; Davidson, Joy; Wang, David

    2014-01-01

    , to support decision-making and for selecting the most efficient processes – all of which are critical for ensuring sustainability of digital curation investment. The evaluation revealed that the most prominent challenges are associated with the models’ usability, their inability to model quality and benefits......This paper presents the results of an evaluation carried out by the EU 4C project to assess how well current digital curation cost and benefit models meet a range of stakeholders’ needs. This work aims to elicit a means of modelling that enables comparing financial information across organisations...... of curation, and the lack of a clear terminology and conceptual description of costs and benefits. The paper provides recommendations on how these gaps in cost and benefit modelling can be bridged....

  14. Controlling Nutritional Status (CONUT) score is a prognostic marker for gastric cancer patients after curative resection.

    Science.gov (United States)

    Kuroda, Daisuke; Sawayama, Hiroshi; Kurashige, Junji; Iwatsuki, Masaaki; Eto, Tsugio; Tokunaga, Ryuma; Kitano, Yuki; Yamamura, Kensuke; Ouchi, Mayuko; Nakamura, Kenichi; Baba, Yoshifumi; Sakamoto, Yasuo; Yamashita, Yoichi; Yoshida, Naoya; Chikamoto, Akira; Baba, Hideo

    2018-03-01

    Controlling Nutritional Status (CONUT), as calculated from serum albumin, total cholesterol concentration, and total lymphocyte count, was previously shown to be useful for nutritional assessment. The current study investigated the potential use of CONUT as a prognostic marker in gastric cancer patients after curative resection. Preoperative CONUT was retrospectively calculated in 416 gastric cancer patients who underwent curative resection at Kumamoto University Hospital from 2005 to 2014. The patients were divided into two groups: CONUT-high (≥4) and CONUT-low (≤3), according to time-dependent receiver operating characteristic (ROC) analysis. The associations of CONUT with clinicopathological factors and survival were evaluated. CONUT-high patients were significantly older (p nutritional status but also for predicting long-term OS in gastric cancer patients after curative resection.

  15. A systematic review of nonrandomized controlled trials on the curative effects of aquatic exercise

    Directory of Open Access Journals (Sweden)

    Kamioka H

    2011-03-01

    Full Text Available Hiroharu Kamioka1, Kiichiro Tsutani2, Yoshiteru Mutoh3, Hiroyasu Okuizum4, Miho Ohta5, Shuichi Handa4, Shinpei Okada6, Jun Kitayuguchi7, Masamitsu Kamada7, Nobuyoshi Shiozawa8, Sang-Jun Park4, Takuya Honda4, Shoko Moriyama41Faculty of Regional Environment Science, Tokyo University of Agriculture, Tokyo, Japan; 2Department of Drug Policy and Management, Graduate School of Pharmaceutical Sciences, 3Department of Physical and Health Education, Graduate School of Education, The University of Tokyo, Tokyo, Japan; 4Mimaki Onsen (Spa Clinic, Tomi City, Japan; 5Laboratory of Aqua, Health, and Sports Medicine, 6Physical Education and Medicine Research Foundation, Nagano, Japan; 7Physical Education and Medicine Research Center Unnan, Unnan City, Japan; 8Department of Longevity and Social Medicine, Okayama University Graduate School of Medicine, Dentistry and Pharmaceutical Sciences, Okayama, JapanBackground: The objectives of this review were to integrate the evidence of curative effects through aquatic exercise and assess the quality of studies based on a review of nonrandomized controlled trials (nRCTs.Methods: Study design was a systematic review of nonrandomized controlled trials. Trials were eligible if they were nonrandomized clinical trials. Studies included one treatment group in which aquatic exercise was applied. We searched the following databases from 2000 up to July 20, 2009: MEDLINE via PubMed, CINAHL, and Ichushi-Web.Results: Twenty-one trials met all inclusion criteria. Languages included were English (N = 9, Japanese (N = 11, and Korean (N = 1. Target diseases were knee and/or hip osteoarthritis, poliomyelitis, chronic kidney disease, discomforts of pregnancy, cardiovascular diseases, and rotator cuff tears. Many studies on nonspecific disease (healthy participants were included. All studies reported significant effectiveness in at least one or more outcomes. However results of evaluations with the TREND and CLEAR-NPT checklists generally

  16. NIRS database of the original research database

    International Nuclear Information System (INIS)

    Morita, Kyoko

    1991-01-01

    Recently, library staffs arranged and compiled the original research papers that have been written by researchers for 33 years since National Institute of Radiological Sciences (NIRS) established. This papers describes how the internal database of original research papers has been created. This is a small sample of hand-made database. This has been cumulating by staffs who have any knowledge about computer machine or computer programming. (author)

  17. Scopus database: a review.

    Science.gov (United States)

    Burnham, Judy F

    2006-03-08

    The Scopus database provides access to STM journal articles and the references included in those articles, allowing the searcher to search both forward and backward in time. The database can be used for collection development as well as for research. This review provides information on the key points of the database and compares it to Web of Science. Neither database is inclusive, but complements each other. If a library can only afford one, choice must be based in institutional needs.

  18. Aviation Safety Issues Database

    Science.gov (United States)

    Morello, Samuel A.; Ricks, Wendell R.

    2009-01-01

    The aviation safety issues database was instrumental in the refinement and substantiation of the National Aviation Safety Strategic Plan (NASSP). The issues database is a comprehensive set of issues from an extremely broad base of aviation functions, personnel, and vehicle categories, both nationally and internationally. Several aviation safety stakeholders such as the Commercial Aviation Safety Team (CAST) have already used the database. This broader interest was the genesis to making the database publically accessible and writing this report.

  19. Geocuration Lessons Learned from the Climate Data Initiative Project

    Science.gov (United States)

    Ramachandran, Rahul; Bugbee, Kaylin; Tilmes, Curt; Pinheiro Privette, Ana

    2015-01-01

    Curation is traditionally defined as the process of collecting and organizing information around a common subject matter or a topic of interest and typically occurs in museums, art galleries, and libraries. The task of organizing data around specific topics or themes is a vibrant and growing effort in the biological sciences but to date this effort has not been actively pursued in the Earth sciences. This presentation will introduce the concept of geocuration, which we define it as the act of searching, selecting, and synthesizing Earth science data/metadata and information from across disciplines and repositories into a single, cohesive, and useful compendium. We also present the Climate Data Initiative (CDI) project as an prototypical example. The CDI project is a systematic effort to manually curate and share openly available climate data from various federal agencies. CDI is a broad multi-agency effort of the U.S. government and seeks to leverage the extensive existing federal climate-relevant data to stimulate innovation and private-sector entrepreneurship to support national climate change preparedness. The geocuration process used in the CDI project, key lessons learned, and suggestions to improve similar geocuration efforts in the future will be part of this presentation.

  20. Geocuration Lessons Learned from the Climate Data Initiative Project

    Science.gov (United States)

    Ramachandran, R.; Bugbee, K.; Tilmes, C.; Privette, A. P.

    2015-12-01

    Curation is traditionally defined as the process of collecting and organizing information around a common subject matter or a topic of interest and typically occurs in museums, art galleries, and libraries. The task of organizing data around specific topics or themes is a vibrant and growing effort in the biological sciences but to date this effort has not been actively pursued in the Earth sciences. This presentation will introduce the concept of geocuration, which we define it as the act of searching, selecting, and synthesizing Earth science data/metadata and information from across disciplines and repositories into a single, cohesive, and useful compendium.We also present the Climate Data Initiative (CDI) project as an exemplar example. The CDI project is a systematic effort to manually curate and share openly available climate data from various federal agencies. CDI is a broad multi-agency effort of the U.S. government and seeks to leverage the extensive existing federal climate-relevant data to stimulate innovation and private-sector entrepreneurship to support national climate-change preparedness. The geocuration process used in CDI project, key lessons learned, and suggestions to improve similar geocuration efforts in the future will be part of this presentation.

  1. Hospital of Diagnosis Influences the Probability of Receiving Curative Treatment for Esophageal Cancer.

    Science.gov (United States)

    van Putten, Margreet; Koëter, Marijn; van Laarhoven, Hanneke W M; Lemmens, Valery E P P; Siersema, Peter D; Hulshof, Maarten C C M; Verhoeven, Rob H A; Nieuwenhuijzen, Grard A P

    2018-02-01

    The aim of this article was to study the influence of hospital of diagnosis on the probability of receiving curative treatment and its impact on survival among patients with esophageal cancer (EC). Although EC surgery is centralized in the Netherlands, the disease is often diagnosed in hospitals that do not perform this procedure. Patients with potentially curable esophageal or gastroesophageal junction tumors diagnosed between 2005 and 2013 who were potentially curable (cT1-3,X, any N, M0,X) were selected from the Netherlands Cancer Registry. Multilevel logistic regression was performed to examine the probability to undergo curative treatment (resection with or without neoadjuvant treatment, definitive chemoradiotherapy, or local tumor excision) according to hospital of diagnosis. Effects of variation in probability of undergoing curative treatment among these hospitals on survival were investigated by Cox regression. All 13,017 patients with potentially curable EC, diagnosed in 91 hospitals, were included. The proportion of patients receiving curative treatment ranged from 37% to 83% and from 45% to 86% in the periods 2005-2009 and 2010-2013, respectively, depending on hospital of diagnosis. After adjustment for patient- and hospital-related characteristics these proportions ranged from 41% to 77% and from 50% to 82%, respectively (both P < 0.001). Multivariable survival analyses showed that patients diagnosed in hospitals with a low probability of undergoing curative treatment had a worse overall survival (hazard ratio = 1.13, 95% confidence interval 1.06-1.20; hazard ratio = 1.15, 95% confidence interval 1.07-1.24). The variation in probability of undergoing potentially curative treatment for EC between hospitals of diagnosis and its impact on survival indicates that treatment decision making in EC may be improved.

  2. TRENDS: The aeronautical post-test database management system

    Science.gov (United States)

    Bjorkman, W. S.; Bondi, M. J.

    1990-01-01

    TRENDS, an engineering-test database operating system developed by NASA to support rotorcraft flight tests, is described. Capabilities and characteristics of the system are presented, with examples of its use in recalling and analyzing rotorcraft flight-test data from a TRENDS database. The importance of system user-friendliness in gaining users' acceptance is stressed, as is the importance of integrating supporting narrative data with numerical data in engineering-test databases. Considerations relevant to the creation and maintenance of flight-test database are discussed and TRENDS' solutions to database management problems are described. Requirements, constraints, and other considerations which led to the system's configuration are discussed and some of the lessons learned during TRENDS' development are presented. Potential applications of TRENDS to a wide range of aeronautical and other engineering tests are identified.

  3. Solving Guesstimation Problems Using the Semantic Web:Four Lessons from an Application

    OpenAIRE

    Bundy, Alan; Sasnauskas, Gintautas; Chan, Michael

    2013-01-01

    We draw on our experience of implementing a semi-automated guesstimation application of the Semantic Web, gort, to draw four lessons, which we claim are of general applicability. These are:1. Inference can unleash the Semantic Web: The full power of the web will only be realised when we can use it to infer new knowledge from old.2. The Semantic Web does not constrain the inference mechanisms: Since we must anyway curate the knowledge we extract from the web, we can take the opportunity to tra...

  4. Automated Oracle database testing

    CERN Multimedia

    CERN. Geneva

    2014-01-01

    Ensuring database stability and steady performance in the modern world of agile computing is a major challenge. Various changes happening at any level of the computing infrastructure: OS parameters & packages, kernel versions, database parameters & patches, or even schema changes, all can potentially harm production services. This presentation shows how an automatic and regular testing of Oracle databases can be achieved in such agile environment.

  5. Inleiding database-systemen

    NARCIS (Netherlands)

    Pels, H.J.; Lans, van der R.F.; Pels, H.J.; Meersman, R.A.

    1993-01-01

    Dit artikel introduceert de voornaamste begrippen die een rol spelen rond databases en het geeft een overzicht van de doelstellingen, de functies en de componenten van database-systemen. Hoewel de functie van een database intuitief vrij duidelijk is, is het toch een in technologisch opzicht complex

  6. In defense of commitment: The curative power of violated expectations.

    Science.gov (United States)

    Murray, Sandra L; Lamarche, Veronica M; Gomillion, Sarah; Seery, Mark D; Kondrak, Cheryl

    2017-11-01

    A new model of commitment defense in romantic relationships is proposed. It assumes that relationships afford a central resource for affirming meaning and purpose in the world. Consequently, violating expectations about the world outside the relationship can precipitate commitment defense inside the relationship. A meta-analysis of 5 experiments, 2 follow-up correlational studies, and a longitudinal study of the transition to first parenthood supported the model. Experimentally violating conventional expectations about the world (e.g., "hard work pays off") motivated less satisfied people to defensively affirm their commitment. Similarly, when becoming a parent naturalistically violated culturally conditioned gendered expectations about the division of household labor, less satisfied new mothers and fathers defensively affirmed their commitment from pre-to-post baby. The findings suggest that violating expected associations in the world outside the relationship motivates vulnerable people to set relationship their relationship right, thereby affirming expected associations in the relationship in the face of an unexpected world. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  7. Database Description - RMOS | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available base Description General information of database Database name RMOS Alternative nam...arch Unit Shoshi Kikuchi E-mail : Database classification Plant databases - Rice Microarray Data and other Gene Expression Database...s Organism Taxonomy Name: Oryza sativa Taxonomy ID: 4530 Database description The Ric...19&lang=en Whole data download - Referenced database Rice Expression Database (RED) Rice full-length cDNA Database... (KOME) Rice Genome Integrated Map Database (INE) Rice Mutant Panel Database (Tos17) Rice Genome Annotation Database

  8. The curative effect analysis of 131I-therapy on patients with Graves' disease

    International Nuclear Information System (INIS)

    Cui Qin; Lu Shujun; Lu Tianhe

    2002-01-01

    To investigate the curative effect of 131 I-therapy on Graves' disease, the authors analyse conditions of patients who have received 131 I-therapy (n -674). These results showed that the incidence of fully recover, improve, Graves' disease and invalid is 80.11%, 7.28%, 11.87% and 0.74% respectively. Therefore, 131 I-therapy on Graves' disease is convenient. It has little side effect, low cost and better curative effect, it is one of the best therapeutic methods to treat hyperthyroidism

  9. Observation of curative effect of 131I in treatment of hyperthyroidism

    International Nuclear Information System (INIS)

    Huang Kebin; Xu Fan; Zhang Yaping; Wang Jingchang; Zhao Mingli; Ye Ming

    2012-01-01

    Objective: To explore the curative effect of 131 I in the treatment of hyperthyroidism. Method: 126 patients with hyperthyroidism were treated with 131 I and the curative effect was analyzed. Result: The results showed that among 126 cases of hyperthyroidism treated with 131 I, 117 cases had recovered and the cure rate was 92.9%. 9 cases were found hypothyroidism in one-year follow-up and the occurrence rate was 7.1%. Conclusion: The treatment of hyperthyroidism with 131 I is safe and effective method. (authors)

  10. Jean-Paul Martinon, ed. - The Curatorial: A Philosophy of Curating

    Directory of Open Access Journals (Sweden)

    Sofia Romualdo

    2015-12-01

    Full Text Available In the words of Jean-Paul Martinon, this book’s editor, The Curatorial: A Philosophy of Curating originated from a “wish to talk about curating”, the same wish that led to the creation, in 2006, of a practice-led PhD programme at Goldsmiths College, called Curatorial/Knowledge. The anthology features contributions from tutors, guest speakers and students, all of whom delve into what “the curatorial” is and what it might mean in the future. Curating, or the act of organizing exhibitions, paral...

  11. Sharing and community curation of mass spectrometry data with Global Natural Products Social Molecular Networking

    DEFF Research Database (Denmark)

    Wang, Mingxun; Carver, Jeremy J.; Pevzner, Pavel

    2016-01-01

    are well-suited to high-throughput characterization of NP, there is a pressing need for an infrastructure to enable sharing and curation of data. We present Global Natural Products Social Molecular Networking (GNPS; http://gnps.ucsd.edu), an open-access knowledge base for community-wide organization...... and sharing of raw, processed or identified tandem mass (MS/MS) spectrometry data. In GNPS, crowdsourced curation of freely available community-wide reference MS libraries will underpin improved annotations. Data-driven social-networking should facilitate identification of spectra and foster collaborations...

  12. Perioperative and long-term outcome of intrahepatic cholangiocarcinoma involving the hepatic hilus after curative-intent resection: comparison with peripheral intrahepatic cholangiocarcinoma and hilar cholangiocarcinoma.

    Science.gov (United States)

    Zhang, Xu-Feng; Bagante, Fabio; Chen, Qinyu; Beal, Eliza W; Lv, Yi; Weiss, Matthew; Popescu, Irinel; Marques, Hugo P; Aldrighetti, Luca; Maithel, Shishir K; Pulitano, Carlo; Bauer, Todd W; Shen, Feng; Poultsides, George A; Soubrane, Olivier; Martel, Guillaume; Koerkamp, B Groot; Guglielmi, Alfredo; Itaru, Endo; Pawlik, Timothy M

    2018-05-01

    Intrahepatic cholangiocarcinoma with hepatic hilus involvement has been either classified as intrahepatic cholangiocarcinoma or hilar cholangiocarcinoma. The present study aimed to investigate the clinicopathologic characteristics and short- and long-term outcomes after curative resection for hilar type intrahepatic cholangiocarcinoma in comparison with peripheral intrahepatic cholangiocarcinoma and hilar cholangiocarcinoma. A total of 912 patients with mass-forming peripheral intrahepatic cholangiocarcinoma, 101 patients with hilar type intrahepatic cholangiocarcinoma, and 159 patients with hilar cholangiocarcinoma undergoing curative resection from 2000 to 2015 were included from two multi-institutional databases. Clinicopathologic characteristics and short- and long-term outcomes were compared among the 3 groups. Patients with hilar type intrahepatic cholangiocarcinoma had more aggressive tumor characteristics (eg, higher frequency of vascular invasion and lymph nodes metastasis) and experienced more extensive resections in comparison with either peripheral intrahepatic cholangiocarcinoma or hilar cholangiocarcinoma patients. The odds of lymphadenectomy and R0 resection rate among patients with hilar type intrahepatic cholangiocarcinoma were comparable with hilar cholangiocarcinoma patients, but higher than peripheral intrahepatic cholangiocarcinoma patients (lymphadenectomy incidence, 85.1% vs 42.5%, P hilar type intrahepatic cholangiocarcinoma experienced a higher rate of technical-related complications compared with peripheral intrahepatic cholangiocarcinoma patients. Of note, hilar type intrahepatic cholangiocarcinoma was associated with worse disease-specific survival and recurrence-free survival after curative resection versus peripheral intrahepatic cholangiocarcinoma (median disease-specific survival, 26.0 vs 54.0 months, P hilar cholangiocarcinoma (median disease-specific survival, 26.0 vs 49.0 months, P = .003; median recurrence-free survival

  13. The BDNYC database of low-mass stars, brown dwarfs, and planetary mass companions

    Science.gov (United States)

    Cruz, Kelle; Rodriguez, David; Filippazzo, Joseph; Gonzales, Eileen; Faherty, Jacqueline K.; Rice, Emily; BDNYC

    2018-01-01

    We present a web-interface to a database of low-mass stars, brown dwarfs, and planetary mass companions. Users can send SELECT SQL queries to the database, perform searches by coordinates or name, check the database inventory on specified objects, and even plot spectra interactively. The initial version of this database contains information for 198 objects and version 2 will contain over 1000 objects. The database currently includes photometric data from 2MASS, WISE, and Spitzer and version 2 will include a significant portion of the publicly available optical and NIR spectra for brown dwarfs. The database is maintained and curated by the BDNYC research group and we welcome contributions from other researchers via GitHub.

  14. An Interoperable Cartographic Database

    Directory of Open Access Journals (Sweden)

    Slobodanka Ključanin

    2007-05-01

    Full Text Available The concept of producing a prototype of interoperable cartographic database is explored in this paper, including the possibilities of integration of different geospatial data into the database management system and their visualization on the Internet. The implementation includes vectorization of the concept of a single map page, creation of the cartographic database in an object-relation database, spatial analysis, definition and visualization of the database content in the form of a map on the Internet. 

  15. Keyword Search in Databases

    CERN Document Server

    Yu, Jeffrey Xu; Chang, Lijun

    2009-01-01

    It has become highly desirable to provide users with flexible ways to query/search information over databases as simple as keyword search like Google search. This book surveys the recent developments on keyword search over databases, and focuses on finding structural information among objects in a database using a set of keywords. Such structural information to be returned can be either trees or subgraphs representing how the objects, that contain the required keywords, are interconnected in a relational database or in an XML database. The structural keyword search is completely different from

  16. Nuclear power economic database

    International Nuclear Information System (INIS)

    Ding Xiaoming; Li Lin; Zhao Shiping

    1996-01-01

    Nuclear power economic database (NPEDB), based on ORACLE V6.0, consists of three parts, i.e., economic data base of nuclear power station, economic data base of nuclear fuel cycle and economic database of nuclear power planning and nuclear environment. Economic database of nuclear power station includes data of general economics, technique, capital cost and benefit, etc. Economic database of nuclear fuel cycle includes data of technique and nuclear fuel price. Economic database of nuclear power planning and nuclear environment includes data of energy history, forecast, energy balance, electric power and energy facilities

  17. Brothers Grimm. [Lesson Plan].

    Science.gov (United States)

    Discovery Communications, Inc., Bethesda, MD.

    Based on Grimm's fairy tales, this lesson plan presents activities designed to help students understand that fairy tales connect them to earlier generations, help them think about present situations, that magic figures prominently in fairy tales, and that fairy tales can inspire readers to create original works of art. The main activity in the…

  18. DSCOVR Contamination Lessons Learned

    Science.gov (United States)

    Graziani, Larissa

    2015-01-01

    The Triana observatory was built at NASA GSFC in the late 1990's, then placed into storage. After approximately ten years it was removed from storage and repurposed as the Deep Space Climate Observatory (DSCOVR). This presentation outlines the contamination control program lessons learned during the integration, test and launch of DSCOVR.

  19. Phagocytosis: history's lessons.

    Science.gov (United States)

    Garg, Manish; Chandawarkar, Rajiv Y

    2013-01-01

    The assimilation of lessons from the past is an essential component of education for scientists of tomorrow. These lessons are not easy to find. History books on science are few and usually highly dramatized and biographies of scientists tend to exaggerate the pomp of scientific discovery. Both underplay the hard and laborious work that is integral to any scientific pursuit. Here we illustrate one such example. A century ago, the Nobel Prize in Medicine was awarded to two scientists: Ilya Metchnikoff, a Russian zoologist, for the discovery ofphagocytosis-a cell-mediated ingestion ofmicrobes; and Paul Ehrlich, a distinguished physician-scientist, for discovering a highly antigen-specific serum-derived antibody-based immune defense. These two diametrically opposing views of the host-pathogen interaction set the stage for a strife that led to seminal advancements in immunology. Mirrored in this journey are important lessons for scientists today--ubiquitously as applicable to modern scientific life as they were a century ago. This commentaryhighlights these lessons--a fitting centenary to a well-deserved recognition.

  20. Recycling Lesson Plan

    Science.gov (United States)

    Okaz, Abeer Ali

    2013-01-01

    This lesson plan designed for grade 2 students has the goal of teaching students about the environmental practice of recycling. Children will learn language words related to recycling such as: "we can recycle"/"we can't recycle" and how to avoid littering with such words as: "recycle paper" and/or "don't throw…

  1. Smart Consumer Lesson Plans.

    Science.gov (United States)

    New Jersey Consortium for Consumer Education, Newark.

    Lesson plans are provided for use with different populations of pre-K through senior high school students in four different areas of consumer education. Eight units in advertising are included: A First Look at Ads (pre-K-Grade 3), Don't Judge a Book by Its Cover (Grades 1-3), Fatal Distraction (Junior High), Package Labeling (Junior High), Product…

  2. CARFMAP: A Curated Pathway Map of Cardiac Fibroblasts.

    Directory of Open Access Journals (Sweden)

    Hieu T Nim

    Full Text Available The adult mammalian heart contains multiple cell types that work in unison under tightly regulated conditions to maintain homeostasis. Cardiac fibroblasts are a significant and unique population of non-muscle cells in the heart that have recently gained substantial interest in the cardiac biology community. To better understand this renaissance cell, it is essential to systematically survey what has been known in the literature about the cellular and molecular processes involved. We have built CARFMAP (http://visionet.erc.monash.edu.au/CARFMAP, an interactive cardiac fibroblast pathway map derived from the biomedical literature using a software-assisted manual data collection approach. CARFMAP is an information-rich interactive tool that enables cardiac biologists to explore the large body of literature in various creative ways. There is surprisingly little overlap between the cardiac fibroblast pathway map, a foreskin fibroblast pathway map, and a whole mouse organism signalling pathway map from the REACTOME database. Among the use cases of CARFMAP is a common task in our cardiac biology laboratory of identifying new genes that are (1 relevant to cardiac literature, and (2 differentially regulated in high-throughput assays. From the expression profiles of mouse cardiac and tail fibroblasts, we employed CARFMAP to characterise cardiac fibroblast pathways. Using CARFMAP in conjunction with transcriptomic data, we generated a stringent list of six genes that would not have been singled out using bioinformatics analyses alone. Experimental validation showed that five genes (Mmp3, Il6, Edn1, Pdgfc and Fgf10 are differentially regulated in the cardiac fibroblast. CARFMAP is a powerful tool for systems analyses of cardiac fibroblasts, facilitating systems-level cardiovascular research.

  3. CARFMAP: A Curated Pathway Map of Cardiac Fibroblasts.

    Science.gov (United States)

    Nim, Hieu T; Furtado, Milena B; Costa, Mauro W; Kitano, Hiroaki; Rosenthal, Nadia A; Boyd, Sarah E

    2015-01-01

    The adult mammalian heart contains multiple cell types that work in unison under tightly regulated conditions to maintain homeostasis. Cardiac fibroblasts are a significant and unique population of non-muscle cells in the heart that have recently gained substantial interest in the cardiac biology community. To better understand this renaissance cell, it is essential to systematically survey what has been known in the literature about the cellular and molecular processes involved. We have built CARFMAP (http://visionet.erc.monash.edu.au/CARFMAP), an interactive cardiac fibroblast pathway map derived from the biomedical literature using a software-assisted manual data collection approach. CARFMAP is an information-rich interactive tool that enables cardiac biologists to explore the large body of literature in various creative ways. There is surprisingly little overlap between the cardiac fibroblast pathway map, a foreskin fibroblast pathway map, and a whole mouse organism signalling pathway map from the REACTOME database. Among the use cases of CARFMAP is a common task in our cardiac biology laboratory of identifying new genes that are (1) relevant to cardiac literature, and (2) differentially regulated in high-throughput assays. From the expression profiles of mouse cardiac and tail fibroblasts, we employed CARFMAP to characterise cardiac fibroblast pathways. Using CARFMAP in conjunction with transcriptomic data, we generated a stringent list of six genes that would not have been singled out using bioinformatics analyses alone. Experimental validation showed that five genes (Mmp3, Il6, Edn1, Pdgfc and Fgf10) are differentially regulated in the cardiac fibroblast. CARFMAP is a powerful tool for systems analyses of cardiac fibroblasts, facilitating systems-level cardiovascular research.

  4. Database Description - RPD | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available ase Description General information of database Database name RPD Alternative name Rice Proteome Database...titute of Crop Science, National Agriculture and Food Research Organization Setsuko Komatsu E-mail: Database... classification Proteomics Resources Plant databases - Rice Organism Taxonomy Name: Oryza sativa Taxonomy ID: 4530 Database... description Rice Proteome Database contains information on protei...and entered in the Rice Proteome Database. The database is searchable by keyword,

  5. Database Description - JSNP | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available base Description General information of database Database name JSNP Alternative nam...n Science and Technology Agency Creator Affiliation: Contact address E-mail : Database...sapiens Taxonomy ID: 9606 Database description A database of about 197,000 polymorphisms in Japanese populat...1):605-610 External Links: Original website information Database maintenance site Institute of Medical Scien...er registration Not available About This Database Database Description Download License Update History of This Database

  6. Database Description - ASTRA | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available abase Description General information of database Database name ASTRA Alternative n...tics Journal Search: Contact address Database classification Nucleotide Sequence Databases - Gene structure,...3702 Taxonomy Name: Oryza sativa Taxonomy ID: 4530 Database description The database represents classified p...(10):1211-6. External Links: Original website information Database maintenance site National Institute of Ad... for user registration Not available About This Database Database Description Dow

  7. Database Description - RED | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available ase Description General information of database Database name RED Alternative name Rice Expression Database...enome Research Unit Shoshi Kikuchi E-mail : Database classification Plant databases - Rice Database classifi...cation Microarray, Gene Expression Organism Taxonomy Name: Oryza sativa Taxonomy ID: 4530 Database descripti... Article title: Rice Expression Database: the gateway to rice functional genomics...nt Science (2002) Dec 7 (12):563-564 External Links: Original website information Database maintenance site

  8. Database Description - PLACE | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available abase Description General information of database Database name PLACE Alternative name A Database...Kannondai, Tsukuba, Ibaraki 305-8602, Japan National Institute of Agrobiological Sciences E-mail : Databas...e classification Plant databases Organism Taxonomy Name: Tracheophyta Taxonomy ID: 58023 Database...99, Vol.27, No.1 :297-300 External Links: Original website information Database maintenance site National In...- Need for user registration Not available About This Database Database Descripti

  9. Active and Social Data Curation: Reinventing the Business of Community-scale Lifecycle Data Management

    Science.gov (United States)

    McDonald, R. H.; Kumar, P.; Plale, B. A.; Myers, J.; Hedstrom, M. L.

    2012-12-01

    Effective long-term curation and preservation of data for community use has historically been limited to high-value and homogeneous collections produced by mission-oriented organizations. The technologies and practices that have been applied in these cases, e.g. relational data bases, development of comprehensive standardized vocabularies, and centralized support for reference data collections, are arguably applicable to the much broader range of data generated by the long tail of investigator-led research, with the logical conclusion of such an argument leading to the call for training, evangelism, and vastly increased funding as the best means of broadening community-scale data management. In this paper, we question this reasoning and explore how alternative approaches focused on the overall data lifecycle and the sociological and business realities of distributed multi-disciplinary research communities might dramatically lower costs, increase value, and consequently drive dramatic advances in our ability to use and re-use data, and ultimately enable more rapid scientific advance. Specifically, we introduce the concepts of active and social curation as a means to decrease coordination costs, align costs and values for individual data producers and data consumers, and improve the immediacy of returns for data curation investments. Further, we describe the specific architecture and services for active and social curation that are being prototyped within the Sustainable Environment - Actionable Data (SEAD) project within NSF's DataNet network and discuss how they are motivated by the long-tail dynamics in the cross-disciplinary sustainability research community.

  10. Introverts and Extraverts Reflecting on the Experience of Parish Ministry: Conversation between Training Incumbents and Curates

    Science.gov (United States)

    Francis, Leslie J.; Smith, Greg

    2016-01-01

    Psychological type theory suggests that introverts and extraverts may approach Christian ministry somewhat differently. This theory was tested within the context of a residential workshop attended by 15 curates, 12 of whom were accompanied by their training incumbents. Twelve themes were identified within responses to the question, "What does…

  11. Data Stewardship: Environmental Data Curation and a Web-of-Repositories

    Directory of Open Access Journals (Sweden)

    Karen S. Baker

    2009-10-01

    Full Text Available Scientific researchers today frequently package measurements and associated metadata as digital datasets in anticipation of storage in data repositories. Through the lens of environmental data stewardship, we consider the data repository as an organizational element central to data curation. One aspect of non-commercial repositories, their distance-from-origin of the data, is explored in terms of near and remote categories. Three idealized repository types are distinguished – local, center, and archive - paralleling research, resource, and reference collection categories respectively. Repository type characteristics such as scope, structure, and goals are discussed. Repository similarities in terms of roles, activities and responsibilities are also examined. Data stewardship is related to care of research data and responsible scientific communication supported by an infrastructure that coordinates curation activities; data curation is defined as a set of repeated and repeatable activities focusing on tending data and creating data products within a particular arena. The concept of “sphere-of-context” is introduced as an aid to distinguishing repository types. Conceptualizing a “web-of-repositories” accommodates a variety of repository types and represents an ecologically inclusive approach to data curation.

  12. [Curative Effects of Hydroxyurea on the Patients with β-thalassaemia Intermadia].

    Science.gov (United States)

    Huang, Li; Yao, Hong-Xia

    2016-06-01

    To investigate the clinical features of β-thalassaemia intermediate (TI) patients and the curative effect and side reactions of hydroxyurea therapys. Twenty nine patients with TI were divided into hydroxyurea therapy group and no hydroxyurea therapy group; the curative effect and side reactions in 2 groups were compared; the situation of blood transfusion in the 2 groups was evaluated. In hydroxyurea therapy group, the hemoglobin level increased after treatment for 3 months; the reticulocyte percentage obviously decreased after treatment for 12 months; the serum ferritin had been maintained at a low level; while in no hydroxyurea therapy group, the levels of hemoglobin and reticulocytes were not significantly improved after treatment, the serum ferritin level gradually increased. In hydroxyurea therapy group, 12 cases were out of blood transfusion after treatment for 12 months, effective rate of treatment was 85.71%; while in no hydroxyurea therapy group, the blood transfusion dependency was not improved after treatment. No serious side reactions were found in all the hydroxyurea treated patients. The hydroxyurea shows a better curative effect on TI patients, no serious side reactions occur in all the patients treated with hydroxyurea, but the long-term curative effect and side reactions should be observed continuously.

  13. Curation-Based Network Marketing: Strategies for Network Growth and Electronic Word-of-Mouth Diffusion

    Science.gov (United States)

    Church, Earnie Mitchell, Jr.

    2013-01-01

    In the last couple of years, a new aspect of online social networking has emerged, in which the strength of social network connections is based not on social ties but mutually shared interests. This dissertation studies these "curation-based" online social networks (CBN) and their suitability for the diffusion of electronic word-of-mouth…

  14. Social Media Selves: College Students' Curation of Self and Others through Facebook

    Science.gov (United States)

    Kasch, David Michael

    2013-01-01

    This qualitative study used cyber-ethnography and grounded theory to explore the ways in which 35 undergraduate students crafted and refined self-presentations on the social network site Facebook. Findings included the identification of two unique forms of self-presentation that students enacted: a "curated self" and a "commodified…

  15. A theoretical and experimental study on solubility of curatives in rubbers

    NARCIS (Netherlands)

    Guo, R.; Talma, Auke; Datta, Rabin; Dierkes, Wilma K.; Noordermeer, Jacobus W.M.

    2009-01-01

    The current study provides indications for improving the properties of dissimilar rubber blends. As solubility is a key factor for the distribution and a driving force for the migration of curatives in dissimilar rubber blends, it is of importance to have precise data of their solubilities. The

  16. The curative effect of a second curettage in persistent trophoblastic disease: a retrospective cohort survey.

    NARCIS (Netherlands)

    Trommel, N.E. van; Massuger, L.F.A.G.; Verheijen, R.; Sweep, C.G.J.; Thomas, C.M.G.

    2005-01-01

    OBJECTIVE: To assess the curative effect of a second curettage in patients with low-risk Persistent Trophoblastic Disease (PTD) after molar pregnancy. METHODS: A retrospective cohort survey was performed on 2122 patients registered with the Dutch Central Registry for Hydatidiform Moles between 1987

  17. Secondary Malignancy As A Manifestation Of Late Toxicity Of Curative Treatment For Testicular Cancer

    International Nuclear Information System (INIS)

    Reckova, M.; Kakalejcik, M.; Beniak, J.; Boljesikova, E.

    2008-01-01

    The case presents the patient with a diagnosis of bladder carcinosarcoma. He was diagnosed 42 years after adjuvant middle abdominal and pelvic radiotherapy for testicular seminoma. We discuss the problem of late toxicity of oncology treatment in patients with potentially curative germ cell tumors of testes together with diagnosis and treatment of patients with bladder carcinoma and carcinosarcoma. (author)

  18. Anastomotic Leak Increases Distant Recurrence and Long-Term Mortality After Curative Resection for Colonic Cancer

    DEFF Research Database (Denmark)

    Krarup, Peter-Martin; Nordholm-Carstensen, Andreas; Jorgensen, Lars N

    2014-01-01

    OBJECTIVE: To investigate the impact of anastomotic leak (AL) on disease recurrence and long-term mortality in patients alive 120 days after curative resection for colonic cancer. BACKGROUND: There is no solid data as to whether AL after colonic cancer surgery increases the risk of disease...

  19. From idol to art : African 'objects with power': a challenge for missionaries, anthropologists and museum curators

    NARCIS (Netherlands)

    Leyten, H.

    2015-01-01

    After his theological studies, Harrie Leyten (1935) worked as a missionary in Ghana for ten years. He studied social anthropology at Oxford University and became Africa curator of the Tropenmuseum in Amsterdam in 1975, and taught at the University of Amsterdam and the Reinwardt Academy in Amsterdam.

  20. PathwayAccess: CellDesigner plugins for pathway databases.

    Science.gov (United States)

    Van Hemert, John L; Dickerson, Julie A

    2010-09-15

    CellDesigner provides a user-friendly interface for graphical biochemical pathway description. Many pathway databases are not directly exportable to CellDesigner models. PathwayAccess is an extensible suite of CellDesigner plugins, which connect CellDesigner directly to pathway databases using respective Java application programming interfaces. The process is streamlined for creating new PathwayAccess plugins for specific pathway databases. Three PathwayAccess plugins, MetNetAccess, BioCycAccess and ReactomeAccess, directly connect CellDesigner to the pathway databases MetNetDB, BioCyc and Reactome. PathwayAccess plugins enable CellDesigner users to expose pathway data to analytical CellDesigner functions, curate their pathway databases and visually integrate pathway data from different databases using standard Systems Biology Markup Language and Systems Biology Graphical Notation. Implemented in Java, PathwayAccess plugins run with CellDesigner version 4.0.1 and were tested on Ubuntu Linux, Windows XP and 7, and MacOSX. Source code, binaries, documentation and video walkthroughs are freely available at http://vrac.iastate.edu/~jlv.

  1. Comparative analysis of four active compounds of Baikal skullcap and its classical TCM prescriptions according to different clinical curative effects

    Directory of Open Access Journals (Sweden)

    Guang-Wei Zhu

    2017-05-01

    Full Text Available Objective: A sensitive HPLC-DAD detection method was established for the comparative analysis of the four active compounds (including baicalin, baicalein, wogonoside and wogonin of Baikal Skullcap and its classical TCM prescriptions according to different clinical curative effects. And analyze the relationship between compatibility of medicines, content and clinical curative effect.

  2. Hazard Analysis Database Report

    CERN Document Server

    Grams, W H

    2000-01-01

    The Hazard Analysis Database was developed in conjunction with the hazard analysis activities conducted in accordance with DOE-STD-3009-94, Preparation Guide for U S . Department of Energy Nonreactor Nuclear Facility Safety Analysis Reports, for HNF-SD-WM-SAR-067, Tank Farms Final Safety Analysis Report (FSAR). The FSAR is part of the approved Authorization Basis (AB) for the River Protection Project (RPP). This document describes, identifies, and defines the contents and structure of the Tank Farms FSAR Hazard Analysis Database and documents the configuration control changes made to the database. The Hazard Analysis Database contains the collection of information generated during the initial hazard evaluations and the subsequent hazard and accident analysis activities. The Hazard Analysis Database supports the preparation of Chapters 3 ,4 , and 5 of the Tank Farms FSAR and the Unreviewed Safety Question (USQ) process and consists of two major, interrelated data sets: (1) Hazard Analysis Database: Data from t...

  3. Database Optimizing Services

    Directory of Open Access Journals (Sweden)

    Adrian GHENCEA

    2010-12-01

    Full Text Available Almost every organization has at its centre a database. The database provides support for conducting different activities, whether it is production, sales and marketing or internal operations. Every day, a database is accessed for help in strategic decisions. The satisfaction therefore of such needs is entailed with a high quality security and availability. Those needs can be realised using a DBMS (Database Management System which is, in fact, software for a database. Technically speaking, it is software which uses a standard method of cataloguing, recovery, and running different data queries. DBMS manages the input data, organizes it, and provides ways of modifying or extracting the data by its users or other programs. Managing the database is an operation that requires periodical updates, optimizing and monitoring.

  4. National Database of Geriatrics

    DEFF Research Database (Denmark)

    Kannegaard, Pia Nimann; Vinding, Kirsten L; Hare-Bruun, Helle

    2016-01-01

    AIM OF DATABASE: The aim of the National Database of Geriatrics is to monitor the quality of interdisciplinary diagnostics and treatment of patients admitted to a geriatric hospital unit. STUDY POPULATION: The database population consists of patients who were admitted to a geriatric hospital unit....... Geriatric patients cannot be defined by specific diagnoses. A geriatric patient is typically a frail multimorbid elderly patient with decreasing functional ability and social challenges. The database includes 14-15,000 admissions per year, and the database completeness has been stable at 90% during the past......, percentage of discharges with a rehabilitation plan, and the part of cases where an interdisciplinary conference has taken place. Data are recorded by doctors, nurses, and therapists in a database and linked to the Danish National Patient Register. DESCRIPTIVE DATA: Descriptive patient-related data include...

  5. Greengenes: Chimera-checked 16S rRNA gene database and workbenchcompatible in ARB

    Energy Technology Data Exchange (ETDEWEB)

    DeSantis, T.Z.; Hugenholtz, P.; Larsen, N.; Rojas, M.; Brodie,E.L; Keller, K.; Huber, T.; Dalevi, D.; Hu, P.; Andersen, G.L.

    2006-02-01

    A 16S rRNA gene database (http://greengenes.lbl.gov) addresses limitations of public repositories by providing chimera-screening, standard alignments and taxonomic classification using multiple published taxonomies. It was revealed that incongruent taxonomic nomenclature exists among curators even at the phylum-level. Putative chimeras were identified in 3% of environmental sequences and 0.2% of records derived from isolates. Environmental sequences were classified into 100 phylum-level lineages within the Archaea and Bacteria.

  6. Tradeoffs in distributed databases

    OpenAIRE

    Juntunen, R. (Risto)

    2016-01-01

    Abstract In a distributed database data is spread throughout the network into separated nodes with different DBMS systems (Date, 2000). According to CAP-theorem three database properties — consistency, availability and partition tolerance cannot be achieved simultaneously in distributed database systems. Two of these properties can be achieved but not all three at the same time (Brewer, 2000). Since this theorem there has b...

  7. Specialist Bibliographic Databases

    OpenAIRE

    Gasparyan, Armen Yuri; Yessirkepov, Marlen; Voronov, Alexander A.; Trukhachev, Vladimir I.; Kostyukova, Elena I.; Gerasimov, Alexey N.; Kitas, George D.

    2016-01-01

    Specialist bibliographic databases offer essential online tools for researchers and authors who work on specific subjects and perform comprehensive and systematic syntheses of evidence. This article presents examples of the established specialist databases, which may be of interest to those engaged in multidisciplinary science communication. Access to most specialist databases is through subscription schemes and membership in professional associations. Several aggregators of information and d...

  8. Supply Chain Initiatives Database

    Energy Technology Data Exchange (ETDEWEB)

    None

    2012-11-01

    The Supply Chain Initiatives Database (SCID) presents innovative approaches to engaging industrial suppliers in efforts to save energy, increase productivity and improve environmental performance. This comprehensive and freely-accessible database was developed by the Institute for Industrial Productivity (IIP). IIP acknowledges Ecofys for their valuable contributions. The database contains case studies searchable according to the types of activities buyers are undertaking to motivate suppliers, target sector, organization leading the initiative, and program or partnership linkages.

  9. Database Description - SAHG | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available base Description General information of database Database name SAHG Alternative nam...h: Contact address Chie Motono Tel : +81-3-3599-8067 E-mail : Database classification Structure Databases - ...e databases - Protein properties Organism Taxonomy Name: Homo sapiens Taxonomy ID: 9606 Database description... Links: Original website information Database maintenance site The Molecular Profiling Research Center for D...stration Not available About This Database Database Description Download License Update History of This Database Site Policy | Contact Us Database Description - SAHG | LSDB Archive ...

  10. Intermodal Passenger Connectivity Database -

    Data.gov (United States)

    Department of Transportation — The Intermodal Passenger Connectivity Database (IPCD) is a nationwide data table of passenger transportation terminals, with data on the availability of connections...

  11. Transporter Classification Database (TCDB)

    Data.gov (United States)

    U.S. Department of Health & Human Services — The Transporter Classification Database details a comprehensive classification system for membrane transport proteins known as the Transporter Classification (TC)...

  12. Residency Allocation Database

    Data.gov (United States)

    Department of Veterans Affairs — The Residency Allocation Database is used to determine allocation of funds for residency programs offered by Veterans Affairs Medical Centers (VAMCs). Information...

  13. Smart Location Database - Service

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Smart Location Database (SLD) summarizes over 80 demographic, built environment, transit service, and destination accessibility attributes for every census block...

  14. Database principles programming performance

    CERN Document Server

    O'Neil, Patrick

    2014-01-01

    Database: Principles Programming Performance provides an introduction to the fundamental principles of database systems. This book focuses on database programming and the relationships between principles, programming, and performance.Organized into 10 chapters, this book begins with an overview of database design principles and presents a comprehensive introduction to the concepts used by a DBA. This text then provides grounding in many abstract concepts of the relational model. Other chapters introduce SQL, describing its capabilities and covering the statements and functions of the programmi

  15. Veterans Administration Databases

    Science.gov (United States)

    The Veterans Administration Information Resource Center provides database and informatics experts, customer service, expert advice, information products, and web technology to VA researchers and others.

  16. IVR EFP Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This database contains trip-level reports submitted by vessels participating in Exempted Fishery projects with IVR reporting requirements.

  17. Towards Sensor Database Systems

    DEFF Research Database (Denmark)

    Bonnet, Philippe; Gehrke, Johannes; Seshadri, Praveen

    2001-01-01

    . These systems lack flexibility because data is extracted in a predefined way; also, they do not scale to a large number of devices because large volumes of raw data are transferred regardless of the queries that are submitted. In our new concept of sensor database system, queries dictate which data is extracted...... from the sensors. In this paper, we define the concept of sensor databases mixing stored data represented as relations and sensor data represented as time series. Each long-running query formulated over a sensor database defines a persistent view, which is maintained during a given time interval. We...... also describe the design and implementation of the COUGAR sensor database system....

  18. Database Publication Practices

    DEFF Research Database (Denmark)

    Bernstein, P.A.; DeWitt, D.; Heuer, A.

    2005-01-01

    There has been a growing interest in improving the publication processes for database research papers. This panel reports on recent changes in those processes and presents an initial cut at historical data for the VLDB Journal and ACM Transactions on Database Systems.......There has been a growing interest in improving the publication processes for database research papers. This panel reports on recent changes in those processes and presents an initial cut at historical data for the VLDB Journal and ACM Transactions on Database Systems....

  19. Smart Location Database - Download

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Smart Location Database (SLD) summarizes over 80 demographic, built environment, transit service, and destination accessibility attributes for every census block...

  20. Database Description - RMG | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available ase Description General information of database Database name RMG Alternative name ...raki 305-8602, Japan National Institute of Agrobiological Sciences E-mail : Database... classification Nucleotide Sequence Databases Organism Taxonomy Name: Oryza sativa Japonica Group Taxonomy ID: 39947 Database...rnal: Mol Genet Genomics (2002) 268: 434–445 External Links: Original website information Database...available URL of Web services - Need for user registration Not available About This Database Database Descri

  1. Database Description - KOME | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available base Description General information of database Database name KOME Alternative nam... Sciences Plant Genome Research Unit Shoshi Kikuchi E-mail : Database classification Plant databases - Rice ...Organism Taxonomy Name: Oryza sativa Taxonomy ID: 4530 Database description Information about approximately ...Hayashizaki Y, Kikuchi S. Journal: PLoS One. 2007 Nov 28; 2(11):e1235. External Links: Original website information Database...OS) Rice mutant panel database (Tos17) A Database of Plant Cis-acting Regulatory

  2. Update History of This Database - Arabidopsis Phenome Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Arabidopsis Phenome Database Update History of This Database Date Update contents 2017/02/27 Arabidopsis Phenome Data...base English archive site is opened. - Arabidopsis Phenome Database (http://jphenom...e.info/?page_id=95) is opened. About This Database Database Description Download License Update History of This Database... Site Policy | Contact Us Update History of This Database - Arabidopsis Phenome Database | LSDB Archive ...

  3. Update History of This Database - SKIP Stemcell Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us SKIP Stemcell Database Update History of This Database Date Update contents 2017/03/13 SKIP Stemcell Database... English archive site is opened. 2013/03/29 SKIP Stemcell Database ( https://www.skip.med.k...eio.ac.jp/SKIPSearch/top?lang=en ) is opened. About This Database Database Description Download License Update History of This Databa...se Site Policy | Contact Us Update History of This Database - SKIP Stemcell Database | LSDB Archive ...

  4. Update History of This Database - Yeast Interacting Proteins Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Yeast Interacting Proteins Database Update History of This Database Date Update contents 201...0/03/29 Yeast Interacting Proteins Database English archive site is opened. 2000/12/4 Yeast Interacting Proteins Database...( http://itolab.cb.k.u-tokyo.ac.jp/Y2H/ ) is released. About This Database Database Description... Download License Update History of This Database Site Policy | Contact Us Update History of This Database... - Yeast Interacting Proteins Database | LSDB Archive ...

  5. Discovering New Global Climate Patterns: Curating a 21-Year High Temporal (Hourly) and Spatial (40km) Resolution Reanalysis Dataset

    Science.gov (United States)

    Hou, C. Y.; Dattore, R.; Peng, G. S.

    2014-12-01

    The National Center for Atmospheric Research's Global Climate Four-Dimensional Data Assimilation (CFDDA) Hourly 40km Reanalysis dataset is a dynamically downscaled dataset with high temporal and spatial resolution. The dataset contains three-dimensional hourly analyses in netCDF format for the global atmospheric state from 1985 to 2005 on a 40km horizontal grid (0.4°grid increment) with 28 vertical levels, providing good representation of local forcing and diurnal variation of processes in the planetary boundary layer. This project aimed to make the dataset publicly available, accessible, and usable in order to provide a unique resource to allow and promote studies of new climate characteristics. When the curation project started, it had been five years since the data files were generated. Also, although the Principal Investigator (PI) had generated a user document at the end of the project in 2009, the document had not been maintained. Furthermore, the PI had moved to a new institution, and the remaining team members were reassigned to other projects. These factors made data curation in the areas of verifying data quality, harvest metadata descriptions, documenting provenance information especially challenging. As a result, the project's curation process found that: Data curator's skill and knowledge helped make decisions, such as file format and structure and workflow documentation, that had significant, positive impact on the ease of the dataset's management and long term preservation. Use of data curation tools, such as the Data Curation Profiles Toolkit's guidelines, revealed important information for promoting the data's usability and enhancing preservation planning. Involving data curators during each stage of the data curation life cycle instead of at the end could improve the curation process' efficiency. Overall, the project showed that proper resources invested in the curation process would give datasets the best chance to fulfill their potential to

  6. Lessons from Goiania

    International Nuclear Information System (INIS)

    Nazari Alves, R.

    2000-01-01

    The lessons learned from the radiological accident of Goiania in 1987 derived from the observations from the Regulatory Agency which was in charge of the decontamination tasks may be consolidated into four classes: Preventive Actions, characterised as those that aim to minimise the probability of occurrence of a radiological accident; Minimisation of time between the moment of the accident occurrence and the beginning of intervention, in case a radiological accident does occur, despite all preventive measures; Intervention, which is correlated to the type of installation, its geographical location, the social classes involved and their contamination vectors; and Follow up, for which well established rules to allow continuing monitoring of the victims and rebuilding of homes are necessary. The greatest lesson of all was the need for integration of the professionals involved, from all organizations. (author)

  7. Download - Trypanosomes Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us Trypanosomes Database Download First of all, please read the license of this database. Data ...1.4 KB) Simple search and download Downlaod via FTP FTP server is sometimes jammed. If it is, access [here]. About This Database Data...base Description Download License Update History of This Database Site Policy | Contact Us Download - Trypanosomes Database | LSDB Archive ...

  8. Database design and database administration for a kindergarten

    OpenAIRE

    Vítek, Daniel

    2009-01-01

    The bachelor thesis deals with creation of database design for a standard kindergarten, installation of the designed database into the database system Oracle Database 10g Express Edition and demonstration of the administration tasks in this database system. The verification of the database was proved by a developed access application.

  9. GEAR: A database of Genomic Elements Associated with drug Resistance

    Science.gov (United States)

    Wang, Yin-Ying; Chen, Wei-Hua; Xiao, Pei-Pei; Xie, Wen-Bin; Luo, Qibin; Bork, Peer; Zhao, Xing-Ming

    2017-01-01

    Drug resistance is becoming a serious problem that leads to the failure of standard treatments, which is generally developed because of genetic mutations of certain molecules. Here, we present GEAR (A database of Genomic Elements Associated with drug Resistance) that aims to provide comprehensive information about genomic elements (including genes, single-nucleotide polymorphisms and microRNAs) that are responsible for drug resistance. Right now, GEAR contains 1631 associations between 201 human drugs and 758 genes, 106 associations between 29 human drugs and 66 miRNAs, and 44 associations between 17 human drugs and 22 SNPs. These relationships are firstly extracted from primary literature with text mining and then manually curated. The drug resistome deposited in GEAR provides insights into the genetic factors underlying drug resistance. In addition, new indications and potential drug combinations can be identified based on the resistome. The GEAR database can be freely accessed through http://gear.comp-sysbio.org. PMID:28294141

  10. Current Challenges in Development of a Database of Three-Dimensional Chemical Structures

    Science.gov (United States)

    Maeda, Miki H.

    2015-01-01

    We are developing a database named 3DMET, a three-dimensional structure database of natural metabolites. There are two major impediments to the creation of 3D chemical structures from a set of planar structure drawings: the limited accuracy of computer programs and insufficient human resources for manual curation. We have tested some 2D–3D converters to convert 2D structure files from external databases. These automatic conversion processes yielded an excessive number of improper conversions. To ascertain the quality of the conversions, we compared IUPAC Chemical Identifier and canonical SMILES notations before and after conversion. Structures whose notations correspond to each other were regarded as a correct conversion in our present work. We found that chiral inversion is the most serious factor during the improper conversion. In the current stage of our database construction, published books or articles have been resources for additions to our database. Chemicals are usually drawn as pictures on the paper. To save human resources, an optical structure reader was introduced. The program was quite useful but some particular errors were observed during our operation. We hope our trials for producing correct 3D structures will help other developers of chemical programs and curators of chemical databases. PMID:26075200

  11. Missing "Links" in Bioinformatics Education: Expanding Students' Conceptions of Bioinformatics Using a Biodiversity Database of Living and Fossil Reef Corals

    Science.gov (United States)

    Nehm, Ross H.; Budd, Ann F.

    2006-01-01

    NMITA is a reef coral biodiversity database that we use to introduce students to the expansive realm of bioinformatics beyond genetics. We introduce a series of lessons that have students use this database, thereby accessing real data that can be used to test hypotheses about biodiversity and evolution while targeting the "National Science …

  12. ExplorEnz: a MySQL database of the IUBMB enzyme nomenclature.

    Science.gov (United States)

    McDonald, Andrew G; Boyce, Sinéad; Moss, Gerard P; Dixon, Henry B F; Tipton, Keith F

    2007-07-27

    We describe the database ExplorEnz, which is the primary repository for EC numbers and enzyme data that are being curated on behalf of the IUBMB. The enzyme nomenclature is incorporated into many other resources, including the ExPASy-ENZYME, BRENDA and KEGG bioinformatics databases. The data, which are stored in a MySQL database, preserve the formatting of chemical and enzyme names. A simple, easy to use, web-based query interface is provided, along with an advanced search engine for more complex queries. The database is publicly available at http://www.enzyme-database.org. The data are available for download as SQL and XML files via FTP. ExplorEnz has powerful and flexible search capabilities and provides the scientific community with the most up-to-date version of the IUBMB Enzyme List.

  13. Directory of IAEA databases

    International Nuclear Information System (INIS)

    1992-12-01

    This second edition of the Directory of IAEA Databases has been prepared within the Division of Scientific and Technical Information (NESI). Its main objective is to describe the computerized information sources available to staff members. This directory contains all databases produced at the IAEA, including databases stored on the mainframe, LAN's and PC's. All IAEA Division Directors have been requested to register the existence of their databases with NESI. For the second edition database owners were requested to review the existing entries for their databases and answer four additional questions. The four additional questions concerned the type of database (e.g. Bibliographic, Text, Statistical etc.), the category of database (e.g. Administrative, Nuclear Data etc.), the available documentation and the type of media used for distribution. In the individual entries on the following pages the answers to the first two questions (type and category) is always listed, but the answers to the second two questions (documentation and media) is only listed when information has been made available

  14. HIV Structural Database

    Science.gov (United States)

    SRD 102 HIV Structural Database (Web, free access)   The HIV Protease Structural Database is an archive of experimentally determined 3-D structures of Human Immunodeficiency Virus 1 (HIV-1), Human Immunodeficiency Virus 2 (HIV-2) and Simian Immunodeficiency Virus (SIV) Proteases and their complexes with inhibitors or products of substrate cleavage.

  15. Balkan Vegetation Database

    NARCIS (Netherlands)

    Vassilev, Kiril; Pedashenko, Hristo; Alexandrova, Alexandra; Tashev, Alexandar; Ganeva, Anna; Gavrilova, Anna; Gradevska, Asya; Assenov, Assen; Vitkova, Antonina; Grigorov, Borislav; Gussev, Chavdar; Filipova, Eva; Aneva, Ina; Knollová, Ilona; Nikolov, Ivaylo; Georgiev, Georgi; Gogushev, Georgi; Tinchev, Georgi; Pachedjieva, Kalina; Koev, Koycho; Lyubenova, Mariyana; Dimitrov, Marius; Apostolova-Stoyanova, Nadezhda; Velev, Nikolay; Zhelev, Petar; Glogov, Plamen; Natcheva, Rayna; Tzonev, Rossen; Boch, Steffen; Hennekens, Stephan M.; Georgiev, Stoyan; Stoyanov, Stoyan; Karakiev, Todor; Kalníková, Veronika; Shivarov, Veselin; Russakova, Veska; Vulchev, Vladimir

    2016-01-01

    The Balkan Vegetation Database (BVD; GIVD ID: EU-00-019; http://www.givd.info/ID/EU-00- 019) is a regional database that consists of phytosociological relevés from different vegetation types from six countries on the Balkan Peninsula (Albania, Bosnia and Herzegovina, Bulgaria, Kosovo, Montenegro

  16. World Database of Happiness

    NARCIS (Netherlands)

    R. Veenhoven (Ruut)

    1995-01-01

    textabstractABSTRACT The World Database of Happiness is an ongoing register of research on subjective appreciation of life. Its purpose is to make the wealth of scattered findings accessible, and to create a basis for further meta-analytic studies. The database involves four sections:
    1.

  17. Dictionary as Database.

    Science.gov (United States)

    Painter, Derrick

    1996-01-01

    Discussion of dictionaries as databases focuses on the digitizing of The Oxford English dictionary (OED) and the use of Standard Generalized Mark-Up Language (SGML). Topics include the creation of a consortium to digitize the OED, document structure, relational databases, text forms, sequence, and discourse. (LRW)

  18. Fire test database

    International Nuclear Information System (INIS)

    Lee, J.A.

    1989-01-01

    This paper describes a project recently completed for EPRI by Impell. The purpose of the project was to develop a reference database of fire tests performed on non-typical fire rated assemblies. The database is designed for use by utility fire protection engineers to locate test reports for power plant fire rated assemblies. As utilities prepare to respond to Information Notice 88-04, the database will identify utilities, vendors or manufacturers who have specific fire test data. The database contains fire test report summaries for 729 tested configurations. For each summary, a contact is identified from whom a copy of the complete fire test report can be obtained. Five types of configurations are included: doors, dampers, seals, wraps and walls. The database is computerized. One version for IBM; one for Mac. Each database is accessed through user-friendly software which allows adding, deleting, browsing, etc. through the database. There are five major database files. One each for the five types of tested configurations. The contents of each provides significant information regarding the test method and the physical attributes of the tested configuration. 3 figs

  19. Children's Culture Database (CCD)

    DEFF Research Database (Denmark)

    Wanting, Birgit

    a Dialogue inspired database with documentation, network (individual and institutional profiles) and current news , paper presented at the research seminar: Electronic access to fiction, Copenhagen, November 11-13, 1996......a Dialogue inspired database with documentation, network (individual and institutional profiles) and current news , paper presented at the research seminar: Electronic access to fiction, Copenhagen, November 11-13, 1996...

  20. Atomic Spectra Database (ASD)

    Science.gov (United States)

    SRD 78 NIST Atomic Spectra Database (ASD) (Web, free access)   This database provides access and search capability for NIST critically evaluated data on atomic energy levels, wavelengths, and transition probabilities that are reasonably up-to-date. The NIST Atomic Spectroscopy Data Center has carried out these critical compilations.

  1. Consumer Product Category Database

    Science.gov (United States)

    The Chemical and Product Categories database (CPCat) catalogs the use of over 40,000 chemicals and their presence in different consumer products. The chemical use information is compiled from multiple sources while product information is gathered from publicly available Material Safety Data Sheets (MSDS). EPA researchers are evaluating the possibility of expanding the database with additional product and use information.

  2. Database in Artificial Intelligence.

    Science.gov (United States)

    Wilkinson, Julia

    1986-01-01

    Describes a specialist bibliographic database of literature in the field of artificial intelligence created by the Turing Institute (Glasgow, Scotland) using the BRS/Search information retrieval software. The subscription method for end-users--i.e., annual fee entitles user to unlimited access to database, document provision, and printed awareness…

  3. NoSQL database scaling

    OpenAIRE

    Žardin, Norbert

    2017-01-01

    NoSQL database scaling is a decision, where system resources or financial expenses are traded for database performance or other benefits. By scaling a database, database performance and resource usage might increase or decrease, such changes might have a negative impact on an application that uses the database. In this work it is analyzed how database scaling affect database resource usage and performance. As a results, calculations are acquired, using which database scaling types and differe...

  4. HEROD: a human ethnic and regional specific omics database.

    Science.gov (United States)

    Zeng, Xian; Tao, Lin; Zhang, Peng; Qin, Chu; Chen, Shangying; He, Weidong; Tan, Ying; Xia Liu, Hong; Yang, Sheng Yong; Chen, Zhe; Jiang, Yu Yang; Chen, Yu Zong

    2017-10-15

    Genetic and gene expression variations within and between populations and across geographical regions have substantial effects on the biological phenotypes, diseases, and therapeutic response. The development of precision medicines can be facilitated by the OMICS studies of the patients of specific ethnicity and geographic region. However, there is an inadequate facility for broadly and conveniently accessing the ethnic and regional specific OMICS data. Here, we introduced a new free database, HEROD, a human ethnic and regional specific OMICS database. Its first version contains the gene expression data of 53 070 patients of 169 diseases in seven ethnic populations from 193 cities/regions in 49 nations curated from the Gene Expression Omnibus (GEO), the ArrayExpress Archive of Functional Genomics Data (ArrayExpress), the Cancer Genome Atlas (TCGA) and the International Cancer Genome Consortium (ICGC). Geographic region information of curated patients was mainly manually extracted from referenced publications of each original study. These data can be accessed and downloaded via keyword search, World map search, and menu-bar search of disease name, the international classification of disease code, geographical region, location of sample collection, ethnic population, gender, age, sample source organ, patient type (patient or healthy), sample type (disease or normal tissue) and assay type on the web interface. The HEROD database is freely accessible at http://bidd2.nus.edu.sg/herod/index.php. The database and web interface are implemented in MySQL, PHP and HTML with all major browsers supported. phacyz@nus.edu.sg. © The Author (2017). Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  5. The LHCb configuration database

    CERN Document Server

    Abadie, L; Van Herwijnen, Eric; Jacobsson, R; Jost, B; Neufeld, N

    2005-01-01

    The aim of the LHCb configuration database is to store information about all the controllable devices of the detector. The experiment's control system (that uses PVSS ) will configure, start up and monitor the detector from the information in the configuration database. The database will contain devices with their properties, connectivity and hierarchy. The ability to store and rapidly retrieve huge amounts of data, and the navigability between devices are important requirements. We have collected use cases to ensure the completeness of the design. Using the entity relationship modelling technique we describe the use cases as classes with attributes and links. We designed the schema for the tables using relational diagrams. This methodology has been applied to the TFC (switches) and DAQ system. Other parts of the detector will follow later. The database has been implemented using Oracle to benefit from central CERN database support. The project also foresees the creation of tools to populate, maintain, and co...

  6. Outcomes of the 'Data Curation for Geobiology at Yellowstone National Park' Workshop

    Science.gov (United States)

    Thomer, A.; Palmer, C. L.; Fouke, B. W.; Rodman, A.; Choudhury, G. S.; Baker, K. S.; Asangba, A. E.; Wickett, K.; DiLauro, T.; Varvel, V.

    2013-12-01

    The continuing proliferation of geological and biological data generated at scientifically significant sites (such as hot springs, coral reefs, volcanic fields and other unique, data-rich locales) has created a clear need for the curation and active management of these data. However, there has been little exploration of what these curation processes and policies would entail. To that end, the Site-Based Data Curation (SBDC) project is developing a framework of guidelines and processes for the curation of research data generated at scientifically significant sites. A workshop was held in April 2013 at Yellowstone National Park (YNP) to gather input from scientists and stakeholders. Workshop participants included nine researchers actively conducting geobiology research at YNP, and seven YNP representatives, including permitting staff and information professionals from the YNP research library and archive. Researchers came from a range of research areas -- geology, molecular and microbial biology, ecology, environmental engineering, and science education. Through group discussions, breakout sessions and hands-on activities, we sought to generate policy recommendations and curation guidelines for the collection, representation, sharing and quality control of geobiological datasets. We report on key themes that emerged from workshop discussions, including: - participants' broad conceptions of the long-term usefulness, reusability and value of data. - the benefits of aggregating site-specific data in general, and geobiological data in particular. - the importance of capturing a dataset's originating context, and the potential usefulness of photographs as a reliable and easy way of documenting context. - researchers' and resource managers' overlapping priorities with regards to 'big picture' data collection and management in the long-term. Overall, we found that workshop participants were enthusiastic and optimistic about future collaboration and development of community

  7. PHI-base: a new interface and further additions for the multi-species pathogen–host interactions database

    Science.gov (United States)

    Urban, Martin; Cuzick, Alayne; Rutherford, Kim; Irvine, Alistair; Pedro, Helder; Pant, Rashmi; Sadanadan, Vidyendra; Khamari, Lokanath; Billal, Santoshkumar; Mohanty, Sagar; Hammond-Kosack, Kim E.

    2017-01-01

    The pathogen–host interactions database (PHI-base) is available at www.phi-base.org. PHI-base contains expertly curated molecular and biological information on genes proven to affect the outcome of pathogen–host interactions reported in peer reviewed research articles. In addition, literature that indicates specific gene alterations that did not affect the disease interaction phenotype are curated to provide complete datasets for comparative purposes. Viruses are not included. Here we describe a revised PHI-base Version 4 data platform with improved search, filtering and extended data display functions. A PHIB-BLAST search function is provided and a link to PHI-Canto, a tool for authors to directly curate their own published data into PHI-base. The new release of PHI-base Version 4.2 (October 2016) has an increased data content containing information from 2219 manually curated references. The data provide information on 4460 genes from 264 pathogens tested on 176 hosts in 8046 interactions. Prokaryotic and eukaryotic pathogens are represented in almost equal numbers. Host species belong ∼70% to plants and 30% to other species of medical and/or environmental importance. Additional data types included into PHI-base 4 are the direct targets of pathogen effector proteins in experimental and natural host organisms. The curation problems encountered and the future directions of the PHI-base project are briefly discussed. PMID:27915230

  8. PHI-base: a new interface and further additions for the multi-species pathogen-host interactions database.

    Science.gov (United States)

    Urban, Martin; Cuzick, Alayne; Rutherford, Kim; Irvine, Alistair; Pedro, Helder; Pant, Rashmi; Sadanadan, Vidyendra; Khamari, Lokanath; Billal, Santoshkumar; Mohanty, Sagar; Hammond-Kosack, Kim E

    2017-01-04

    The pathogen-host interactions database (PHI-base) is available at www.phi-base.org PHI-base contains expertly curated molecular and biological information on genes proven to affect the outcome of pathogen-host interactions reported in peer reviewed research articles. In addition, literature that indicates specific gene alterations that did not affect the disease interaction phenotype are curated to provide complete datasets for comparative purposes. Viruses are not included. Here we describe a revised PHI-base Version 4 data platform with improved search, filtering and extended data display functions. A PHIB-BLAST search function is provided and a link to PHI-Canto, a tool for authors to directly curate their own published data into PHI-base. The new release of PHI-base Version 4.2 (October 2016) has an increased data content containing information from 2219 manually curated references. The data provide information on 4460 genes from 264 pathogens tested on 176 hosts in 8046 interactions. Prokaryotic and eukaryotic pathogens are represented in almost equal numbers. Host species belong ∼70% to plants and 30% to other species of medical and/or environmental importance. Additional data types included into PHI-base 4 are the direct targets of pathogen effector proteins in experimental and natural host organisms. The curation problems encountered and the future directions of the PHI-base project are briefly discussed. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  9. Database Description - SKIP Stemcell Database | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us SKIP Stemcell Database Database Description General information of database Database name SKIP Stemcell Database...rsity Journal Search: Contact address http://www.skip.med.keio.ac.jp/en/contact/ Database classification Human Genes and Diseases Dat...abase classification Stemcell Article Organism Taxonomy Name: Homo sapiens Taxonomy ID: 9606 Database...ks: Original website information Database maintenance site Center for Medical Genetics, School of medicine, ...lable Web services Not available URL of Web services - Need for user registration Not available About This Database Database

  10. CEBS: a comprehensive annotated database of toxicological data

    Science.gov (United States)

    Lea, Isabel A.; Gong, Hui; Paleja, Anand; Rashid, Asif; Fostel, Jennifer

    2017-01-01

    The Chemical Effects in Biological Systems database (CEBS) is a comprehensive and unique toxicology resource that compiles individual and summary animal data from the National Toxicology Program (NTP) testing program and other depositors into a single electronic repository. CEBS has undergone significant updates in recent years and currently contains over 11 000 test articles (exposure agents) and over 8000 studies including all available NTP carcinogenicity, short-term toxicity and genetic toxicity studies. Study data provided to CEBS are manually curated, accessioned and subject to quality assurance review prior to release to ensure high quality. The CEBS database has two main components: data collection and data delivery. To accommodate the breadth of data produced by NTP, the CEBS data collection component is an integrated relational design that allows the flexibility to capture any type of electronic data (to date). The data delivery component of the database comprises a series of dedicated user interface tables containing pre-processed data that support each component of the user interface. The user interface has been updated to include a series of nine Guided Search tools that allow access to NTP summary and conclusion data and larger non-NTP datasets. The CEBS database can be accessed online at http://www.niehs.nih.gov/research/resources/databases/cebs/. PMID:27899660

  11. HIPdb: a database of experimentally validated HIV inhibiting peptides.

    Science.gov (United States)

    Qureshi, Abid; Thakur, Nishant; Kumar, Manoj

    2013-01-01

    Besides antiretroviral drugs, peptides have also demonstrated potential to inhibit the Human immunodeficiency virus (HIV). For example, T20 has been discovered to effectively block the HIV entry and was approved by the FDA as a novel anti-HIV peptide (AHP). We have collated all experimental information on AHPs at a single platform. HIPdb is a manually curated database of experimentally verified HIV inhibiting peptides targeting various steps or proteins involved in the life cycle of HIV e.g. fusion, integration, reverse transcription etc. This database provides experimental information of 981 peptides. These are of varying length obtained from natural as well as synthetic sources and tested on different cell lines. Important fields included are peptide sequence, length, source, target, cell line, inhibition/IC(50), assay and reference. The database provides user friendly browse, search, sort and filter options. It also contains useful services like BLAST and 'Map' for alignment with user provided sequences. In addition, predicted structure and physicochemical properties of the peptides are also included. HIPdb database is freely available at http://crdd.osdd.net/servers/hipdb. Comprehensive information of this database will be helpful in selecting/designing effective anti-HIV peptides. Thus it may prove a useful resource to researchers for peptide based therapeutics development.

  12. Kalium: a database of potassium channel toxins from scorpion venom.

    Science.gov (United States)

    Kuzmenkov, Alexey I; Krylov, Nikolay A; Chugunov, Anton O; Grishin, Eugene V; Vassilevski, Alexander A

    2016-01-01

    Kalium (http://kaliumdb.org/) is a manually curated database that accumulates data on potassium channel toxins purified from scorpion venom (KTx). This database is an open-access resource, and provides easy access to pages of other databases of interest, such as UniProt, PDB, NCBI Taxonomy Browser, and PubMed. General achievements of Kalium are a strict and easy regulation of KTx classification based on the unified nomenclature supported by researchers in the field, removal of peptides with partial sequence and entries supported by transcriptomic information only, classification of β-family toxins, and addition of a novel λ-family. Molecules presented in the database can be processed by the Clustal Omega server using a one-click option. Molecular masses of mature peptides are calculated and available activity data are compiled for all KTx. We believe that Kalium is not only of high interest to professional toxinologists, but also of general utility to the scientific community.Database URL:http://kaliumdb.org/. © The Author(s) 2016. Published by Oxford University Press.

  13. Initial analysis and curation plans for MUSES-C asteroidal samples

    Science.gov (United States)

    Yano, H.; Kushiro, I.; Fujiwara, A.

    In the MUSES-C mission, sample return of several hundred mg to several g in to- tal is expected from the surface of the S-type near Earth asteroid 1998SF36 in 2007. The MUSES-C samples are expected to be more similar to micrometeorites than large pieces of rocks. Also the initial analysis to characterize general aspects of returned samples can consume only 15 % of its total mass and must complete the whole anal- yses including the database building before international AO for detailed analyses opens in less than a year. Confident exercise of non-destructive, micro-analyses when- ever possible are thus vital for the SMUSES-C Asteroidal Sample Preliminary Exam- ination Team (MASPET)T, which will be formed by the ISAS MUSES-C team, the international partners from NASA and Australia and Sall-JapanT meteoritic scientists to be selected as outsourcing parts of the initial analyses. In 2000-2001, in the pur- pose to survey what kinds and levels of micro-analysis techniques in respective fields, from major elements and mineralogy to trace and isotopic elements and organics, are available in Japan at present, ISAS welcomed a total of 11 applications to the first round open competition for the MASPET candidates. The initial evaluation was made by multiple domestic peer reviews. Nine out of 11 were then provided two kinds of Sasteroid sample analogsT that no applicant knew what they were in advance by the Selection Committee (chair: I. Kushiro) in order to conduct proposed analysis with self-claimed amount of samples (100 mg max) in self-claimed duration (6 months max). The proponents must demonstrate how much their technical capabilities, ana- lytical precision, and usefulness of the derived results for subsequent detailed analyses are worth being included in the MASPET studies. After the completion of multiple, international peer reviews, the Selection Committee compiled evaluations and recom- mended the finalists of this round competition. However, it is also recognized

  14. Hazard Analysis Database Report

    Energy Technology Data Exchange (ETDEWEB)

    GAULT, G.W.

    1999-10-13

    The Hazard Analysis Database was developed in conjunction with the hazard analysis activities conducted in accordance with DOE-STD-3009-94, Preparation Guide for US Department of Energy Nonreactor Nuclear Facility Safety Analysis Reports, for the Tank Waste Remediation System (TWRS) Final Safety Analysis Report (FSAR). The FSAR is part of the approved TWRS Authorization Basis (AB). This document describes, identifies, and defines the contents and structure of the TWRS FSAR Hazard Analysis Database and documents the configuration control changes made to the database. The TWRS Hazard Analysis Database contains the collection of information generated during the initial hazard evaluations and the subsequent hazard and accident analysis activities. The database supports the preparation of Chapters 3,4, and 5 of the TWRS FSAR and the USQ process and consists of two major, interrelated data sets: (1) Hazard Evaluation Database--Data from the results of the hazard evaluations; and (2) Hazard Topography Database--Data from the system familiarization and hazard identification.

  15. Masterwork Art Lesson: Kandinsky Watercolors.

    Science.gov (United States)

    LiPira, Michael

    2003-01-01

    Presents an art lesson used with sixth-grade students which also can be used with other grade levels. Explains that the artwork of Wassily Kandinsky served as inspiration for this lesson. Explains that the students learned about abstract art and used watercolors to create their own paintings in the style of Kandinsky. (CMK)

  16. Bead Game Simulation. Lesson Plan.

    Science.gov (United States)

    Ripp, Ken

    This lesson plan offers students the opportunity to participate in the three basic economic systems (market, command, and tradition). By working in each of the systems, students will internalize the fundamental values present in each system and will gain insights into the basic advantages and disadvantages of each system. The lesson plan provides…

  17. Simple and Practical Efficiency Lessons

    Science.gov (United States)

    Kolpin, Van

    2018-01-01

    The derivation of conditions necessary for Pareto efficient production and exchange is a lesson frequently showcased in microeconomic theory textbooks. Traditional delivery of this lesson is, however, limited in its scope of application and can be unnecessarily convoluted. The author shows that the universe of application is greatly expanded and a…

  18. Keiko, Killer Whale. [Lesson Plan].

    Science.gov (United States)

    Discovery Communications, Inc., Bethesda, MD.

    This lesson plan presents activities designed to help students understand that Keiko, the killer whale, lived for a long time in an aquarium and had to be taught to live independently; and that computer users can get updates on how Keiko is doing. The main activity of the lesson involves middle school students working in small groups to produce a…

  19. Lesson Study and History Education

    Science.gov (United States)

    Halvorsen, Anne-Lise; Kesler Lund, Alisa

    2013-01-01

    This article examines the experiences of a group of fifth-grade teachers who used lesson study, a teacher-driven form of professional development, to teach history in a project supported by a Teaching American History Grant. The project addressed the following questions: What does a lesson study cycle for history education look like? What…

  20. Database for propagation models

    Science.gov (United States)

    Kantak, Anil V.

    1991-07-01

    A propagation researcher or a systems engineer who intends to use the results of a propagation experiment is generally faced with various database tasks such as the selection of the computer software, the hardware, and the writing of the programs to pass the data through the models of interest. This task is repeated every time a new experiment is conducted or the same experiment is carried out at a different location generating different data. Thus the users of this data have to spend a considerable portion of their time learning how to implement the computer hardware and the software towards the desired end. This situation may be facilitated considerably if an easily accessible propagation database is created that has all the accepted (standardized) propagation phenomena models approved by the propagation research community. Also, the handling of data will become easier for the user. Such a database construction can only stimulate the growth of the propagation research it if is available to all the researchers, so that the results of the experiment conducted by one researcher can be examined independently by another, without different hardware and software being used. The database may be made flexible so that the researchers need not be confined only to the contents of the database. Another way in which the database may help the researchers is by the fact that they will not have to document the software and hardware tools used in their research since the propagation research community will know the database already. The following sections show a possible database construction, as well as properties of the database for the propagation research.

  1. When a Patient Declines Curative Care: Management of a Ruptured Aortic Aneurysm

    Directory of Open Access Journals (Sweden)

    Sangeeta Lamba

    2013-09-01

    Full Text Available The management of major vascular emergencies in the emergency department (ED involves rapid, aggressive resuscitation followed by emergent definitive surgery. However, for some patients this traditional approach may not be consistent with their goals and values. We explore the appropriate way to determine best treatment practices when patients elect to forego curative care in the ED, while reviewing such a case. We present the case of a 72-year-old patient who presented to the ED with a ruptured abdominal aortic aneurysm, but refused surgery. We discuss the transition of the patient from a curative to a comfort care approach with appropriate direct referral to hospice from the ED. Using principles of autonomy, decision-making capacity, informed consent, prognostication, and goals-of-care, ED clinicians are best able to align their approach with patients’ goals and values. [West J Emerg Med. 2013;14(5:555–558.

  2. Results of adjuvant chemo radiation after curative surgery for gastric cancer. A retrospective study

    International Nuclear Information System (INIS)

    Muller, Bettina; Balbontin, Paulina; Trujillo, Cristian; Becerra, Sergio; Sola, Antonio; Neveu, Rodrigo; Fernandez, Roberto; Buchholtz, Martin; Villanueva, Luis; Cerda, Berta

    2009-01-01

    Background: Survival rates after curative surgery for gastric cancer are disappointing. Therefore adjuvant therapeutic strategies are required. Aim: To analyze survival and side effects of treatment among gastric cancer patients treated with adjuvant chemoradiotherapy after curative resection of gastric adenocarcinoma. Material and methods: Retrospective review of medical records of 74 patients aged 20 to 74 years, treated with complete resection of gastric adenocarcinoma followed by adjuvant chemo radiation. Survival analysis was based on the records and information from the National Mortality Registry. Results: Five years survival fluctuated from 50% among patients in stage 1B to 25% among those is stage IV. Significant acute toxicity was observed in 23 patients (31%). No patients died due to acute toxicity. Eleven patients (16.4%) developed significant late toxicity, with two possible deaths related to treatment. Conclusions: Postoperative chemoradiotherapy is feasible in our experience. Continuos infusion of 5- fluoruracil is recommended to reduce toxicity

  3. Curating Research: e-Merging New Roles and Responsibilities in the European Landscape (Conference Report

    Directory of Open Access Journals (Sweden)

    Inge Angevaare

    2009-11-01

    Full Text Available On 17 April 2009 LIBER, the Koninklijke Bibliotheek (KB and the Netherlands Coalition for Digital Preservation (NCDD co-organised LIBER’s first workshop on digital curation/preservation, providing an introduction to the theme of digital curation and different perspectives on the roles research libraries can and/or must play in keeping the digital records of science safe for future generations. Speakers included Eileen Fenton (Portico, Dale Peters (DRIVER, Maria Heijne (TU Delft Library, Jeffrey van der Hoeven (KB, PARSE.insight and ninety workshop attendees. The paper includes the conference report and some results of the PARSE.insight online survey into digital preservation practices of LIBER libraries.

  4. WEB-BASED DATABASE ON RENEWAL TECHNOLOGIES ...

    Science.gov (United States)

    As U.S. utilities continue to shore up their aging infrastructure, renewal needs now represent over 43% of annual expenditures compared to new construction for drinking water distribution and wastewater collection systems (Underground Construction [UC], 2016). An increased understanding of renewal options will ultimately assist drinking water utilities in reducing water loss and help wastewater utilities to address infiltration and inflow issues in a cost-effective manner. It will also help to extend the service lives of both drinking water and wastewater mains. This research effort involved collecting case studies on the use of various trenchless pipeline renewal methods and providing the information in an online searchable database. The overall objective was to further support technology transfer and information sharing regarding emerging and innovative renewal technologies for water and wastewater mains. The result of this research is a Web-based, searchable database that utility personnel can use to obtain technology performance and cost data, as well as case study references. The renewal case studies include: technologies used; the conditions under which the technology was implemented; costs; lessons learned; and utility contact information. The online database also features a data mining tool for automated review of the technologies selected and cost data. Based on a review of the case study results and industry data, several findings are presented on tren

  5. Product Licenses Database Application

    CERN Document Server

    Tonkovikj, Petar

    2016-01-01

    The goal of this project is to organize and centralize the data about software tools available to CERN employees, as well as provide a system that would simplify the license management process by providing information about the available licenses and their expiry dates. The project development process is consisted of two steps: modeling the products (software tools), product licenses, legal agreements and other data related to these entities in a relational database and developing the front-end user interface so that the user can interact with the database. The result is an ASP.NET MVC web application with interactive views for displaying and managing the data in the underlying database.

  6. LandIT Database

    DEFF Research Database (Denmark)

    Iftikhar, Nadeem; Pedersen, Torben Bach

    2010-01-01

    and reporting purposes. This paper presents the LandIT database; which is result of the LandIT project, which refers to an industrial collaboration project that developed technologies for communication and data integration between farming devices and systems. The LandIT database in principal is based...... on the ISOBUS standard; however the standard is extended with additional requirements, such as gradual data aggregation and flexible exchange of farming data. This paper describes the conceptual and logical schemas of the proposed database based on a real-life farming case study....

  7. JICST Factual Database(2)

    Science.gov (United States)

    Araki, Keisuke

    The computer programme, which builds atom-bond connection tables from nomenclatures, is developed. Chemical substances with their nomenclature and varieties of trivial names or experimental code numbers are inputted. The chemical structures of the database are stereospecifically stored and are able to be searched and displayed according to stereochemistry. Source data are from laws and regulations of Japan, RTECS of US and so on. The database plays a central role within the integrated fact database service of JICST and makes interrelational retrieval possible.

  8. RNA STRAND: The RNA Secondary Structure and Statistical Analysis Database

    Directory of Open Access Journals (Sweden)

    Andronescu Mirela

    2008-08-01

    Full Text Available Abstract Background The ability to access, search and analyse secondary structures of a large set of known RNA molecules is very important for deriving improved RNA energy models, for evaluating computational predictions of RNA secondary structures and for a better understanding of RNA folding. Currently there is no database that can easily provide these capabilities for almost all RNA molecules with known secondary structures. Results In this paper we describe RNA STRAND – the RNA secondary STRucture and statistical ANalysis Database, a curated database containing known secondary structures of any type and organism. Our new database provides a wide collection of known RNA secondary structures drawn from public databases, searchable and downloadable in a common format. Comprehensive statistical information on the secondary structures in our database is provided using the RNA Secondary Structure Analyser, a new tool we have developed to analyse RNA secondary structures. The information thus obtained is valuable for understanding to which extent and with which probability certain structural motifs can appear. We outline several ways in which the data provided in RNA STRAND can facilitate research on RNA structure, including the improvement of RNA energy models and evaluation of secondary structure prediction programs. In order to keep up-to-date with new RNA secondary structure experiments, we offer the necessary tools to add solved RNA secondary structures to our database and invite researchers to contribute to RNA STRAND. Conclusion RNA STRAND is a carefully assembled database of trusted RNA secondary structures, with easy on-line tools for searching, analyzing and downloading user selected entries, and is publicly available at http://www.rnasoft.ca/strand.

  9. LeishCyc: a biochemical pathways database for Leishmania major

    Directory of Open Access Journals (Sweden)

    Doyle Maria A

    2009-06-01

    Full Text Available Abstract Background Leishmania spp. are sandfly transmitted protozoan parasites that cause a spectrum of diseases in more than 12 million people worldwide. Much research is now focusing on how these parasites adapt to the distinct nutrient environments they encounter in the digestive tract of the sandfly vector and the phagolysosome compartment of mammalian macrophages. While data mining and annotation of the genomes of three Leishmania species has provided an initial inventory of predicted metabolic components and associated pathways, resources for integrating this information into metabolic networks and incorporating data from transcript, protein, and metabolite profiling studies is currently lacking. The development of a reliable, expertly curated, and widely available model of Leishmania metabolic networks is required to facilitate systems analysis, as well as discovery and prioritization of new drug targets for this important human pathogen. Description The LeishCyc database was initially built from the genome sequence of Leishmania major (v5.2, based on the annotation published by the Wellcome Trust Sanger Institute. LeishCyc was manually curated to remove errors, correct automated predictions, and add information from the literature. The ongoing curation is based on public sources, literature searches, and our own experimental and bioinformatics studies. In a number of instances we have improved on the original genome annotation, and, in some ambiguous cases, collected relevant information from the literature in order to help clarify gene or protein annotation in the future. All genes in LeishCyc are linked to the corresponding entry in GeneDB (Wellcome Trust Sanger Institute. Conclusion The LeishCyc database describes Leishmania major genes, gene products, metabolites, their relationships and biochemical organization into metabolic pathways. LeishCyc provides a systematic approach to organizing the evolving information about Leishmania

  10. Predicting prognosis in hepatocellular carcinoma after curative surgery with common clinicopathologic parameters

    International Nuclear Information System (INIS)

    Hao, Ke; Sham, Pak C; Poon, Ronnie TP; Luk, John M; Lee, Nikki PY; Mao, Mao; Zhang, Chunsheng; Ferguson, Mark D; Lamb, John; Dai, Hongyue; Ng, Irene O

    2009-01-01

    Surgical resection is one important curative treatment for hepatocellular carcinoma (HCC), but the prognosis following surgery differs substantially and such large variation is mainly unexplained. A review of the literature yields a number of clinicopathologic parameters associated with HCC prognosis. However, the results are not consistent due to lack of systemic approach to establish a prediction model incorporating all these parameters. We conducted a retrospective analysis on the common clinicopathologic parameters from a cohort of 572 ethnic Chinese HCC patients who received curative surgery. The cases were randomly divided into training (n = 272) and validation (n = 300) sets. Each parameter was individually tested and the significant parameters were entered into a linear classifier for model building, and the prediction accuracy was assessed in the validation set Our findings based on the training set data reveal 6 common clinicopathologic parameters (tumor size, number of tumor nodules, tumor stage, venous infiltration status, and serum α-fetoprotein and total albumin levels) that were significantly associated with the overall HCC survival and disease-free survival (time to recurrence). We next built a linear classifier model by multivariate Cox regression to predict prognostic outcomes of HCC patients after curative surgery This analysis detected a considerable fraction of variance in HCC prognosis and the area under the ROC curve was about 70%. We further evaluated the model using two other protocols; leave-one-out procedure (n = 264) and independent validation (n = 300). Both were found to have excellent prediction power. The predicted score could separate patients into distinct groups with respect to survival (p-value = 1.8e-12) and disease free survival (p-value = 3.2e-7). This described model will provide valuable guidance on prognosis after curative surgery for HCC in clinical practice. The adaptive nature allows easy accommodation for future new

  11. BELIEF dashboard - a web-based curation interface to support generation of BEL networks

    OpenAIRE

    Madan, Sumit; Hodapp, Sven; Fluck, Juliane

    2015-01-01

    The relevance of network-based approaches in systems biology to achieve a better understanding of biological mechanisms has increased enormously. The Biological Expression Language (BEL) is well designed to collate findings from scientific literature into biological network models. To facilitate encoding and biocuration of such findings in BEL, a free and user-friendly web-based curation interface called BELIEF Dashboard has been developed. The interface incorporates an information extraction...

  12. Community Intelligence in Knowledge Curation: An Application to Managing Scientific Nomenclature

    OpenAIRE

    Dai, Lin; Xu, Chao; Tian, Ming; Sang, Jian; Zou, Dong; Li, Ang; Liu, Guocheng; Chen, Fei; Wu, Jiayan; Xiao, Jingfa; Wang, Xumin; Yu, Jun; Zhang, Zhang

    2013-01-01

    Harnessing community intelligence in knowledge curation bears significant promise in dealing with communication and education in the flood of scientific knowledge. As knowledge is accumulated at ever-faster rates, scientific nomenclature, a particular kind of knowledge, is concurrently generated in all kinds of fields. Since nomenclature is a system of terms used to name things in a particular discipline, accurate translation of scientific nomenclature in different languages is of critical im...

  13. Updates to the Cool Season Food Legume Genome Database: Resources for pea, lentil, faba bean and chickpea genetics, genomics and breeding

    Science.gov (United States)

    The Cool Season Food Legume Genome database (CSFL, www.coolseasonfoodlegume.org) is an online resource for genomics, genetics, and breeding research for chickpea, lentil,pea, and faba bean. The user-friendly and curated website allows for all publicly available map,marker,trait, gene,transcript, ger...

  14. Distribution and utilization of curative primary healthcare services in Lahej, Yemen.

    Science.gov (United States)

    Bawazir, A A; Bin Hawail, T S; Al-Sakkaf, K A Z; Basaleem, H O; Muhraz, A F; Al-Shehri, A M

    2013-09-01

    No evidence-based data exist on the availability, accessibility and utilization of healthcare services in Lahej Governorate, Yemen. The aim of this study was to assess the distribution and utilization of curative services in primary healthcare units and centres in Lahej. Cross-sectional study (clustering sample). This study was conducted in three of the 15 districts in Lahej between December 2009 and August 2010. Household members were interviewed using a questionnaire to determine sociodemographic characteristics and types of healthcare services available in the area. The distribution of health centres, health units and hospitals did not match the size of the populations or areas of the districts included in this study. Geographical accessibility was the main obstacle to utilization. Factors associated with the utilization of curative services were significantly related to the time required to reach the nearest facility, seeking curative services during illness and awareness of the availability of health facilities (P < 0.01). There is an urgent need to look critically and scientifically at the distribution of healthcare services in the region in order to ensure accessibility and quality of services. Copyright © 2013 The Royal Society for Public Health. Published by Elsevier Ltd. All rights reserved.

  15. [Two Cases of Curative Resection of Locally Advanced Rectal Cancer after Preoperative Chemotherapy].

    Science.gov (United States)

    Mitsuhashi, Noboru; Shimizu, Yoshiaki; Kuboki, Satoshi; Yoshitomi, Hideyuki; Kato, Atsushi; Ohtsuka, Masayuki; Shimizu, Hiroaki; Miyazaki, Masaru

    2015-11-01

    Reports of conversion in cases of locally advanced colorectal cancer have been increasing. Here, we present 2 cases in which curative resection of locally advanced rectal cancer accompanied by intestinal obstruction was achieved after establishing a stoma and administering chemotherapy. The first case was of a 46-year-old male patient diagnosed with upper rectal cancer and intestinal obstruction. Because of a high level of retroperitoneal invasion, after establishing a sigmoid colostomy, 13 courses of mFOLFOX6 plus Pmab were administered. Around 6 months after the initial surgery, low anterior resection for rectal cancer and surgery to close the stoma were performed. Fourteen days after curative resection, the patient was discharged from the hospital. The second case was of a 66-year-old male patient with a circumferential tumor extending from Rs to R, accompanied by right ureter infiltration and sub-intestinal obstruction. After establishing a sigmoid colostomy, 11 courses of mFOLFOX6 plus Pmab were administered. Five months after the initial surgery, anterior resection of the rectum and surgery to close the stoma were performed. Twenty days after curative resection, the patient was released from the hospital. No recurrences have been detected in either case.

  16. Data Albums: An Event Driven Search, Aggregation and Curation Tool for Earth Science

    Science.gov (United States)

    Ramachandran, Rahul; Kulkarni, Ajinkya; Maskey, Manil; Bakare, Rohan; Basyal, Sabin; Li, Xiang; Flynn, Shannon

    2014-01-01

    Approaches used in Earth science research such as case study analysis and climatology studies involve discovering and gathering diverse data sets and information to support the research goals. To gather relevant data and information for case studies and climatology analysis is both tedious and time consuming. Current Earth science data systems are designed with the assumption that researchers access data primarily by instrument or geophysical parameter. In cases where researchers are interested in studying a significant event, they have to manually assemble a variety of datasets relevant to it by searching the different distributed data systems. This paper presents a specialized search, aggregation and curation tool for Earth science to address these challenges. The search rool automatically creates curated 'Data Albums', aggregated collections of information related to a specific event, containing links to relevant data files [granules] from different instruments, tools and services for visualization and analysis, and information about the event contained in news reports, images or videos to supplement research analysis. Curation in the tool is driven via an ontology based relevancy ranking algorithm to filter out non relevant information and data.

  17. A comparative study to analyze the cost of curative care at primary health center in Ahmedabad

    Directory of Open Access Journals (Sweden)

    Mathur Neeta

    2010-01-01

    Full Text Available Objectives: To determine the unit cost of curative care provided at Primary Health Centers (PHCs and to examine the variation in unit cost in different PHCs. Materials and Methods: The present study was carried out in three PHCs of Ahmedabad district namely Sanathal, Nandej, and Uperdal, between 1 April, 2006 and 31 March, 2007. For estimating the cost of a health program, information on all the physical and human resources that were basic inputs to the PHC services were collected and grouped into two categories, non-recurrent (capital resources vehicles, buildings, etc. and recurrent resources (salaries, drugs, vaccines, contraceptives, maintenance, etc.. To generate the required data, two types of schedules were developed, daily time schedule and PHC/SC (Subcenter information schedule. Results: The unit cost of curative care was lowest (Rs. 29.43 for the Sanathal PHC and highest (Rs. 88.26 for the Uperdal PHC, followed by the Nandej PHC with Rs. 40.88, implying severe underutilization of curative care at the Uperdal PHC. Conclusions: Location of health facilities is a problem at many places. As relocation is not possible or even feasible, strengthening of infrastructure and facilities at these centers can be taken up immediately.

  18. Stop the Bleeding: the Development of a Tool to Streamline NASA Earth Science Metadata Curation Efforts

    Science.gov (United States)

    le Roux, J.; Baker, A.; Caltagirone, S.; Bugbee, K.

    2017-12-01

    The Common Metadata Repository (CMR) is a high-performance, high-quality repository for Earth science metadata records, and serves as the primary way to search NASA's growing 17.5 petabytes of Earth science data holdings. Released in 2015, CMR has the capability to support several different metadata standards already being utilized by NASA's combined network of Earth science data providers, or Distributed Active Archive Centers (DAACs). The Analysis and Review of CMR (ARC) Team located at Marshall Space Flight Center is working to improve the quality of records already in CMR with the goal of making records optimal for search and discovery. This effort entails a combination of automated and manual review, where each NASA record in CMR is checked for completeness, accuracy, and consistency. This effort is highly collaborative in nature, requiring communication and transparency of findings amongst NASA personnel, DAACs, the CMR team and other metadata curation teams. Through the evolution of this project it has become apparent that there is a need to document and report findings, as well as track metadata improvements in a more efficient manner. The ARC team has collaborated with Element 84 in order to develop a metadata curation tool to meet these needs. In this presentation, we will provide an overview of this metadata curation tool and its current capabilities. Challenges and future plans for the tool will also be discussed.

  19. Xylella fastidiosa comparative genomic database is an information resource to explore the annotation, genomic features, and biology of different strains

    Directory of Open Access Journals (Sweden)

    Alessandro M. Varani

    2012-01-01

    Full Text Available The Xylella fastidiosa comparative genomic database is a scientific resource with the aim to provide a user-friendly interface for accessing high-quality manually curated genomic annotation and comparative sequence analysis, as well as for identifying and mapping prophage-like elements, a marked feature of Xylella genomes. Here we describe a database and tools for exploring the biology of this important plant pathogen. The hallmarks of this database are the high quality genomic annotation, the functional and comparative genomic analysis and the identification and mapping of prophage-like elements. It is available from web site http://www.xylella.lncc.br.

  20. Classroom Management and Lesson Planning(4)

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    Lesson PlanningTask 1As teachers,we all need to plan our lessons before we teach.Make a list of things that you think need tobe included in a lesson plan.Then compare and discuss your list with another teacher.Also think about reasonswhy we need to plan our lessons.