WorldWideScience

Sample records for high-quality computational libraries

  1. Construction of High-Quality Camel Immune Antibody Libraries.

    Science.gov (United States)

    Romão, Ema; Poignavent, Vianney; Vincke, Cécile; Ritzenthaler, Christophe; Muyldermans, Serge; Monsion, Baptiste

    2018-01-01

    Single-domain antibodies libraries of heavy-chain only immunoglobulins from camelids or shark are enriched for high-affinity antigen-specific binders by a short in vivo immunization. Thus, potent binders are readily retrieved from relatively small-sized libraries of 10 7 -10 8 individual transformants, mostly after phage display and panning on a purified target. However, the remaining drawback of this strategy arises from the need to generate a dedicated library, for nearly every envisaged target. Therefore, all the procedures that shorten and facilitate the construction of an immune library of best possible quality are definitely a step forward. In this chapter, we provide the protocol to generate a high-quality immune VHH library using the Golden Gate Cloning strategy employing an adapted phage display vector where a lethal ccdB gene has to be substituted by the VHH gene. With this procedure, the construction of the library can be shortened to less than a week starting from bleeding the animal. Our libraries exceed 10 8 individual transformants and close to 100% of the clones harbor a phage display vector having an insert with the length of a VHH gene. These libraries are also more economic to make than previous standard approaches using classical restriction enzymes and ligations. The quality of the Nanobodies that are retrieved from immune libraries obtained by Golden Gate Cloning is identical to those from immune libraries made according to the classical procedure.

  2. Availability and Overlap of Quality Computer Science Journal Holdings in Selected University Libraries in Malaysia

    OpenAIRE

    Zainab, A.N.; Ng, S.L.

    2003-01-01

    The study reveals the availability status of quality journals in the field of computer science held in the libraries of the University of Malaya, (UM), University of Science Malaysia (USM), University of Technology Malaysia (UTM), National University of Malaysia (UKM) and University Putra Malaysia (UPM). These universities are selected since they offer degree programmes in computer science. The study also investigates the degree of overlaps and unique titles in the five libraries. The Univers...

  3. Computers, Education and the Library at The Bronx High School of Science.

    Science.gov (United States)

    Nachbar, Sondra; Sussman, Valerie

    1988-01-01

    Describes the services and programs offered by the library at The Bronx High School of Science. Topics discussed include the library collection; a basic library skills mini-course for freshmen and incoming sophomores; current uses of the library's computer system; and plans to automate the library's card catalog and circulation records.…

  4. BEAGLE: an application programming interface and high-performance computing library for statistical phylogenetics.

    Science.gov (United States)

    Ayres, Daniel L; Darling, Aaron; Zwickl, Derrick J; Beerli, Peter; Holder, Mark T; Lewis, Paul O; Huelsenbeck, John P; Ronquist, Fredrik; Swofford, David L; Cummings, Michael P; Rambaut, Andrew; Suchard, Marc A

    2012-01-01

    Phylogenetic inference is fundamental to our understanding of most aspects of the origin and evolution of life, and in recent years, there has been a concentration of interest in statistical approaches such as Bayesian inference and maximum likelihood estimation. Yet, for large data sets and realistic or interesting models of evolution, these approaches remain computationally demanding. High-throughput sequencing can yield data for thousands of taxa, but scaling to such problems using serial computing often necessitates the use of nonstatistical or approximate approaches. The recent emergence of graphics processing units (GPUs) provides an opportunity to leverage their excellent floating-point computational performance to accelerate statistical phylogenetic inference. A specialized library for phylogenetic calculation would allow existing software packages to make more effective use of available computer hardware, including GPUs. Adoption of a common library would also make it easier for other emerging computing architectures, such as field programmable gate arrays, to be used in the future. We present BEAGLE, an application programming interface (API) and library for high-performance statistical phylogenetic inference. The API provides a uniform interface for performing phylogenetic likelihood calculations on a variety of compute hardware platforms. The library includes a set of efficient implementations and can currently exploit hardware including GPUs using NVIDIA CUDA, central processing units (CPUs) with Streaming SIMD Extensions and related processor supplementary instruction sets, and multicore CPUs via OpenMP. To demonstrate the advantages of a common API, we have incorporated the library into several popular phylogenetic software packages. The BEAGLE library is free open source software licensed under the Lesser GPL and available from http://beagle-lib.googlecode.com. An example client program is available as public domain software.

  5. Computers in Academic Architecture Libraries.

    Science.gov (United States)

    Willis, Alfred; And Others

    1992-01-01

    Computers are widely used in architectural research and teaching in U.S. schools of architecture. A survey of libraries serving these schools sought information on the emphasis placed on computers by the architectural curriculum, accessibility of computers to library staff, and accessibility of computers to library patrons. Survey results and…

  6. Croatian library leaders’ views on (their library quality

    Directory of Open Access Journals (Sweden)

    Kornelija Petr Balog

    2014-04-01

    Full Text Available The purpose of this paper is to determine and describe the library culture in Croatian public libraries. Semi-structured interviews with 14 library directors (ten public and four academic were conducted. The tentative discussion topics were: definition of quality, responsibility for quality, satisfaction with library services, familiarization with user perspective of library and librarians, monitoring of user expectations and opinions. These interviews incorporate some of the findings of the project Evaluation of library and information services: public and academic libraries. The project investigates library culture in Croatian public and academic libraries and their preparedness for activities of performance measurement. The interviews reveal that library culture has changed positively in the past few years and that library leaders have positive attitude towards quality and evaluation activities. Library culture in Croatian libraries is a relatively new concept and as such was not actively developed and/or created. This article looks into the library culture of Croatian libraries, but at the same time investigates whether there is any trace of culture of assessment in them. Also, this article brings the latest update on views, opinions and atmosphere in Croatian public and academic libraries.

  7. Case Studies in Library Computer Systems.

    Science.gov (United States)

    Palmer, Richard Phillips

    Twenty descriptive case studies of computer applications in a variety of libraries are presented in this book. Computerized circulation, serial and acquisition systems in public, high school, college, university and business libraries are included. Each of the studies discusses: 1) the environment in which the system operates, 2) the objectives of…

  8. Cloud Computing in KAUST Library: Beyond Remote Hosting

    KAUST Repository

    Yu, Yi

    2013-12-01

    Enterprise computing is the key strategic approach for KAUST to build its modern IT landscape. In such a strategic direction and technical environment, the library tries to establish library technology by catching new trends which help to make the library more efficient and sufficient. This paper focuses on the cloud computing development in the KAUST library, by using real world scenarios and first-hand experiences to describe what cloud computing means for KAUST library. It addresses the difficulties that were met by the library during the implementation process, how cloud computing affects the functional performance and work procedure of the library, how it impacts the style and modal of the library’s technical service and systems administration, how it changes the relationships and cooperation among the involved players (the library, campus IT and vendors), and what the benefits and disadvantages are. The story of cloud computing at KAUST will share the knowledge and lessons that the KAUST library learnt during its development, and will also point out the future direction of cloud computing at KAUST.

  9. The Impact of Information Technology on Library Anxiety: The Role of Computer Attitudes

    Directory of Open Access Journals (Sweden)

    Qun G. Jiao

    2017-09-01

    Full Text Available Over the past two decades, computer-based technologies have become dominant forces to shape and reshape the products and services the academic library has to offer. The applicationo of library technologies has had a profound impact on the way library resources are being used. Although many students continue to experience high levels of library anxiety, it is likely that the new technologies in the library have led to them experiencing other forms of negative affective states that may be, in part, a function of their attitude towards computers. This study investigates whether students' computer attitudes predict levels of library anxiety.

  10. AstroCV: Astronomy computer vision library

    Science.gov (United States)

    González, Roberto E.; Muñoz, Roberto P.; Hernández, Cristian A.

    2018-04-01

    AstroCV processes and analyzes big astronomical datasets, and is intended to provide a community repository of high performance Python and C++ algorithms used for image processing and computer vision. The library offers methods for object recognition, segmentation and classification, with emphasis in the automatic detection and classification of galaxies.

  11. The Adoption of Cloud Computing Technology for Library Services ...

    African Journals Online (AJOL)

    The study investigated the rationales for the adoption of cloud computing technology for library services in NOUN Library. Issues related to the existing computer network available in NOUN library such as LAN, WAN, rationales for the adoption of cloud computing in NOUN library such as the need to disclose their collections ...

  12. Multi-line split DNA synthesis: a novel combinatorial method to make high quality peptide libraries

    Directory of Open Access Journals (Sweden)

    Ueno Shingo

    2004-09-01

    Full Text Available Abstract Background We developed a method to make a various high quality random peptide libraries for evolutionary protein engineering based on a combinatorial DNA synthesis. Results A split synthesis in codon units was performed with mixtures of bases optimally designed by using a Genetic Algorithm program. It required only standard DNA synthetic reagents and standard DNA synthesizers in three lines. This multi-line split DNA synthesis (MLSDS is simply realized by adding a mix-and-split process to normal DNA synthesis protocol. Superiority of MLSDS method over other methods was shown. We demonstrated the synthesis of oligonucleotide libraries with 1016 diversity, and the construction of a library with random sequence coding 120 amino acids containing few stop codons. Conclusions Owing to the flexibility of the MLSDS method, it will be able to design various "rational" libraries by using bioinformatics databases.

  13. The computer library of experimental neutron data

    International Nuclear Information System (INIS)

    Bychkov, V.M.; Manokhin, V.N.; Surgutanov, V.V.

    1976-05-01

    The paper describes the computer library of experimental neutron data at the Obninsk Nuclear Data Centre. The format of the library (EXFOR) and the system of programmes for supplying the library are briefly described. (author)

  14. Application of Quality Assurance Mechanisms for Effective Library ...

    African Journals Online (AJOL)

    The main strategy used for implementation of quality assurance was integration of the library` services quality assurance agenda into the university structures and the quality assurance mechanisms available and used in Nigerian university libraries were programme accreditation and benchmarking of library systems.

  15. The NEA computer program library: a possible GDMS application

    International Nuclear Information System (INIS)

    Schuler, W.

    1978-01-01

    NEA Computer Program library maintains a series of eleven sequential computer files, used for linked applications in managing their stock of computer codes for nuclear reactor calculations, storing index and program abstract information, and administering their service to requesters. The high data redundancy beween the files suggests that a data base approach would be valid and this paper suggests a possible 'schema' for an CODASYL GDMS

  16. A lightweight communication library for distributed computing

    International Nuclear Information System (INIS)

    Groen, Derek; Rieder, Steven; Zwart, Simon Portegies; Grosso, Paola; Laat, Cees de

    2010-01-01

    We present MPWide, a platform-independent communication library for performing message passing between computers. Our library allows coupling of several local message passing interface (MPI) applications through a long-distance network and is specifically optimized for such communications. The implementation is deliberately kept lightweight and platform independent, and the library can be installed and used without administrative privileges. The only requirements are a C++ compiler and at least one open port to a wide-area network on each site. In this paper we present the library, describe the user interface, present performance tests and apply MPWide in a large-scale cosmological N-body simulation on a network of two computers, one in Amsterdam and the other in Tokyo.

  17. Computer Science Professionals and Greek Library Science

    Science.gov (United States)

    Dendrinos, Markos N.

    2008-01-01

    This paper attempts to present the current state of computer science penetration into librarianship in terms of both workplace and education issues. The shift from material libraries into digital libraries is mirrored in the corresponding shift from librarians into information scientists. New library data and metadata, as well as new automated…

  18. Assessing the Quality of Academic Libraries on the Web: The Development and Testing of Criteria.

    Science.gov (United States)

    Chao, Hungyune

    2002-01-01

    This study develops and tests an instrument useful for evaluating the quality of academic library Web sites. Discusses criteria for print materials and human-computer interfaces; user-based perspectives; the use of factor analysis; a survey of library experts; testing reliability through analysis of variance; and regression models. (Contains 53…

  19. The Student/Library Computer Science Collaborative

    Science.gov (United States)

    Hahn, Jim

    2015-01-01

    With funding from an Institute of Museum and Library Services demonstration grant, librarians of the Undergraduate Library at the University of Illinois at Urbana-Champaign partnered with students in computer science courses to design and build student-centered mobile apps. The grant work called for demonstration of student collaboration…

  20. Assessing users satisfaction with service quality in Slovenian public library

    Directory of Open Access Journals (Sweden)

    Igor Podbrežnik

    2016-07-01

    Full Text Available Purpose: A research was made into user satisfaction with regard to the quality of library services in one of the Slovenian public libraries. The aim was to establish the type of service quality level actually expected by the users, and to determine their satisfaction with the current quality level of available library services.Methodology: The research was performed by means of the SERVQUAL measuring tool which was used to determine the size and direction of the gap between the detected and the expected quality of library services among public library users.Results: Different groups of users provide different assessments of specific quality factors, and a library cannot satisfy the expectations of each and every user if most quality factors display discrepancies between the estimated perception and expectations. The users expect more reliable services and more qualified library staff members who would understand and allocate time for each user’s individual needs. The largest discrepancies from the expectations are detected among users in the under-35 age group and among the more experienced and skilled library users. The results of factor analysis confirm the fact that a higher number of quality factors can be explained by three common factors affecting the satisfaction of library users. A strong connection between user satisfaction and their assessment of the integral quality of services and loyalty has been established.Research restrictions: The research results should not be generalised and applied to all Slovenian public libraries since they differ in many important aspects. In addition, a non-random sampling method was used.Research originality/Applicability: The conducted research illustrates the use of a measuring tool that was developed with the aim of determining the satisfaction of users with the quality of library services in Slovenian public libraries. Keywords: public library, user satisfaction, quality of library services, user

  1. News from the Library: A one-stop-shop for computing literature: ACM Digital Library

    CERN Multimedia

    CERN Library

    2011-01-01

    The Association for Computing Machinery, ACM, is the world’s largest educational and scientific computing society. Among others, the ACM provides the computing field's premier Digital Library and serves its members and the computing profession with leading-edge publications, conferences, and career resources.   ACM Digital Library is available to the CERN community. The most popular journal here at CERN is Communications of the ACM. However, the collection offers access to a series of other valuable important academic journals such as Journal of the ACM and even fulltext of a series of classical books. In addition, users have access to the ACM Guide to Computing Literature, the most comprehensive bibliographic database focusing on computing, integrated with ACM’s full-text articles and including features such as ACM Author Profile Pages - which provides bibliographic and bibliometric data for over 1,000,000 authors in the field. ACM Digital Library is an excellent com...

  2. A lightweight communication library for distributed computing

    NARCIS (Netherlands)

    Groen, D.; Rieder, S.; Grosso, P.; de Laat, C.; Portegies Zwart, S.

    2010-01-01

    We present MPWide, a platform-independent communication library for performing message passing between computers. Our library allows coupling of several local message passing interface (MPI) applications through a long-distance network and is specifically optimized for such communications. The

  3. Libraries\\\\\\' Nationwide Membership (Ghadir Quality Assessment Using SERVQUAL: Co-Libraries\\\\\\' Perspective

    Directory of Open Access Journals (Sweden)

    Sirous Alidousti

    2012-02-01

    Full Text Available Ghadir Program has been developed to make direct access to academic universities’ resources. This program has been implemented in Ministry of Science, Research and Education by Iranian Research Institute for Information Science and Technology as the coordinating centre since 1999 in 240 libraries after a period of pilot run. After a couple of days, it was necessary to assess the quality of this program to make changes if required. Therefore, here the results of assessing the quality of services provided by the coordinating center from the viewpoint of participant libraries were presented. Servqual applied as the base for this quality assessment. The population of this research was the entire participants which consist of 240 libraries affiliated to 66 universities and research centers. The questionnaire as the research instrument was sent to the libraries’ managers and they were asked to give them to Ghadir Program agents to fill it as well. As the result, among the questionnaires returned form 131 libraries, 178 ones were analyzable. The satisfaction of participant libraries from the services received and the gap between these services and their expectations from the coordinating center was investigated. According to the findings, the satisfaction of libraries from this center was more than average (3.5 from 5. In the Servqual dimensions, empathy and responsiveness had the maximum and reliability had the minimum gap between services received and expectations from the coordinating center.

  4. Integrating total quality management in a library setting

    CERN Document Server

    Jurow, Susan

    2013-01-01

    Improve the delivery of library services by implementing total quality management (TQM), a system of continuous improvement employing participative management and centered on the needs of customers. Although TQM was originally designed for and successfully applied in business and manufacturing settings, this groundbreaking volume introduces strategies for translating TQM principles from the profit-based manufacturing sector to the library setting. Integrating Total Quality Management in a Library Setting shows librarians how to improve library services by implementing strategies such as employ

  5. A reliable computational workflow for the selection of optimal screening libraries.

    Science.gov (United States)

    Gilad, Yocheved; Nadassy, Katalin; Senderowitz, Hanoch

    2015-01-01

    The experimental screening of compound collections is a common starting point in many drug discovery projects. Successes of such screening campaigns critically depend on the quality of the screened library. Many libraries are currently available from different vendors yet the selection of the optimal screening library for a specific project is challenging. We have devised a novel workflow for the rational selection of project-specific screening libraries. The workflow accepts as input a set of virtual candidate libraries and applies the following steps to each library: (1) data curation; (2) assessment of ADME/T profile; (3) assessment of the number of promiscuous binders/frequent HTS hitters; (4) assessment of internal diversity; (5) assessment of similarity to known active compound(s) (optional); (6) assessment of similarity to in-house or otherwise accessible compound collections (optional). For ADME/T profiling, Lipinski's and Veber's rule-based filters were implemented and a new blood brain barrier permeation model was developed and validated (85 and 74 % success rate for training set and test set, respectively). Diversity and similarity descriptors which demonstrated best performances in terms of their ability to select either diverse or focused sets of compounds from three databases (Drug Bank, CMC and CHEMBL) were identified and used for diversity and similarity assessments. The workflow was used to analyze nine common screening libraries available from six vendors. The results of this analysis are reported for each library providing an assessment of its quality. Furthermore, a consensus approach was developed to combine the results of these analyses into a single score for selecting the optimal library under different scenarios. We have devised and tested a new workflow for the rational selection of screening libraries under different scenarios. The current workflow was implemented using the Pipeline Pilot software yet due to the usage of generic

  6. DMG-α--a computational geometry library for multimolecular systems.

    Science.gov (United States)

    Szczelina, Robert; Murzyn, Krzysztof

    2014-11-24

    The DMG-α library grants researchers in the field of computational biology, chemistry, and biophysics access to an open-sourced, easy to use, and intuitive software for performing fine-grained geometric analysis of molecular systems. The library is capable of computing power diagrams (weighted Voronoi diagrams) in three dimensions with 3D periodic boundary conditions, computing approximate projective 2D Voronoi diagrams on arbitrarily defined surfaces, performing shape properties recognition using α-shape theory and can do exact Solvent Accessible Surface Area (SASA) computation. The software is written mainly as a template-based C++ library for greater performance, but a rich Python interface (pydmga) is provided as a convenient way to manipulate the DMG-α routines. To illustrate possible applications of the DMG-α library, we present results of sample analyses which allowed to determine nontrivial geometric properties of two Escherichia coli-specific lipids as emerging from molecular dynamics simulations of relevant model bilayers.

  7. INTRIGOSS: A new Library of High Resolution Synthetic Spectra

    Science.gov (United States)

    Franchini, Mariagrazia; Morossi, Carlo; Di Marcancantonio, Paolo; Chavez, Miguel; GES-Builders

    2018-01-01

    INTRIGOSS (INaf Trieste Grid Of Synthetic Spectra) is a new High Resolution (HiRes) synthetic spectral library designed for studying F, G, and K stars. The library is based on atmosphere models computed with specified individual element abundances via ATLAS12 code. Normalized SPectra (NSP) and surface Flux SPectra (FSP), in the 4800-5400 Å wavelength range, were computed by means of the SPECTRUM code. The synthetic spectra are computed with an atomic and bi-atomic molecular line list including "bona fide" Predicted Lines (PLs) built by tuning loggf to reproduce very high SNR Solar spectrum and the UVES-U580 spectra of five cool giants extracted from the Gaia-ESO survey (GES). The astrophysical gf-values were then assessed by using more than 2000 stars with homogenous and accurate atmosphere parameters and detailed chemical composition from GES. The validity and greater accuracy of INTRIGOSS NSPs and FSPs with respect to other available spectral libraries is discussed. INTRIGOSS will be available on the web and will be a valuable tool for both stellar atmospheric parameters and stellar population studies.

  8. Library and information services: impact on patient care quality.

    Science.gov (United States)

    Marshall, Joanne Gard; Morgan, Jennifer Craft; Thompson, Cheryl A; Wells, Amber L

    2014-01-01

    The purpose of this paper is to explore library and information service impact on patient care quality. A large-scale critical incident survey of physicians and residents at 56 library sites serving 118 hospitals in the USA and Canada. Respondents were asked to base their answers on a recent incident in which they had used library resources to search for information related to a specific clinical case. Of 4,520 respondents, 75 percent said that they definitely or probably handled patient care differently using information obtained through the library. In a multivariate analysis, three summary clinical outcome measures were used as value and impact indicators: first, time saved; second, patient care changes; and third, adverse events avoided. The outcomes were examined in relation to four information access methods: first, asking librarian for assistance; second, performing search in a physical library; third, searching library's web site; or fourth, searching library resources on an institutional intranet. All library access methods had consistently positive relationships with the clinical outcomes, providing evidence that library services have a positive impact on patient care quality. Electronic collections and services provided by the library and the librarian contribute to patient care quality.

  9. Assessing service quality satisfying the expectations of library customers

    CERN Document Server

    Hernon, Peter; Dugan, Robert

    2015-01-01

    Academic and public libraries are continuing to transform as the information landscape changes, expanding their missions into new service roles that call for improved organizational performance and accountability. Since Assessing Service Quality premiered in 1998, receiving the prestigious Highsmith Library Literature Award, scores of library managers and administrators have trusted its guidance for applying a customer-centered approach to service quality and performance evaluation. This extensively revised and updated edition explores even further the ways technology influences both the experiences of library customers and the ways libraries themselves can assess those experiences.

  10. A Generic High-performance GPU-based Library for PDE solvers

    DEFF Research Database (Denmark)

    Glimberg, Stefan Lemvig; Engsig-Karup, Allan Peter

    , the privilege of high-performance parallel computing is now in principle accessible for many scientific users, no matter their economic resources. Though being highly effective units, GPUs and parallel architectures in general, pose challenges for software developers to utilize their efficiency. Sequential...... legacy codes are not always easily parallelized and the time spent on conversion might not pay o in the end. We present a highly generic C++ library for fast assembling of partial differential equation (PDE) solvers, aiming at utilizing the computational resources of GPUs. The library requires a minimum...... of GPU computing knowledge, while still oering the possibility to customize user-specic solvers at kernel level if desired. Spatial dierential operators are based on matrix free exible order nite dierence approximations. These matrix free operators minimize both memory consumption and main memory access...

  11. Academic Libraries and Quality: An Analysis and Evaluation Framework

    Science.gov (United States)

    Atkinson, Jeremy

    2017-01-01

    The paper proposes and describes a framework for academic library quality to be used by new and more experienced library practitioners and by others involved in considering the quality of academic libraries' services and provision. The framework consists of eight themes and a number of questions to examine within each theme. The framework was…

  12. Construction of high quality Gateway™ entry libraries and their application to yeast two-hybrid for the monocot model plant Brachypodium distachyon

    Directory of Open Access Journals (Sweden)

    Kumimoto Roderick W

    2011-05-01

    Full Text Available Abstract Background Monocots, especially the temperate grasses, represent some of the most agriculturally important crops for both current food needs and future biofuel development. Because most of the agriculturally important grass species are difficult to study (e.g., they often have large, repetitive genomes and can be difficult to grow in laboratory settings, developing genetically tractable model systems is essential. Brachypodium distachyon (hereafter Brachypodium is an emerging model system for the temperate grasses. To fully realize the potential of this model system, publicly accessible discovery tools are essential. High quality cDNA libraries that can be readily adapted for multiple downstream purposes are a needed resource. Additionally, yeast two-hybrid (Y2H libraries are an important discovery tool for protein-protein interactions and are not currently available for Brachypodium. Results We describe the creation of two high quality, publicly available Gateway™ cDNA entry libraries and their derived Y2H libraries for Brachypodium. The first entry library represents cloned cDNA populations from both short day (SD, 8/16-h light/dark and long day (LD, 20/4-h light/dark grown plants, while the second library was generated from hormone treated tissues. Both libraries have extensive genome coverage (~5 × 107 primary clones each and average clone lengths of ~1.5 Kb. These entry libraries were then used to create two recombination-derived Y2H libraries. Initial proof-of-concept screens demonstrated that a protein with known interaction partners could readily re-isolate those partners, as well as novel interactors. Conclusions Accessible community resources are a hallmark of successful biological model systems. Brachypodium has the potential to be a broadly useful model system for the grasses, but still requires many of these resources. The Gateway™ compatible entry libraries created here will facilitate studies for multiple user

  13. Study of application technology of ultra-high speed computer to the elucidation of complex phenomena

    International Nuclear Information System (INIS)

    Sekiguchi, Tomotsugu

    1996-01-01

    The basic design of numerical information library in the decentralized computer network was explained at the first step of constructing the application technology of ultra-high speed computer to the elucidation of complex phenomena. Establishment of the system makes possible to construct the efficient application environment of ultra-high speed computer system to be scalable with the different computing systems. We named the system Ninf (Network Information Library for High Performance Computing). The summary of application technology of library was described as follows: the application technology of library under the distributed environment, numeric constants, retrieval of value, library of special functions, computing library, Ninf library interface, Ninf remote library and registration. By the system, user is able to use the program concentrating the analyzing technology of numerical value with high precision, reliability and speed. (S.Y.)

  14. Surviving the future academic libraries, quality and assessment

    CERN Document Server

    Munde, Gail

    2009-01-01

    Every academic library strives to make improvements - in its services, its effectiveness, and its contributions to overall university success. Every librarian wants to improve library quality, but few are knowledgeable or enthusiastic about the means and mechanisms of quality improvement. This book assists librarians to make sense of data collection, assessment, and comparative evaluation as stepping stones to transformative quality improvement. Creating value lies in a library's ability to understand, communicate and measure what matters to users, and what can be measured can be managed to successful outcomes. Complex and fragmented subject matter is synthesized into clear and logical presentation Focuses on current research and best practices International in scope.

  15. Interfacing the Paramesh Computational Libraries to the Cactus Computational Framework, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — We will design and implement an interface between the Paramesh computational libraries, developed and used by groups at NASA GSFC, and the Cactus computational...

  16. Cloud computing: An innovative tool for library services

    OpenAIRE

    Sahu, R.

    2015-01-01

    Cloud computing is a new technique of information communication technology because of its potential benefits such as reduced cost, accessible anywhere any time as well as its elasticity and flexibility. In this Paper defines cloud Computing, definition, essential characteristics, model of cloud computing, components of cloud, advantages & drawbacks of cloud computing and also describe cloud computing in libraries.

  17. Library Computing

    Science.gov (United States)

    Library Computing, 1985

    1985-01-01

    Special supplement to "Library Journal" and "School Library Journal" covers topics of interest to school, public, academic, and special libraries planning for automation: microcomputer use, readings in automation, online searching, databases of microcomputer software, public access to microcomputers, circulation, creating a…

  18. Computer Programme Library at Ispra: Service to IAEA Member States

    International Nuclear Information System (INIS)

    1974-01-01

    In 1964, in view of the increasingly important role of computers in atomic energy, the Nuclear Energy Agency of the Organization for Economic Co-operation and Development established a Computer Programme Library (CPL) at Ispra, Italy. The original purpose of this Library, then consisting chiefly of programmes for reactor calculations, was to improve communication between the originators of the computer programmes and the scientists and engineers in Member States of the OECD who used them, thus furthering the efficient and economic use of the many large and expensive computers in the countries concerned. During the first few years of the Library's operation, a growing interest in its activities became evident in countries which were not members of the OECD. The Director General of the International Atomic Energy Agency, aware that such countries which were Member States of the IAEA could benefit from the Library's services, concluded an agreement with the Nuclear Energy Agency in 1967 to the effect that these services would be extended to all IAEA Member States. In March 1968, the Director General announced by circular letter that these services were available, and requested the governments of Member States that were not members of OECD to nominate establishments to participate in the activities of the Computer Programme Library. By August 1973, 32 such establishments were regularly using the services of the CPL

  19. Quality and the academic library reviewing, assessing and enhancing service provision

    CERN Document Server

    2016-01-01

    Quality and the Academic Library: Reviewing, Assessing and Enhancing Service Provision provides an in-depth review and analysis of quality management and service quality in academic libraries. All aspects of quality are considered in the book, including quality assessment, quality review, and quality enhancement. An overview of quality management and service quality concepts, principles, and methods leads to a detailed consideration of how they have been applied in universities and their libraries. A case study approach is used with different perspectives provided from the different stakeholders involved in the quality processes. All contributors adopt a critical reflection approach, reflecting on the implications, impact, and significance of the activities undertaken and the conclusions that can be drawn for future developments. The book concludes with an overall reflection on quality management and service quality in academic libraries with a final analysis of priorities for the future.

  20. Benchmarking, Total Quality Management, and Libraries.

    Science.gov (United States)

    Shaughnessy, Thomas W.

    1993-01-01

    Discussion of the use of Total Quality Management (TQM) in higher education and academic libraries focuses on the identification, collection, and use of reliable data. Methods for measuring quality, including benchmarking, are described; performance measures are considered; and benchmarking techniques are examined. (11 references) (MES)

  1. The use of quality benchmarking in assessing web resources for the dermatology virtual branch library of the National electronic Library for Health (NeLH).

    Science.gov (United States)

    Kamel Boulos, M N; Roudsari, A V; Gordon, C; Muir Gray, J A

    2001-01-01

    In 1998, the U.K. National Health Service Information for Health Strategy proposed the implementation of a National electronic Library for Health to provide clinicians, healthcare managers and planners, patients and the public with easy, round the clock access to high quality, up-to-date electronic information on health and healthcare. The Virtual Branch Libraries are among the most important components of the National electronic Library for Health. They aim at creating online knowledge based communities, each concerned with some specific clinical and other health-related topics. This study is about the envisaged Dermatology Virtual Branch Libraries of the National electronic Library for Health. It aims at selecting suitable dermatology Web resources for inclusion in the forthcoming Virtual Branch Libraries after establishing preliminary quality benchmarking rules for this task. Psoriasis, being a common dermatological condition, has been chosen as a starting point. Because quality is a principal concern of the National electronic Library for Health, the study includes a review of the major quality benchmarking systems available today for assessing health-related Web sites. The methodology of developing a quality benchmarking system has been also reviewed. Aided by metasearch Web tools, candidate resources were hand-selected in light of the reviewed benchmarking systems and specific criteria set by the authors. Over 90 professional and patient-oriented Web resources on psoriasis and dermatology in general are suggested for inclusion in the forthcoming Dermatology Virtual Branch Libraries. The idea of an all-in knowledge-hallmarking instrument for the National electronic Library for Health is also proposed based on the reviewed quality benchmarking systems. Skilled, methodical, organized human reviewing, selection and filtering based on well-defined quality appraisal criteria seems likely to be the key ingredient in the envisaged National electronic Library for

  2. Data-flow oriented visual programming libraries for scientific computing

    NARCIS (Netherlands)

    Maubach, J.M.L.; Drenth, W.D.; Sloot, P.M.A.

    2002-01-01

    The growing release of scientific computational software does not seem to aid the implementation of complex numerical algorithms. Released libraries lack a common standard interface with regard to for instance finite element, difference or volume discretizations. And, libraries written in standard

  3. Orientation and Functions of Library in Quality Education of College

    Science.gov (United States)

    Yang, Lan

    2011-01-01

    Quality education is the core of college education. Libraries are the second class for students due to the extremely important position and function in quality education. Libraries are the best place for cultivating students' morals, the important front for improving students' scientific and cultural qualities, and the effective facilities for…

  4. High-performance computing — an overview

    Science.gov (United States)

    Marksteiner, Peter

    1996-08-01

    An overview of high-performance computing (HPC) is given. Different types of computer architectures used in HPC are discussed: vector supercomputers, high-performance RISC processors, various parallel computers like symmetric multiprocessors, workstation clusters, massively parallel processors. Software tools and programming techniques used in HPC are reviewed: vectorizing compilers, optimization and vector tuning, optimization for RISC processors; parallel programming techniques like shared-memory parallelism, message passing and data parallelism; and numerical libraries.

  5. Computationally optimized deimmunization libraries yield highly mutated enzymes with low immunogenicity and enhanced activity.

    Science.gov (United States)

    Salvat, Regina S; Verma, Deeptak; Parker, Andrew S; Kirsch, Jack R; Brooks, Seth A; Bailey-Kellogg, Chris; Griswold, Karl E

    2017-06-27

    Therapeutic proteins of wide-ranging function hold great promise for treating disease, but immune surveillance of these macromolecules can drive an antidrug immune response that compromises efficacy and even undermines safety. To eliminate widespread T-cell epitopes in any biotherapeutic and thereby mitigate this key source of detrimental immune recognition, we developed a Pareto optimal deimmunization library design algorithm that optimizes protein libraries to account for the simultaneous effects of combinations of mutations on both molecular function and epitope content. Active variants identified by high-throughput screening are thus inherently likely to be deimmunized. Functional screening of an optimized 10-site library (1,536 variants) of P99 β-lactamase (P99βL), a component of ADEPT cancer therapies, revealed that the population possessed high overall fitness, and comprehensive analysis of peptide-MHC II immunoreactivity showed the population possessed lower average immunogenic potential than the wild-type enzyme. Although similar functional screening of an optimized 30-site library (2.15 × 10 9 variants) revealed reduced population-wide fitness, numerous individual variants were found to have activity and stability better than the wild type despite bearing 13 or more deimmunizing mutations per enzyme. The immunogenic potential of one highly active and stable 14-mutation variant was assessed further using ex vivo cellular immunoassays, and the variant was found to silence T-cell activation in seven of the eight blood donors who responded strongly to wild-type P99βL. In summary, our multiobjective library-design process readily identified large and mutually compatible sets of epitope-deleting mutations and produced highly active but aggressively deimmunized constructs in only one round of library screening.

  6. Introduction of library administration system using an office computer in the smallscale library

    International Nuclear Information System (INIS)

    Itabashi, Keizo; Ishikawa, Masashi

    1984-01-01

    Research Information Center was established in Fusion Research Center at Naka site as a new section of Department of Technical Information of Japan Atomic Energy Research Institute. A library materials management system utilizing an office computer was introduced to provide good services. The system is a total-system centered on services at counter except purchase business and the serviced materials are books, reports, journals and pamphlets. The system has produced good effects on many aspects, e.g. a significantly easy inventory of library materials, and complete removal of user's handwriting for borrowing materials, by using an optical chracter recognition handscanner. Those improvements have resulted in better image of the library. (author)

  7. Evaluation of computer-based library services at Kenneth Dike ...

    African Journals Online (AJOL)

    This study evaluated computer-based library services/routines at Kenneth Dike Library, University of Ibadan. Four research questions were developed and answered. A survey research design was adopted; using questionnaire as the instrument for data collection. A total of 200 respondents randomly selected from 10 ...

  8. Managing your library and its quality the ISO 9001 way

    CERN Document Server

    Balagué, Núria

    2011-01-01

    This book, divided into two parts, provides an introduction to the quality management issues and gives a general overview to the use of ISO 9001 in the library environment. The second part presents the main features of ISO 9001:2008 with practical comments and examples on how to implement its clauses in libraries. Whether in the public or in the private sector, libraries can be seen as service organisations: they act in very dynamic environments where users are increasingly demanding new types of services. Thus the adoption of a quality management system helps each library in meeting the needs

  9. Introduction of library administration system using an office computer in the smallscale library

    Energy Technology Data Exchange (ETDEWEB)

    Itabashi, Keizo; Ishikawa, Masashi

    1984-01-01

    Research Information Center was established in Fusion Research Center at Naka site as a new section of Department of Technical Information of Japan Atomic Energy Research Institute. A library materials management system utilizing an office computer was introduced to provide good services. The system is a total-system centered on services at counter except purchase business and the serviced materials are books, reports, journals and pamphlets. The system has produced good effects on many aspects, e.g. a significantly easy inventory of library materials, and complete removal of user's handwriting for borrowing materials, by using an optical chracter recognition handscanner. Those improvements have resulted in better image of the library.

  10. Quality Assurance Strategies for User Friendly School Libraries in ...

    African Journals Online (AJOL)

    This paper examines the methods of school library services with the purpose to promote quality assurance-in- Nigeria school libraries. The recurring theme in this paper is the imperative that school libraries in Nigeria should be committed and contributed to the transformation agenda which is the key issues of equity and ...

  11. The quality infrastructure measuring, analyzing, and improving library services

    CERN Document Server

    Murphy, Sarah Anne

    2013-01-01

    Summarizing specific tools for measuring service quality alongside tips for using these tools most effectively, this book helps libraries of all kinds take a programmatic approach to measuring, analyzing, and improving library services.

  12. Evaluation and quality libraries performance : a view from Portugal and Brazil

    OpenAIRE

    Melo, Luiza Baptista; Sampaio, Maria Imaculada

    2003-01-01

    The main purpose of this poster is to give a view of some projects, on evaluation and quality in libraries performances, which have been developed in Portugal and Brazil during the last ten years. Library services quality is a concept that librarians around of the world has been adopted, that is a way for assessing user needs, satisfaction, and better performance at the libraries. The two academic librarians present a brief description of some methods to obtain performance and quality measure...

  13. Software Quality Assurance and Verification for the MPACT Library Generation Process

    Energy Technology Data Exchange (ETDEWEB)

    Liu, Yuxuan [Univ. of Michigan, Ann Arbor, MI (United States); Williams, Mark L. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Wiarda, Dorothea [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Clarno, Kevin T. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Kim, Kang Seog [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Celik, Cihangir [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2017-05-01

    This report fulfills the requirements for the Consortium for the Advanced Simulation of Light-Water Reactors (CASL) milestone L2:RTM.P14.02, “SQA and Verification for MPACT Library Generation,” by documenting the current status of the software quality, verification, and acceptance testing of nuclear data libraries for MPACT. It provides a brief overview of the library generation process, from general-purpose evaluated nuclear data files (ENDF/B) to a problem-dependent cross section library for modeling of light-water reactors (LWRs). The software quality assurance (SQA) programs associated with each of the software used to generate the nuclear data libraries are discussed; specific tests within the SCALE/AMPX and VERA/XSTools repositories are described. The methods and associated tests to verify the quality of the library during the generation process are described in detail. The library generation process has been automated to a degree to (1) ensure that it can be run without user intervention and (2) to ensure that the library can be reproduced. Finally, the acceptance testing process that will be performed by representatives from the Radiation Transport Methods (RTM) Focus Area prior to the production library’s release is described in detail.

  14. Computer-aided control of high-quality cast iron

    Directory of Open Access Journals (Sweden)

    S. Pietrowski

    2008-04-01

    Full Text Available The study discusses the possibility of control of the high-quality grey cast iron and ductile iron using the author’s genuine computer programs. The programs have been developed with the help of algorithms based on statistical relationships that are said to exist between the characteristic parameters of DTA curves and properties, like Rp0,2, Rm, A5 and HB. It has been proved that the spheroidisation and inoculation treatment of cast iron changes in an important way the characteristic parameters of DTA curves, thus enabling a control of these operations as regards their correctness and effectiveness, along with the related changes in microstructure and mechanical properties of cast iron. Moreover, some examples of statistical relationships existing between the typical properties of ductile iron and its control process were given for cases of the melts consistent and inconsistent with the adopted technology.A test stand for control of the high-quality cast iron and respective melts has been schematically depicted.

  15. Computationally designed libraries for rapid enzyme stabilization

    NARCIS (Netherlands)

    Wijma, Hein J.; Floor, Robert J.; Jekel, Peter A.; Baker, David; Marrink, Siewert J.; Janssen, Dick B.

    The ability to engineer enzymes and other proteins to any desired stability would have wide-ranging applications. Here, we demonstrate that computational design of a library with chemically diverse stabilizing mutations allows the engineering of drastically stabilized and fully functional variants

  16. Employing Virtualization in Library Computing: Use Cases and Lessons Learned

    Directory of Open Access Journals (Sweden)

    Arwen Hutt

    2009-09-01

    Full Text Available This paper provides a broad overview of virtualization technology and describes several examples of its use at the University of California, San Diego Libraries. Libraries can leverage virtualization to address many long-standing library computing challenges, but careful planning is needed to determine if this technology is the right solution for a specific need. This paper outlines both technical and usability considerations, and concludes with a discussion of potential enterprise impacts on the library infrastructure.

  17. Measuring Customer Satisfaction and Quality of Service in Special Libraries.

    Science.gov (United States)

    White, Marilyn Domas; Abels, Eileen G.; Nitecki, Danuta

    This project tested the appropriateness of SERVQUAL (i.e., an instrument widely used in the service industry for assessing service quality based on repeated service encounters rather than a particular service encounter) to measure service quality in special libraries and developed a modified version for special libraries. SERVQUAL is based on an…

  18. The Use of Quality Benchmarking in Assessing Web Resources for the Dermatology Virtual Branch Library of the National electronic Library for Health (NeLH)

    OpenAIRE

    Boulos, MN Kamel; Roudsari, AV; Gordon, C; Gray, JA Muir

    2001-01-01

    Background In 1998, the U.K. National Health Service Information for Health Strategy proposed the implementation of a National electronic Library for Health to provide clinicians, healthcare managers and planners, patients and the public with easy, round the clock access to high quality, up-to-date electronic information on health and healthcare. The Virtual Branch Libraries are among the most important components of the National electronic Library for Health . They aim at creating online kno...

  19. Generalized Portable SHMEM Library for High Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Parzyszek, Krzysztof [Iowa State Univ., Ames, IA (United States)

    2003-01-01

    This dissertation describes the efforts to design and implement the Generalized Portable SHMEM library, GPSHMEM, as well as supplementary tools. There are two major components of the GPSHMEM project: the GPSHMEM library itself and the Fortran 77 source-to-source translator. The rest of this thesis is divided into two parts. Part I introduces the shared memory model and the distributed shared memory model. It explains the motivation behind GPSHMEM and presents its functionality and performance results. Part II is entirely devoted to the Fortran 77 translator call fgpp. The need for such a tool is demonstrated, functionality goals are stated, and the design issues are presented along with the development of the solutions.

  20. Computer Software: Copyright and Licensing Considerations for Schools and Libraries. ERIC Digest.

    Science.gov (United States)

    Reed, Mary Hutchings

    This digest notes that the terms and conditions of computer software package license agreements control the use of software in schools and libraries, and examines the implications of computer software license agreements for classroom use and for library lending policies. Guidelines are provided for interpreting the Copyright Act, and insuring the…

  1. Subroutine library for error estimation of matrix computation (Ver. 1.0)

    International Nuclear Information System (INIS)

    Ichihara, Kiyoshi; Shizawa, Yoshihisa; Kishida, Norio

    1999-03-01

    'Subroutine Library for Error Estimation of Matrix Computation' is a subroutine library which aids the users in obtaining the error ranges of the linear system's solutions or the Hermitian matrices' eigenvalues. This library contains routines for both sequential computers and parallel computers. The subroutines for linear system error estimation calculate norms of residual vectors, matrices's condition numbers, error bounds of solutions and so on. The subroutines for error estimation of Hermitian matrix eigenvalues derive the error ranges of the eigenvalues according to the Korn-Kato's formula. The test matrix generators supply the matrices appeared in the mathematical research, the ones randomly generated and the ones appeared in the application programs. This user's manual contains a brief mathematical background of error analysis on linear algebra and usage of the subroutines. (author)

  2. Cloud Computing Platform for an Online Model Library System

    Directory of Open Access Journals (Sweden)

    Mingang Chen

    2013-01-01

    Full Text Available The rapid developing of digital content industry calls for online model libraries. For the efficiency, user experience, and reliability merits of the model library, this paper designs a Web 3D model library system based on a cloud computing platform. Taking into account complex models, which cause difficulties in real-time 3D interaction, we adopt the model simplification and size adaptive adjustment methods to make the system with more efficient interaction. Meanwhile, a cloud-based architecture is developed to ensure the reliability and scalability of the system. The 3D model library system is intended to be accessible by online users with good interactive experiences. The feasibility of the solution has been tested by experiments.

  3. Automated design of degenerate codon libraries.

    Science.gov (United States)

    Mena, Marco A; Daugherty, Patrick S

    2005-12-01

    Degenerate codon libraries are frequently used in protein engineering and evolution studies but are often limited to targeting a small number of positions to adequately limit the search space. To mitigate this, codon degeneracy can be limited using heuristics or previous knowledge of the targeted positions. To automate design of libraries given a set of amino acid sequences, an algorithm (LibDesign) was developed that generates a set of possible degenerate codon libraries, their resulting size, and their score relative to a user-defined scoring function. A gene library of a specified size can then be constructed that is representative of the given amino acid distribution or that includes specific sequences or combinations thereof. LibDesign provides a new tool for automated design of high-quality protein libraries that more effectively harness existing sequence-structure information derived from multiple sequence alignment or computational protein design data.

  4. High Energy Physics Forum for Computational Excellence: Working Group Reports (I. Applications Software II. Software Libraries and Tools III. Systems)

    Energy Technology Data Exchange (ETDEWEB)

    Habib, Salman [Argonne National Lab. (ANL), Argonne, IL (United States); Roser, Robert [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); LeCompte, Tom [Argonne National Lab. (ANL), Argonne, IL (United States); Marshall, Zach [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Borgland, Anders [SLAC National Accelerator Lab., Menlo Park, CA (United States); Viren, Brett [Brookhaven National Lab. (BNL), Upton, NY (United States); Nugent, Peter [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Asai, Makato [SLAC National Accelerator Lab., Menlo Park, CA (United States); Bauerdick, Lothar [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Finkel, Hal [Argonne National Lab. (ANL), Argonne, IL (United States); Gottlieb, Steve [Indiana Univ., Bloomington, IN (United States); Hoeche, Stefan [SLAC National Accelerator Lab., Menlo Park, CA (United States); Sheldon, Paul [Vanderbilt Univ., Nashville, TN (United States); Vay, Jean-Luc [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Elmer, Peter [Princeton Univ., NJ (United States); Kirby, Michael [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Patton, Simon [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Potekhin, Maxim [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Yanny, Brian [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Calafiura, Paolo [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Dart, Eli [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Gutsche, Oliver [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Izubuchi, Taku [Brookhaven National Lab. (BNL), Upton, NY (United States); Lyon, Adam [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Petravick, Don [Univ. of Illinois, Urbana-Champaign, IL (United States). National Center for Supercomputing Applications (NCSA)

    2015-10-29

    Computing plays an essential role in all aspects of high energy physics. As computational technology evolves rapidly in new directions, and data throughput and volume continue to follow a steep trend-line, it is important for the HEP community to develop an effective response to a series of expected challenges. In order to help shape the desired response, the HEP Forum for Computational Excellence (HEP-FCE) initiated a roadmap planning activity with two key overlapping drivers -- 1) software effectiveness, and 2) infrastructure and expertise advancement. The HEP-FCE formed three working groups, 1) Applications Software, 2) Software Libraries and Tools, and 3) Systems (including systems software), to provide an overview of the current status of HEP computing and to present findings and opportunities for the desired HEP computational roadmap. The final versions of the reports are combined in this document, and are presented along with introductory material.

  5. High Energy Physics Forum for Computational Excellence: Working Group Reports (I. Applications Software II. Software Libraries and Tools III. Systems)

    Energy Technology Data Exchange (ETDEWEB)

    Habib, Salman [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Roser, Robert [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States)

    2015-10-28

    Computing plays an essential role in all aspects of high energy physics. As computational technology evolves rapidly in new directions, and data throughput and volume continue to follow a steep trend-line, it is important for the HEP community to develop an effective response to a series of expected challenges. In order to help shape the desired response, the HEP Forum for Computational Excellence (HEP-FCE) initiated a roadmap planning activity with two key overlapping drivers -- 1) software effectiveness, and 2) infrastructure and expertise advancement. The HEP-FCE formed three working groups, 1) Applications Software, 2) Software Libraries and Tools, and 3) Systems (including systems software), to provide an overview of the current status of HEP computing and to present findings and opportunities for the desired HEP computational roadmap. The final versions of the reports are combined in this document, and are presented along with introductory material.

  6. Experiences Using an Open Source Software Library to Teach Computer Vision Subjects

    Science.gov (United States)

    Cazorla, Miguel; Viejo, Diego

    2015-01-01

    Machine vision is an important subject in computer science and engineering degrees. For laboratory experimentation, it is desirable to have a complete and easy-to-use tool. In this work we present a Java library, oriented to teaching computer vision. We have designed and built the library from the scratch with emphasis on readability and…

  7. Determination of a Screening Metric for High Diversity DNA Libraries.

    Science.gov (United States)

    Guido, Nicholas J; Handerson, Steven; Joseph, Elaine M; Leake, Devin; Kung, Li A

    2016-01-01

    The fields of antibody engineering, enzyme optimization and pathway construction rely increasingly on screening complex variant DNA libraries. These highly diverse libraries allow researchers to sample a maximized sequence space; and therefore, more rapidly identify proteins with significantly improved activity. The current state of the art in synthetic biology allows for libraries with billions of variants, pushing the limits of researchers' ability to qualify libraries for screening by measuring the traditional quality metrics of fidelity and diversity of variants. Instead, when screening variant libraries, researchers typically use a generic, and often insufficient, oversampling rate based on a common rule-of-thumb. We have developed methods to calculate a library-specific oversampling metric, based on fidelity, diversity, and representation of variants, which informs researchers, prior to screening the library, of the amount of oversampling required to ensure that the desired fraction of variant molecules will be sampled. To derive this oversampling metric, we developed a novel alignment tool to efficiently measure frequency counts of individual nucleotide variant positions using next-generation sequencing data. Next, we apply a method based on the "coupon collector" probability theory to construct a curve of upper bound estimates of the sampling size required for any desired variant coverage. The calculated oversampling metric will guide researchers to maximize their efficiency in using highly variant libraries.

  8. Determination of a Screening Metric for High Diversity DNA Libraries.

    Directory of Open Access Journals (Sweden)

    Nicholas J Guido

    Full Text Available The fields of antibody engineering, enzyme optimization and pathway construction rely increasingly on screening complex variant DNA libraries. These highly diverse libraries allow researchers to sample a maximized sequence space; and therefore, more rapidly identify proteins with significantly improved activity. The current state of the art in synthetic biology allows for libraries with billions of variants, pushing the limits of researchers' ability to qualify libraries for screening by measuring the traditional quality metrics of fidelity and diversity of variants. Instead, when screening variant libraries, researchers typically use a generic, and often insufficient, oversampling rate based on a common rule-of-thumb. We have developed methods to calculate a library-specific oversampling metric, based on fidelity, diversity, and representation of variants, which informs researchers, prior to screening the library, of the amount of oversampling required to ensure that the desired fraction of variant molecules will be sampled. To derive this oversampling metric, we developed a novel alignment tool to efficiently measure frequency counts of individual nucleotide variant positions using next-generation sequencing data. Next, we apply a method based on the "coupon collector" probability theory to construct a curve of upper bound estimates of the sampling size required for any desired variant coverage. The calculated oversampling metric will guide researchers to maximize their efficiency in using highly variant libraries.

  9. Business Students' Perception of University Library Service Quality and Satisfaction

    Science.gov (United States)

    Hsu, Maxwell K.; Cummings, Richard G.; Wang, Stephen W.

    2014-01-01

    The main purpose of this study is to examine the college students' perception of library services, and to what extent the quality of library services influences students' satisfaction. The findings depict the relationship between academic libraries and their users in today's digital world and identify critical factors that may sustain a viable…

  10. A QUALITY FUNCTION DEPLOYMENT APPLICATION IN IMPROVING LIBRARY SERVISES: THE CASE OF DUZCE UNIVERSITY CENTRAL LIBRARY

    OpenAIRE

    ÜNAL, Aslıhan; YILDIZ, Mehmet Selami

    2017-01-01

    University libraries have a great importance in accessing information fortheir intermediary role. In today’s world, developments in various areas leadsto new user needs and expectations. Libraries are obliged to improve currentservices and to adapt new developments. The purpose of this research is tocontribute to the improvement of the library services of Duzce University -astate university, was founded in 2006- by following Quality Function Deploymentmethodology. As a result of the research ...

  11. Think Quality! The Deming Approach Does Work in Libraries.

    Science.gov (United States)

    Mackey, Terry; Mackey, Kitty

    1992-01-01

    Presents W. Edwards Deming's Total Quality Management method and advocates its adoption in libraries. The 14 points that form the basis of Deming's philosophy are discussed in the context of the library setting. A flow chart of the reference process and user survey questions are included. (MES)

  12. Construction of high-quality Caco-2 three-frame cDNA library and its application to yeast two-hybrid for the human astrovirus protein-protein interaction.

    Science.gov (United States)

    Zhao, Wei; Li, Xin; Liu, Wen-Hui; Zhao, Jian; Jin, Yi-Ming; Sui, Ting-Ting

    2014-09-01

    Human epithelial colorectal adenocarcinoma (Caco-2) cells are widely used as an in vitro model of the human small intestinal mucosa. Caco-2 cells are host cells of the human astrovirus (HAstV) and other enteroviruses. High quality cDNA libraries are pertinent resources and critical tools for protein-protein interaction research, but are currently unavailable for Caco-2 cells. To construct a three-open reading frame, full length-expression cDNA library from the Caco-2 cell line for application to HAstV protein-protein interaction screening, total RNA was extracted from Caco-2 cells. The switching mechanism at the 5' end of the RNA transcript technique was used for cDNA synthesis. Double-stranded cDNA was digested by Sfi I and ligated to reconstruct a pGADT7-Sfi I three-frame vector. The ligation mixture was transformed into Escherichia coli HST08 premium electro cells by electroporation to construct the primary cDNA library. The library capacity was 1.0×10(6)clones. Gel electrophoresis results indicated that the fragments ranged from 0.5kb to 4.2kb. Randomly picked clones show that the recombination rate was 100%. The three-frame primary cDNA library plasmid mixture (5×10(5)cfu) was also transformed into E. coli HST08 premium electro cells, and all clones were harvested to amplify the cDNA library. To detect the sufficiency of the cDNA library, HAstV capsid protein as bait was screened and tested against the Caco-2 cDNA library by a yeast two-hybrid (Y2H) system. A total of 20 proteins were found to interact with the capsid protein. These results showed that a high-quality three-frame cDNA library from Caco-2 cells was successfully constructed. This library was efficient for the application to the Y2H system, and could be used for future research. Copyright © 2014 Elsevier B.V. All rights reserved.

  13. Large-scale DNA Barcode Library Generation for Biomolecule Identification in High-throughput Screens.

    Science.gov (United States)

    Lyons, Eli; Sheridan, Paul; Tremmel, Georg; Miyano, Satoru; Sugano, Sumio

    2017-10-24

    High-throughput screens allow for the identification of specific biomolecules with characteristics of interest. In barcoded screens, DNA barcodes are linked to target biomolecules in a manner allowing for the target molecules making up a library to be identified by sequencing the DNA barcodes using Next Generation Sequencing. To be useful in experimental settings, the DNA barcodes in a library must satisfy certain constraints related to GC content, homopolymer length, Hamming distance, and blacklisted subsequences. Here we report a novel framework to quickly generate large-scale libraries of DNA barcodes for use in high-throughput screens. We show that our framework dramatically reduces the computation time required to generate large-scale DNA barcode libraries, compared with a naїve approach to DNA barcode library generation. As a proof of concept, we demonstrate that our framework is able to generate a library consisting of one million DNA barcodes for use in a fragment antibody phage display screening experiment. We also report generating a general purpose one billion DNA barcode library, the largest such library yet reported in literature. Our results demonstrate the value of our novel large-scale DNA barcode library generation framework for use in high-throughput screening applications.

  14. The specification of Stampi, a message passing library for distributed parallel computing

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Takemiya, Hiroshi; Koide, Hiroshi

    2000-03-01

    At CCSE, Center for Promotion of Computational Science and Engineering, a new message passing library for heterogeneous and distributed parallel computing has been developed, and it is called as Stampi. Stampi enables us to communicate between any combination of parallel computers as well as workstations. Currently, a Stampi system is constructed from Stampi library and Stampi/Java. It provides functions to connect a Stampi application with not only those on COMPACS, COMplex Parallel Computer System, but also applets which work on WWW browsers. This report summarizes the specifications of Stampi and details the development of its system. (author)

  15. Human Aspects of High Tech in Special Libraries.

    Science.gov (United States)

    Bichteler, Julie

    1986-01-01

    This investigation of library employees who spend significant portion of time in online computer interaction provides information on intellectual, psychological, social, and physical aspects of their work. Long- and short-term effects of special libraries are identified and solutions to "technostress" problems are suggested. (16…

  16. SOA-based digital library services and composition in biomedical applications.

    Science.gov (United States)

    Zhao, Xia; Liu, Enjie; Clapworthy, Gordon J; Viceconti, Marco; Testi, Debora

    2012-06-01

    Carefully collected, high-quality data are crucial in biomedical visualization, and it is important that the user community has ready access to both this data and the high-performance computing resources needed by the complex, computational algorithms that will process it. Biological researchers generally require data, tools and algorithms from multiple providers to achieve their goals. This paper illustrates our response to the problems that result from this. The Living Human Digital Library (LHDL) project presented in this paper has taken advantage of Web Services to build a biomedical digital library infrastructure that allows clinicians and researchers not only to preserve, trace and share data resources, but also to collaborate at the data-processing level. Copyright © 2010 Elsevier Ireland Ltd. All rights reserved.

  17. Quality management as a strategy to support collection development in legal libraries

    Directory of Open Access Journals (Sweden)

    Ana Cláudia Carvalho de Miranda

    2016-05-01

    Full Text Available Libraries are seen as non-profit organizations, their main objective is to offer quality in information mediation, in order to meet the user´s specific information needs. Thus, as organizations, libraries need to apply management strategies that aim to guarantee its effectiveness and efficiency to ensure its survival. Amongst some strategies is quality management. This article aims to propose a debate regarding quality management applied to Collections Development in Legal Libraries. In order to attend these objectives, literature review was realized considering the areas of Quality, Collection Development. In this process, the research identifies the main characteristics applied to Law Libraries, characterizes Legal information as well as points out the main challenges in this area and possible solutions. Based on the reflections given, it can be confirmed that the role of the legal library is to disseminate legal information in the institutions to which they belong to, support legal decisions, work with law in its raw state so it can be interpreted, aiming to reach certain standards contributing to peace in society. It is concluded that for this process to be fulfilled it is necessary the effective implementation of strategies for ensuring the quality of the provided services.

  18. Can the ISO-standards on quality management be useful to libraries and how?

    DEFF Research Database (Denmark)

    Johannsen, Carl Gustav

    1994-01-01

    The aim of this paper is to discuss whether the ISO 9000 concept of quality assurance is useful in a library context or not. First, the basics of the ISO 9000 series are explained. Then, the nature of the concept is discussed and the critical views raised by quality management experts are reviewed...... concluded that the whole discussion on ISO 9000 seems a little futile because the real critical success factor for libraries aiming at Total Quality does not seem to lie in the choice of a specific quality model or guru but more is a question of commitment project on quality assurance in special libraries...

  19. Are Public Libraries Improving Quality of Education? When the Provision of Public Goods Is Not Enough. Policy Research Working Paper 7429

    Science.gov (United States)

    Rodriguez-Lesmes, Paul; Trujillo, Jose Daniel; Valderrama, Daniel

    2015-01-01

    This paper analyzes the relation between public, education-related infrastructure and the quality of education in schools. The analysis uses a case study of the establishment of two large, high-quality public libraries in low-income areas in Bogotá, Colombia. It assesses the impact of these libraries on the quality of education by comparing…

  20. A High-Quality Reference Genome for the Invasive Mosquitofish Gambusia affinis Using a Chicago Library

    Directory of Open Access Journals (Sweden)

    Sandra L. Hoffberg

    2018-06-01

    Full Text Available The western mosquitofish, Gambusia affinis, is a freshwater poecilid fish native to the southeastern United States but with a global distribution due to widespread human introduction. Gambusia affinis has been used as a model species for a broad range of evolutionary and ecological studies. We sequenced the genome of a male G. affinis to facilitate genetic studies in diverse fields including invasion biology and comparative genetics. We generated Illumina short read data from paired-end libraries and in vitro proximity-ligation libraries. We obtained 54.9× coverage, N50 contig length of 17.6 kb, and N50 scaffold length of 6.65 Mb. Compared to two other species in the Poeciliidae family, G. affinis has slightly fewer genes that have shorter total, exon, and intron length on average. Using a set of universal single-copy orthologs in fish genomes, we found 95.5% of these genes were complete in the G. affinis assembly. The number of transposable elements in the G. affinis assembly is similar to those of closely related species. The high-quality genome sequence and annotations we report will be valuable resources for scientists to map the genetic architecture of traits of interest in this species.

  1. A High-Quality Reference Genome for the Invasive Mosquitofish Gambusia affinis Using a Chicago Library.

    Science.gov (United States)

    Hoffberg, Sandra L; Troendle, Nicholas J; Glenn, Travis C; Mahmud, Ousman; Louha, Swarnali; Chalopin, Domitille; Bennetzen, Jeffrey L; Mauricio, Rodney

    2018-04-27

    The western mosquitofish, Gambusia affinis, is a freshwater poecilid fish native to the southeastern United States but with a global distribution due to widespread human introduction. Gambusia affinis has been used as a model species for a broad range of evolutionary and ecological studies. We sequenced the genome of a male G. affinis to facilitate genetic studies in diverse fields including invasion biology and comparative genetics. We generated Illumina short read data from paired-end libraries and in vitro proximity-ligation libraries. We obtained 54.9× coverage, N50 contig length of 17.6 kb, and N50 scaffold length of 6.65 Mb. Compared to two other species in the Poeciliidae family, G. affinis has slightly fewer genes that have shorter total, exon, and intron length on average. Using a set of universal single-copy orthologs in fish genomes, we found 95.5% of these genes were complete in the G. affinis assembly. The number of transposable elements in the G. affinis assembly is similar to those of closely related species. The high-quality genome sequence and annotations we report will be valuable resources for scientists to map the genetic architecture of traits of interest in this species. Copyright © 2018, G3: Genes, Genomes, Genetics.

  2. Software for the computized neutron data library of the SOKRATOR system

    International Nuclear Information System (INIS)

    Kolesov, V.E.; Krivtsov, A.S.; Solov'ev, N.A.

    1976-01-01

    When preparing data for nuclear reactors and shield computations using the evaluated nuclear data library it is necessary to have a set of special service programs to maintain the library itself. In this paper the structure of this set is discussed and a brief description of some programs is presented

  3. Utilization of KENO-IV computer code with HANSEN-ROACH library

    International Nuclear Information System (INIS)

    Lima Barros, M. de; Vellozo, S.O.

    1982-01-01

    Several analysis with KENO-IV computer code, which is based in the Monte Carlo method, and the cross section library HANSEN-ROACH, were done, aiming to present the more convenient form to execute criticality calculations with this computer code and this cross sections. (E.G.) [pt

  4. Developing Quality Managers and Quality Management: The Challenge to Leadership in Library Organizations.

    Science.gov (United States)

    Orenstein, David

    1999-01-01

    Considers the use of Total Quality Management (TQM) tools by library managers to develop quality relationships between staff and customers. Topics include building a shared vision, putting the needs of customers before organizational politics, cooperation between levels of employees, communication, teamwork, trust, redesigning processes and…

  5. Using the Intel Math Kernel Library on Peregrine | High-Performance

    Science.gov (United States)

    Computing | NREL the Intel Math Kernel Library on Peregrine Using the Intel Math Kernel Library on Peregrine Learn how to use the Intel Math Kernel Library (MKL) with Peregrine system software. MKL architectures. Core math functions in MKL include BLAS, LAPACK, ScaLAPACK, sparse solvers, fast Fourier

  6. Chaste: an open source C++ library for computational physiology and biology.

    KAUST Repository

    Mirams, Gary R; Arthurs, Christopher J; Bernabeu, Miguel O; Bordas, Rafel; Cooper, Jonathan; Corrias, Alberto; Davit, Yohan; Dunn, Sara-Jane; Fletcher, Alexander G; Harvey, Daniel G; Marsh, Megan E; Osborne, James M; Pathmanathan, Pras; Pitt-Francis, Joe; Southern, James; Zemzemi, Nejib; Gavaghan, David J

    2013-01-01

    Chaste - Cancer, Heart And Soft Tissue Environment - is an open source C++ library for the computational simulation of mathematical models developed for physiology and biology. Code development has been driven by two initial applications: cardiac electrophysiology and cancer development. A large number of cardiac electrophysiology studies have been enabled and performed, including high-performance computational investigations of defibrillation on realistic human cardiac geometries. New models for the initiation and growth of tumours have been developed. In particular, cell-based simulations have provided novel insight into the role of stem cells in the colorectal crypt. Chaste is constantly evolving and is now being applied to a far wider range of problems. The code provides modules for handling common scientific computing components, such as meshes and solvers for ordinary and partial differential equations (ODEs/PDEs). Re-use of these components avoids the need for researchers to 're-invent the wheel' with each new project, accelerating the rate of progress in new applications. Chaste is developed using industrially-derived techniques, in particular test-driven development, to ensure code quality, re-use and reliability. In this article we provide examples that illustrate the types of problems Chaste can be used to solve, which can be run on a desktop computer. We highlight some scientific studies that have used or are using Chaste, and the insights they have provided. The source code, both for specific releases and the development version, is available to download under an open source Berkeley Software Distribution (BSD) licence at http://www.cs.ox.ac.uk/chaste, together with details of a mailing list and links to documentation and tutorials.

  7. Chaste: an open source C++ library for computational physiology and biology.

    Directory of Open Access Journals (Sweden)

    Gary R Mirams

    Full Text Available Chaste - Cancer, Heart And Soft Tissue Environment - is an open source C++ library for the computational simulation of mathematical models developed for physiology and biology. Code development has been driven by two initial applications: cardiac electrophysiology and cancer development. A large number of cardiac electrophysiology studies have been enabled and performed, including high-performance computational investigations of defibrillation on realistic human cardiac geometries. New models for the initiation and growth of tumours have been developed. In particular, cell-based simulations have provided novel insight into the role of stem cells in the colorectal crypt. Chaste is constantly evolving and is now being applied to a far wider range of problems. The code provides modules for handling common scientific computing components, such as meshes and solvers for ordinary and partial differential equations (ODEs/PDEs. Re-use of these components avoids the need for researchers to 're-invent the wheel' with each new project, accelerating the rate of progress in new applications. Chaste is developed using industrially-derived techniques, in particular test-driven development, to ensure code quality, re-use and reliability. In this article we provide examples that illustrate the types of problems Chaste can be used to solve, which can be run on a desktop computer. We highlight some scientific studies that have used or are using Chaste, and the insights they have provided. The source code, both for specific releases and the development version, is available to download under an open source Berkeley Software Distribution (BSD licence at http://www.cs.ox.ac.uk/chaste, together with details of a mailing list and links to documentation and tutorials.

  8. Chaste: an open source C++ library for computational physiology and biology.

    KAUST Repository

    Mirams, Gary R

    2013-03-14

    Chaste - Cancer, Heart And Soft Tissue Environment - is an open source C++ library for the computational simulation of mathematical models developed for physiology and biology. Code development has been driven by two initial applications: cardiac electrophysiology and cancer development. A large number of cardiac electrophysiology studies have been enabled and performed, including high-performance computational investigations of defibrillation on realistic human cardiac geometries. New models for the initiation and growth of tumours have been developed. In particular, cell-based simulations have provided novel insight into the role of stem cells in the colorectal crypt. Chaste is constantly evolving and is now being applied to a far wider range of problems. The code provides modules for handling common scientific computing components, such as meshes and solvers for ordinary and partial differential equations (ODEs/PDEs). Re-use of these components avoids the need for researchers to \\'re-invent the wheel\\' with each new project, accelerating the rate of progress in new applications. Chaste is developed using industrially-derived techniques, in particular test-driven development, to ensure code quality, re-use and reliability. In this article we provide examples that illustrate the types of problems Chaste can be used to solve, which can be run on a desktop computer. We highlight some scientific studies that have used or are using Chaste, and the insights they have provided. The source code, both for specific releases and the development version, is available to download under an open source Berkeley Software Distribution (BSD) licence at http://www.cs.ox.ac.uk/chaste, together with details of a mailing list and links to documentation and tutorials.

  9. In-House Automation of a Small Library Using a Mainframe Computer.

    Science.gov (United States)

    Waranius, Frances B.; Tellier, Stephen H.

    1986-01-01

    An automated library routine management system was developed in-house to create system unique to the Library and Information Center, Lunar and Planetary Institute, Houston, Texas. A modular approach was used to allow continuity in operations and services as system was implemented. Acronyms and computer accounts and file names are appended.…

  10. McMPI – a managed-code message passing interface library for high performance communication in C#

    OpenAIRE

    Holmes, Daniel John

    2012-01-01

    This work endeavours to achieve technology transfer between established best-practice in academic high-performance computing and current techniques in commercial high-productivity computing. It shows that a credible high-performance message-passing communication library, with semantics and syntax following the Message-Passing Interface (MPI) Standard, can be built in pure C# (one of the .Net suite of computer languages). Message-passing has been the dominant paradigm in high-pe...

  11. Increasing the use of 'smart' pump drug libraries by nurses: a continuous quality improvement project.

    Science.gov (United States)

    Harding, Andrew D

    2012-01-01

    The use of infusion pumps that incorporate "smart" technology (smart pumps) can reduce the risks associated with receiving IV therapies. Smart pump technology incorporates safeguards such as a list of high-alert medications, soft and hard dosage limits, and a drug library that can be tailored to specific patient care areas. Its use can help to improve patient safety and to avoid potentially catastrophic harm associated with medication errors. But when one independent community hospital in Massachusetts switched from older mechanical pumps to smart pumps, it neglected to assign an "owner" to oversee the implementation process. One result was that nurses were using the smart pump library for only 37% of all infusions.To increase pump library usage percentage-thereby reducing the risks associated with infusion and improving patient safety-the hospital undertook a continuous quality improvement project over a four-month period in 2009. With the involvement of direct care nurses, and using quantitative data available from the smart pump software, the nursing quality and pharmacy quality teams identified ways to improve pump and pump library use. A secondary goal was to calculate the hospital's return on investment for the purchase of the smart pumps. Several interventions were developed and, on the first of each month, implemented. By the end of the project, pump library usage had nearly doubled; and the hospital had completely recouped its initial investment.

  12. University Rankings: How Well Do They Measure Library Service Quality?

    Science.gov (United States)

    Jackson, Brian

    2015-01-01

    University rankings play an increasingly large role in shaping the goals of academic institutions and departments, while removing universities themselves from the evaluation process. This study compares the library-related results of two university ranking publications with scores on the LibQUAL+™ survey to identify if library service quality--as…

  13. A Computer Library for Ray Tracing in Analytical Media

    International Nuclear Information System (INIS)

    Miqueles, Eduardo; Coimbra, Tiago A; Figueiredo, J J S de

    2013-01-01

    Ray tracing technique is an important tool not only for forward but also for inverse problems in Geophysics, which most of the seismic processing steps depends on. However, implementing ray tracing codes can be very time consuming. This article presents a computer library to trace rays in 2.5D media composed by stack of layers. The velocity profile inside each layer is such that the eikonal equation can be analitically solved. Therefore, the ray tracing within such profile is made fast and accurately. The great advantage of an analytical ray tracing library is the numerical precision of the quantities computed and the fast execution of the implemented codes. Although ray tracing programs already exist for a long time, for example the seis package by Cervený, with a numerical approach to compute the ray. Regardless of the fact that numerical methods can solve more general problems, the analytical ones could be part of a more sofisticated simulation process, where the ray tracing time is completely relevant. We demonstrate the feasibility of our codes using numerical examples.

  14. Total Quality Management in Libraries: A Sourcebook.

    Science.gov (United States)

    O'Neil, Rosanna M., Comp.

    Total Quality Management (TQM) brings together the best aspects of organizational excellence by driving out fear, offering customer-driven products and services, doing it right the first time by eliminating error, and maintaining inventory control without waste. Libraries are service organizations which are constantly trying to improve service.…

  15. Computer Simulation of the Circulation Subsystem of a Library

    Science.gov (United States)

    Shaw, W. M., Jr.

    1975-01-01

    When circulation data are used as input parameters for a computer simulation of a library's circulation subsystem, the results of the simulation provide information on book availability and delays. The model may be used to simulate alternative loan policies. (Author/LS)

  16. Computer tape library for collective dose assessments in the European Community

    International Nuclear Information System (INIS)

    Jones, J.A.; Kelly, G.N.

    1981-01-01

    A computer tape library has been established which can readily be applied to enable calculations to be made of the collective dose to the population of the European Community from radioactive effluent discharges within the EC. The background to the development of the tape library and details of its contents, method of application and availability are summarised in this note. (author)

  17. Usage of Academic Libraries : The Role of Service Quality, Resources, and User Characteristics

    OpenAIRE

    Hamad Ben Ibrahim Al Omran - Translation

    2004-01-01

    A survival study to 3 academic libraries located in Pennsylvania , it aiming at measuring the effects of service quality and users profiles on the usage of academic libraries. the study contained 188 users from the 3 libraries.

  18. Service Quality: An Unobtrusive Investigation of Interlibrary Loan in Large Public Libraries in Canada.

    Science.gov (United States)

    Hebert, Francoise

    1994-01-01

    Describes a study that investigated the quality of interlibrary loan services in Canadian public libraries from the library's and the user's perspectives and then compared results. Measures of interlibrary loan performance are reviewed; an alternative conceptualization of service quality is discussed; and SERVQUAL, a measure of service quality, is…

  19. The Content of U .S. Library's Total Quality Management and Its Developmental Barriers

    Directory of Open Access Journals (Sweden)

    Chen-Yung Fan

    1998-12-01

    Full Text Available The total quality management (TQM is a process, a technique, a management style, a goal and a tool. Most important of all, TQM means quality service. However, these are concepts largely unfamiliar to libraries but which seem to have strong potential to offer new insights and strategies in library management. This paper attempts toexpound the theoretical advantages of TQM and the barriers of its application in U.S. libraries. [Article content in Chinese

  20. The Scythe Statistical Library: An Open Source C++ Library for Statistical Computation

    Directory of Open Access Journals (Sweden)

    Daniel Pemstein

    2011-08-01

    Full Text Available The Scythe Statistical Library is an open source C++ library for statistical computation. It includes a suite of matrix manipulation functions, a suite of pseudo-random number generators, and a suite of numerical optimization routines. Programs written using Scythe are generally much faster than those written in commonly used interpreted languages, such as R and proglang{MATLAB}; and can be compiled on any system with the GNU GCC compiler (and perhaps with other C++ compilers. One of the primary design goals of the Scythe developers has been ease of use for non-expert C++ programmers. Ease of use is provided through three primary mechanisms: (1 operator and function over-loading, (2 numerous pre-fabricated utility functions, and (3 clear documentation and example programs. Additionally, Scythe is quite flexible and entirely extensible because the source code is available to all users under the GNU General Public License.

  1. Usage of Academic Libraries : The Role of Service Quality, Resources, and User Characteristics

    Directory of Open Access Journals (Sweden)

    Hamad Ben Ibrahim Al Omran - Translation

    2004-12-01

    Full Text Available A survival study to 3 academic libraries located in Pennsylvania , it aiming at measuring the effects of service quality and users profiles on the usage of academic libraries. the study contained 188 users from the 3 libraries.

  2. Development library of finite elements for computer-aided design system of reed sensors

    Science.gov (United States)

    Kozlov, A. S.; Shmakov, N. A.; Tkalich, V. L.; Labkovskaia, R. I.; Kalinkina, M. E.; Pirozhnikova, O. I.

    2018-05-01

    The article is devoted to the development of a modern highly reliable element base of devices for security and fire alarm systems, in particular, to the improvement of the quality of contact cores (reed and membrane) of reed sensors. Modeling of elastic sensitive elements uses quadrangular elements of plates and shells, considered in the system of curvilinear orthogonal coordinates. The developed mathematical models and the formed finite element library are designed for systems of automated design of reed switch detectors to create competitive devices alarms. The finite element library is used for the automated system production of reed switch detectors both in series production and in the implementation of individual orders.

  3. Mass Digitisation by Libraries: Issues concerning Organisation, Quality and Efficiency

    Directory of Open Access Journals (Sweden)

    Astrid Verheusen

    2008-03-01

    Full Text Available Ever since the world-wide web made it possible to display graphics on the Internet, libraries have been scanning their older documents and pictures to provide access to them. From the middle of the 1990s thousands of libraries of all sizes began scanning parts of their collections, provided these with metadata and made them available on the web. The emphasis in these first, rather small, digitisation projects was on experimenting with different techniques for both scanning and building interfaces for the Internet. Along the way, methods for quality assurance, project management and business models became more professional. In line with the progress made in the field of digitisation, a profound knowledge of best practices has been developed. However, this knowledge is not available for all cultural heritage institutions who want to digitise their collections. Most of the smaller institutions lack experience and, moreover, the means to digitise in an efficient way. At the same time, the larger libraries are moving towards large-scale digitisation of historical texts while Google has already digitised millions of books from several libraries around the world. Although many libraries welcome the unprecedented access to all this information, Google has also been criticized for the inferior quality of their images, the emphasis on the English language, the violation of copyright laws and the lack of attention for preservation issues. The question therefore arises: can libraries do better than Google?

  4. Real-time computer treatment of THz passive device images with the high image quality

    Science.gov (United States)

    Trofimov, Vyacheslav A.; Trofimov, Vladislav V.

    2012-06-01

    We demonstrate real-time computer code improving significantly the quality of images captured by the passive THz imaging system. The code is not only designed for a THz passive device: it can be applied to any kind of such devices and active THz imaging systems as well. We applied our code for computer processing of images captured by four passive THz imaging devices manufactured by different companies. It should be stressed that computer processing of images produced by different companies requires using the different spatial filters usually. The performance of current version of the computer code is greater than one image per second for a THz image having more than 5000 pixels and 24 bit number representation. Processing of THz single image produces about 20 images simultaneously corresponding to various spatial filters. The computer code allows increasing the number of pixels for processed images without noticeable reduction of image quality. The performance of the computer code can be increased many times using parallel algorithms for processing the image. We develop original spatial filters which allow one to see objects with sizes less than 2 cm. The imagery is produced by passive THz imaging devices which captured the images of objects hidden under opaque clothes. For images with high noise we develop an approach which results in suppression of the noise after using the computer processing and we obtain the good quality image. With the aim of illustrating the efficiency of the developed approach we demonstrate the detection of the liquid explosive, ordinary explosive, knife, pistol, metal plate, CD, ceramics, chocolate and other objects hidden under opaque clothes. The results demonstrate the high efficiency of our approach for the detection of hidden objects and they are a very promising solution for the security problem.

  5. A practical introduction to computer vision with OpenCV

    CERN Document Server

    Dawson-Howe, Kenneth

    2014-01-01

    Explains the theory behind basic computer vision and provides a bridge from the theory to practical implementation using the industry standard OpenCV libraries Computer Vision is a rapidly expanding area and it is becoming progressively easier for developers to make use of this field due to the ready availability of high quality libraries (such as OpenCV 2).  This text is intended to facilitate the practical use of computer vision with the goal being to bridge the gap between the theory and the practical implementation of computer vision. The book will explain how to use the relevant OpenCV

  6. Computational design of chimeric protein libraries for directed evolution.

    Science.gov (United States)

    Silberg, Jonathan J; Nguyen, Peter Q; Stevenson, Taylor

    2010-01-01

    The best approach for creating libraries of functional proteins with large numbers of nondisruptive amino acid substitutions is protein recombination, in which structurally related polypeptides are swapped among homologous proteins. Unfortunately, as more distantly related proteins are recombined, the fraction of variants having a disrupted structure increases. One way to enrich the fraction of folded and potentially interesting chimeras in these libraries is to use computational algorithms to anticipate which structural elements can be swapped without disturbing the integrity of a protein's structure. Herein, we describe how the algorithm Schema uses the sequences and structures of the parent proteins recombined to predict the structural disruption of chimeras, and we outline how dynamic programming can be used to find libraries with a range of amino acid substitution levels that are enriched in variants with low Schema disruption.

  7. Towards better digital pathology workflows: programming libraries for high-speed sharpness assessment of Whole Slide Images.

    Science.gov (United States)

    Ameisen, David; Deroulers, Christophe; Perrier, Valérie; Bouhidel, Fatiha; Battistella, Maxime; Legrès, Luc; Janin, Anne; Bertheau, Philippe; Yunès, Jean-Baptiste

    2014-01-01

    Since microscopic slides can now be automatically digitized and integrated in the clinical workflow, quality assessment of Whole Slide Images (WSI) has become a crucial issue. We present a no-reference quality assessment method that has been thoroughly tested since 2010 and is under implementation in multiple sites, both public university-hospitals and private entities. It is part of the FlexMIm R&D project which aims to improve the global workflow of digital pathology. For these uses, we have developed two programming libraries, in Java and Python, which can be integrated in various types of WSI acquisition systems, viewers and image analysis tools. Development and testing have been carried out on a MacBook Pro i7 and on a bi-Xeon 2.7GHz server. Libraries implementing the blur assessment method have been developed in Java, Python, PHP5 and MySQL5. For web applications, JavaScript, Ajax, JSON and Sockets were also used, as well as the Google Maps API. Aperio SVS files were converted into the Google Maps format using VIPS and Openslide libraries. We designed the Java library as a Service Provider Interface (SPI), extendable by third parties. Analysis is computed in real-time (3 billion pixels per minute). Tests were made on 5000 single images, 200 NDPI WSI, 100 Aperio SVS WSI converted to the Google Maps format. Applications based on our method and libraries can be used upstream, as calibration and quality control tool for the WSI acquisition systems, or as tools to reacquire tiles while the WSI is being scanned. They can also be used downstream to reacquire the complete slides that are below the quality threshold for surgical pathology analysis. WSI may also be displayed in a smarter way by sending and displaying the regions of highest quality before other regions. Such quality assessment scores could be integrated as WSI's metadata shared in clinical, research or teaching contexts, for a more efficient medical informatics workflow.

  8. Towards better digital pathology workflows: programming libraries for high-speed sharpness assessment of Whole Slide Images

    Science.gov (United States)

    2014-01-01

    Background Since microscopic slides can now be automatically digitized and integrated in the clinical workflow, quality assessment of Whole Slide Images (WSI) has become a crucial issue. We present a no-reference quality assessment method that has been thoroughly tested since 2010 and is under implementation in multiple sites, both public university-hospitals and private entities. It is part of the FlexMIm R&D project which aims to improve the global workflow of digital pathology. For these uses, we have developed two programming libraries, in Java and Python, which can be integrated in various types of WSI acquisition systems, viewers and image analysis tools. Methods Development and testing have been carried out on a MacBook Pro i7 and on a bi-Xeon 2.7GHz server. Libraries implementing the blur assessment method have been developed in Java, Python, PHP5 and MySQL5. For web applications, JavaScript, Ajax, JSON and Sockets were also used, as well as the Google Maps API. Aperio SVS files were converted into the Google Maps format using VIPS and Openslide libraries. Results We designed the Java library as a Service Provider Interface (SPI), extendable by third parties. Analysis is computed in real-time (3 billion pixels per minute). Tests were made on 5000 single images, 200 NDPI WSI, 100 Aperio SVS WSI converted to the Google Maps format. Conclusions Applications based on our method and libraries can be used upstream, as calibration and quality control tool for the WSI acquisition systems, or as tools to reacquire tiles while the WSI is being scanned. They can also be used downstream to reacquire the complete slides that are below the quality threshold for surgical pathology analysis. WSI may also be displayed in a smarter way by sending and displaying the regions of highest quality before other regions. Such quality assessment scores could be integrated as WSI's metadata shared in clinical, research or teaching contexts, for a more efficient medical informatics

  9. PuLSE: Quality control and quantification of peptide sequences explored by phage display libraries.

    Science.gov (United States)

    Shave, Steven; Mann, Stefan; Koszela, Joanna; Kerr, Alastair; Auer, Manfred

    2018-01-01

    The design of highly diverse phage display libraries is based on assumption that DNA bases are incorporated at similar rates within the randomized sequence. As library complexity increases and expected copy numbers of unique sequences decrease, the exploration of library space becomes sparser and the presence of truly random sequences becomes critical. We present the program PuLSE (Phage Library Sequence Evaluation) as a tool for assessing randomness and therefore diversity of phage display libraries. PuLSE runs on a collection of sequence reads in the fastq file format and generates tables profiling the library in terms of unique DNA sequence counts and positions, translated peptide sequences, and normalized 'expected' occurrences from base to residue codon frequencies. The output allows at-a-glance quantitative quality control of a phage library in terms of sequence coverage both at the DNA base and translated protein residue level, which has been missing from toolsets and literature. The open source program PuLSE is available in two formats, a C++ source code package for compilation and integration into existing bioinformatics pipelines and precompiled binaries for ease of use.

  10. OpenCL-Based Linear Algebra Libraries for High-Performance Computing, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — Despite its promise, OpenCL adoption is slow, owing to a lack of libraries and tools. Vendors have shown few signs of plans to provide OpenCL libraries, and were...

  11. User Requirements Analysis For Digital Library Application Using Quality Function Deployment.

    Science.gov (United States)

    Wulandari, Lily; Sularto, Lana; Yusnitasari, Tristyanti; Ikasari, Diana

    2017-03-01

    This study attemp to build Smart Digital Library to be used by the wider community wherever they are. The system is built in the form of Smart Digital Library portal which uses semantic similarity method (Semantic Similarity) to search journals, articles or books by title or author name. This method is also used to determine the recommended books to be read by visitors of Smart Digital Library based on testimony from a previous reader automatically. Steps being taken in the development of Smart Digital Library system is the analysis phase, design phase, testing and implementation phase. At this stage of the analysis using WebQual for the preparation of the instruments to be distributed to the respondents and the data obtained from the respondents will be processed using Quality Function Deployment. In the analysis phase has the purpose of identifying consumer needs and technical requirements. The analysis was performed to a digital library on the web digital library Gunadarma University, Bogor Institute of Agriculture, University of Indonesia, etc. The questionnaire was distributed to 200 respondents. The research methodology begins with the collection of user requirements and analyse it using QFD. Application design is funded by the government through a program of Featured Universities Research by the Directorate General of Higher Education (DIKTI). Conclusions from this research are identified which include the Consumer Requirements of digital library application. The elements of the consumers requirements consists of 13 elements and 25 elements of Engineering Characteristics digital library requirements. Therefore the design of digital library applications that will be built, is designed according to the findings by eliminating features that are not needed by restaurant based on QFD House of Quality.

  12. Library Design-Facilitated High-Throughput Sequencing of Synthetic Peptide Libraries.

    Science.gov (United States)

    Vinogradov, Alexander A; Gates, Zachary P; Zhang, Chi; Quartararo, Anthony J; Halloran, Kathryn H; Pentelute, Bradley L

    2017-11-13

    A methodology to achieve high-throughput de novo sequencing of synthetic peptide mixtures is reported. The approach leverages shotgun nanoliquid chromatography coupled with tandem mass spectrometry-based de novo sequencing of library mixtures (up to 2000 peptides) as well as automated data analysis protocols to filter away incorrect assignments, noise, and synthetic side-products. For increasing the confidence in the sequencing results, mass spectrometry-friendly library designs were developed that enabled unambiguous decoding of up to 600 peptide sequences per hour while maintaining greater than 85% sequence identification rates in most cases. The reliability of the reported decoding strategy was additionally confirmed by matching fragmentation spectra for select authentic peptides identified from library sequencing samples. The methods reported here are directly applicable to screening techniques that yield mixtures of active compounds, including particle sorting of one-bead one-compound libraries and affinity enrichment of synthetic library mixtures performed in solution.

  13. HEPLIB '91: International users meeting on the support and environments of high energy physics computing

    International Nuclear Information System (INIS)

    Johnstad, H.

    1991-01-01

    The purpose of this meeting is to discuss the current and future HEP computing support and environments from the perspective of new horizons in accelerator, physics, and computing technologies. Topics of interest to the Meeting include (but are limited to): the forming of the HEPLIB world user group for High Energy Physic computing; mandate, desirables, coordination, organization, funding; user experience, international collaboration; the roles of national labs, universities, and industry; range of software, Monte Carlo, mathematics, physics, interactive analysis, text processors, editors, graphics, data base systems, code management tools; program libraries, frequency of updates, distribution; distributed and interactive computing, data base systems, user interface, UNIX operating systems, networking, compilers, Xlib, X-Graphics; documentation, updates, availability, distribution; code management in large collaborations, keeping track of program versions; and quality assurance, testing, conventions, standards

  14. Learning openCV computer vision with the openCV library

    CERN Document Server

    Bradski, Gary

    2008-01-01

    Learning OpenCV puts you right in the middle of the rapidly expanding field of computer vision. Written by the creators of OpenCV, the widely used free open-source library, this book introduces you to computer vision and demonstrates how you can quickly build applications that enable computers to see" and make decisions based on the data. With this book, any developer or hobbyist can get up and running with the framework quickly, whether it's to build simple or sophisticated vision applications

  15. Multicapillary SDS-gel electrophoresis for the analysis of fluorescently labeled mAb preparations: a high throughput quality control process for the production of QuantiPlasma and PlasmaScan mAb libraries.

    Science.gov (United States)

    Székely, Andrea; Szekrényes, Akos; Kerékgyártó, Márta; Balogh, Attila; Kádas, János; Lázár, József; Guttman, András; Kurucz, István; Takács, László

    2014-08-01

    Molecular heterogeneity of mAb preparations is the result of various co- and post-translational modifications and to contaminants related to the production process. Changes in molecular composition results in alterations of functional performance, therefore quality control and validation of therapeutic or diagnostic protein products is essential. A special case is the consistent production of mAb libraries (QuantiPlasma™ and PlasmaScan™) for proteome profiling, quality control of which represents a challenge because of high number of mAbs (>1000). Here, we devise a generally applicable multicapillary SDS-gel electrophoresis process for the analysis of fluorescently labeled mAb preparations for the high throughput quality control of mAbs of the QuantiPlasma™ and PlasmaScan™ libraries. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  16. Library of subroutines to produce one- and two-dimensional statistical distributions on the ES-1010 computer

    International Nuclear Information System (INIS)

    Vzorov, I.K.; Ivanov, V.V.

    1978-01-01

    A library of subroutines to produce 1- and 2-dimensional distribution on the ES-1010 computer is described. 1-dimensional distribution is represented as the histogram, 2-dimensional one is represented as the table. The library provides such opportunities as booking and deleting, filling and clearing histograms (tables), arithmetic operations with them, and printing histograms (tables) on the computer printer with variable printer line. All subroutines are written in FORTRAN-4 language and can be called from the program written in FORTRAN or in ASSEMBLER. This library can be implemented on all computer systems that offer a FORTRAN-4 compiler

  17. High-throughput identification of higher-κ dielectrics from an amorphous N2-doped HfO2–TiO2 library

    International Nuclear Information System (INIS)

    Chang, K.-S.; Lu, W.-C.; Wu, C.-Y.; Feng, H.-C.

    2014-01-01

    Highlights: • Amorphous N 2 -doped HfO 2 –TiO 2 libraries were fabricated using sputtering. • Structure and quality of the dielectric and interfacial layers were investigated. • κ (54), J L < 10 −6 A/cm 2 , and equivalent oxide thickness (1 nm) were identified. - Abstract: High-throughput sputtering was used to fabricate high-quality, amorphous, thin HfO 2 –TiO 2 and N 2 -doped HfO 2 –TiO 2 (HfON–TiON) gate dielectric libraries. Electron probe energy dispersive spectroscopy was used to investigate the structures, compositions, and qualities of the dielectric and interfacial layers of these libraries to determine their electrical properties. A κ value of approximately 54, a leakage current density <10 −6 A/cm 2 , and an equivalent oxide thickness of approximately 1 nm were identified in an HfON–TiON library within a composition range of 68–80 at.% Ti. This library exhibits promise for application in highly advanced metal–oxide–semiconductor (higher-κ) gate stacks

  18. High Res at High Speed: Automated Delivery of High-Resolution Images from Digital Library Collections

    Science.gov (United States)

    Westbrook, R. Niccole; Watkins, Sean

    2012-01-01

    As primary source materials in the library are digitized and made available online, the focus of related library services is shifting to include new and innovative methods of digital delivery via social media, digital storytelling, and community-based and consortial image repositories. Most images on the Web are not of sufficient quality for most…

  19. Cloud Computing Quality

    Directory of Open Access Journals (Sweden)

    Anamaria Şiclovan

    2013-02-01

    Full Text Available Cloud computing was and it will be a new way of providing Internet services and computers. This calculation approach is based on many existing services, such as the Internet, grid computing, Web services. Cloud computing as a system aims to provide on demand services more acceptable as price and infrastructure. It is exactly the transition from computer to a service offered to the consumers as a product delivered online. This paper is meant to describe the quality of cloud computing services, analyzing the advantages and characteristics offered by it. It is a theoretical paper.Keywords: Cloud computing, QoS, quality of cloud computing

  20. Manual for Assessing the Quality of Health Sciences Libraries in Hospitals.

    Science.gov (United States)

    Bain, Christine A.; And Others

    In compliance with legislation enacted by New York State in 1981, a manual of recommendations and standards for hospital libraries was developed (1) to assist in assessment of the quality of hospital library services; (2) to enable measurement of the improvement in individual institutions; and (3) to be used as criteria in the administration of…

  1. Robust DNA Isolation and High-throughput Sequencing Library Construction for Herbarium Specimens.

    Science.gov (United States)

    Saeidi, Saman; McKain, Michael R; Kellogg, Elizabeth A

    2018-03-08

    Herbaria are an invaluable source of plant material that can be used in a variety of biological studies. The use of herbarium specimens is associated with a number of challenges including sample preservation quality, degraded DNA, and destructive sampling of rare specimens. In order to more effectively use herbarium material in large sequencing projects, a dependable and scalable method of DNA isolation and library preparation is needed. This paper demonstrates a robust, beginning-to-end protocol for DNA isolation and high-throughput library construction from herbarium specimens that does not require modification for individual samples. This protocol is tailored for low quality dried plant material and takes advantage of existing methods by optimizing tissue grinding, modifying library size selection, and introducing an optional reamplification step for low yield libraries. Reamplification of low yield DNA libraries can rescue samples derived from irreplaceable and potentially valuable herbarium specimens, negating the need for additional destructive sampling and without introducing discernible sequencing bias for common phylogenetic applications. The protocol has been tested on hundreds of grass species, but is expected to be adaptable for use in other plant lineages after verification. This protocol can be limited by extremely degraded DNA, where fragments do not exist in the desired size range, and by secondary metabolites present in some plant material that inhibit clean DNA isolation. Overall, this protocol introduces a fast and comprehensive method that allows for DNA isolation and library preparation of 24 samples in less than 13 h, with only 8 h of active hands-on time with minimal modifications.

  2. Implementing of AMPX-II system for a univac computer neutron cross-section libraries

    International Nuclear Information System (INIS)

    Sancho, J.; Verdu, G.; Serradell, V.

    1984-01-01

    The AMPX-II system, developed at ORNL, is constituted by a modular set of computer programs, for generation and handling of several nuclear data libraries. The processing starts from ENDF/B library. Along this paper, we refer mainly to the modules related with neutron cross section libraries: master, working and weighted. These modules have been implemented recently for a UNIVAC 1100/60 computer in the Universidad Politecnica de Valencia (Spain). In order to run the programs in that machine it has been necessary to introduce a number of modifications into their programing structure. The main difficulties found in this work and the need of verification for the new versions are also pointed out. We also refer to the results obtained from the execution of a set of little sample problems. (author)

  3. Computational hybrid anthropometric paediatric phantom library for internal radiation dosimetry

    DEFF Research Database (Denmark)

    Xie, Tianwu; Kuster, Niels; Zaidi, Habib

    2017-01-01

    for children demonstrated that they follow the same trend when correlated with age. The constructed hybrid computational phantom library opens up the prospect of comprehensive radiation dosimetry calculations and risk assessment for the paediatric population of different age groups and diverse anthropometric...

  4. Abstracts of computer programs and data libraries pertaining to photon production data

    Energy Technology Data Exchange (ETDEWEB)

    White, J.E.; Manneschmidt, J.B.; Finch, S.Y.; Dickens, J.K.

    1998-06-01

    Abstracts, or descriptions, of computer programs and data libraries pertaining to Photon Production Data (Measurements, Evaluations and Calculations) maintained in the collections of the Radiation Safety Information Computational Center, Oak Ridge, Tennessee USA and at the OECD/NEA Data Bank, Paris, are collected in this document.

  5. Abstracts of computer programs and data libraries pertaining to photon production data

    International Nuclear Information System (INIS)

    White, J.E.; Manneschmidt, J.B.; Finch, S.Y.; Dickens, J.K.

    1998-01-01

    Abstracts, or descriptions, of computer programs and data libraries pertaining to Photon Production Data (Measurements, Evaluations and Calculations) maintained in the collections of the Radiation Safety Information Computational Center, Oak Ridge, Tennessee USA and at the OECD/NEA Data Bank, Paris, are collected in this document

  6. Applying Machine Learning and High Performance Computing to Water Quality Assessment and Prediction

    Directory of Open Access Journals (Sweden)

    Ruijian Zhang

    2017-12-01

    Full Text Available Water quality assessment and prediction is a more and more important issue. Traditional ways either take lots of time or they can only do assessments. In this research, by applying machine learning algorithm to a long period time of water attributes’ data; we can generate a decision tree so that it can predict the future day’s water quality in an easy and efficient way. The idea is to combine the traditional ways and the computer algorithms together. Using machine learning algorithms, the assessment of water quality will be far more efficient, and by generating the decision tree, the prediction will be quite accurate. The drawback of the machine learning modeling is that the execution takes quite long time, especially when we employ a better accuracy but more time-consuming algorithm in clustering. Therefore, we applied the high performance computing (HPC System to deal with this problem. Up to now, the pilot experiments have achieved very promising preliminary results. The visualized water quality assessment and prediction obtained from this project would be published in an interactive website so that the public and the environmental managers could use the information for their decision making.

  7. Library Automation Report, 1996. Multimedia Computers in U.S. Public Schools, 1995-96.

    Science.gov (United States)

    Quality Education Data, Inc., Denver, CO.

    District library media directors face dual demands including competition for limited educational dollars and the need to meet increasingly sophisticated student research requests. To solve these dilemmas, many districts are automating their schools' library media centers. Quality Education Data (QED) is an education research firm providing…

  8. Debunking the Computer Science Digital Library: Lessons Learned in Collection Development at Seneca College of Applied Arts & Technology

    Science.gov (United States)

    Buczynski, James Andrew

    2005-01-01

    Developing a library collection to support the curriculum of Canada's largest computer studies school has debunked many myths about collecting computer science and technology information resources. Computer science students are among the heaviest print book and e-book users in the library. Circulation statistics indicate that the demand for print…

  9. Arc4nix: A cross-platform geospatial analytical library for cluster and cloud computing

    Science.gov (United States)

    Tang, Jingyin; Matyas, Corene J.

    2018-02-01

    Big Data in geospatial technology is a grand challenge for processing capacity. The ability to use a GIS for geospatial analysis on Cloud Computing and High Performance Computing (HPC) clusters has emerged as a new approach to provide feasible solutions. However, users lack the ability to migrate existing research tools to a Cloud Computing or HPC-based environment because of the incompatibility of the market-dominating ArcGIS software stack and Linux operating system. This manuscript details a cross-platform geospatial library "arc4nix" to bridge this gap. Arc4nix provides an application programming interface compatible with ArcGIS and its Python library "arcpy". Arc4nix uses a decoupled client-server architecture that permits geospatial analytical functions to run on the remote server and other functions to run on the native Python environment. It uses functional programming and meta-programming language to dynamically construct Python codes containing actual geospatial calculations, send them to a server and retrieve results. Arc4nix allows users to employ their arcpy-based script in a Cloud Computing and HPC environment with minimal or no modification. It also supports parallelizing tasks using multiple CPU cores and nodes for large-scale analyses. A case study of geospatial processing of a numerical weather model's output shows that arcpy scales linearly in a distributed environment. Arc4nix is open-source software.

  10. Library Assessment and Quality Assurance - Creating a Staff-Driven and User-Focused Development Process

    Directory of Open Access Journals (Sweden)

    Håkan Carlsson

    2016-09-01

    Full Text Available Objective – Gothenburg University Library has implemented a process with the goal to combine quality assurance and strategic planning activities. The process has bottom-up and top-down features designed to generate strong staff-involvement and long-term strategic stability. Methods – In 2008 the library started implementing a system in which each library team should state a number of improvement activities for the upcoming year. In order to focus the efforts, the system has gradually been improved by closely coupling a number of assessment activities, such as surveys and statistics, and connecting the activities to the long-term strategic plan of the library. Results – The activities of the library are now more systematically guided by both library staff and users. The system has resulted in increased understanding within different staff groups of changing external and internal demands, as well as the need for continuous change to library activities. Conclusion – Library assessment and external intelligence are important for tracking and improving library activities. Quality assurance and strategic planning are intricate parts in sustainable development of better and more effective services. The process becomes more effective when staff-driven and built upon systematic knowledge of present activities and users.

  11. USGS Digital Spectral Library splib06a

    Science.gov (United States)

    Clark, Roger N.; Swayze, Gregg A.; Wise, Richard A.; Livo, K. Eric; Hoefen, Todd M.; Kokaly, Raymond F.; Sutley, Stephen J.

    2007-01-01

    ), one simply needs a diagnostic absorption band. The mapping system uses continuum-removed reference spectral features fitted to features in observed spectra. Spectral features for such algorithms can be obtained from a spectrum of a sample containing large amounts of contaminants, including those that add other spectral features, as long as the shape of the diagnostic feature of interest is not modified. If, however, the data are needed for radiative transfer models to derive mineral abundances from reflectance spectra, then completely uncontaminated spectra are required. This library contains spectra that span a range of quality, with purity indicators to flag spectra for (or against) particular uses. Acquiring spectral measurements and performing sample characterizations for this library has taken about 15 person-years of effort. Software to manage the library and provide scientific analysis capability is provided (Clark, 1980, 1993). A personal computer (PC) reader for the library is also available (Livo and others, 1993). The program reads specpr binary files (Clark, 1980, 1993) and plots spectra. Another program that reads the specpr format is written in IDL (Kokaly, 2005). In our view, an ideal spectral library consists of samples covering a very wide range of materials, has large wavelength range with very high precision, and has enough sample analyses and documentation to establish the quality of the spectra. Time and available resources limit what can be achieved. Ideally, for each mineral, the sample analysis would include X-ray diffraction (XRD), electron microprobe (EM) or X-ray fluorescence (XRF), and petrographic microscopic analyses. For some minerals, such as iron oxides, additional analyses such as Mossbauer would be helpful. We have found that to make the basic spectral measurements, provide XRD, EM or XRF analyses, and microscopic analyses, document the results, and complete an entry of one spectral library sample, all takes about

  12. Measuring Quality in Special Libraries: Lessons from Service Marketing.

    Science.gov (United States)

    White, Marilyn Domas; Abels, Eileen G.

    1995-01-01

    Surveys the service marketing literature for models and data-gathering instruments measuring service quality, particularly the instruments SERVQUAL and SERVPERF, and assesses their applicability to special libraries and information centers. Topics include service characteristics and definitions of service; performance-minus-expectations and…

  13. Validation of MCNP6 Version 1.0 with the ENDF/B-VII.1 Cross Section Library for Plutonium Metals, Oxides, and Solutions on the High Performance Computing Platform Moonlight

    Energy Technology Data Exchange (ETDEWEB)

    Chapman, Bryan Scott [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Gough, Sean T. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-12-05

    This report documents a validation of the MCNP6 Version 1.0 computer code on the high performance computing platform Moonlight, for operations at Los Alamos National Laboratory (LANL) that involve plutonium metals, oxides, and solutions. The validation is conducted using the ENDF/B-VII.1 continuous energy group cross section library at room temperature. The results are for use by nuclear criticality safety personnel in performing analysis and evaluation of various facility activities involving plutonium materials.

  14. Analysis of a comprehensive quality assurance program with computer-enhanced monitors

    International Nuclear Information System (INIS)

    Arenson, R.L.; Mintz, M.C.; Goldstein, E.; Stevens, J.F.; Jovais, C.

    1987-01-01

    The authors' quality assurance (QA) program provides communication pathways among its constituent committees, which include patient care, professional review, medical staff, missed case, quality control, safety, and management committees. The QA monitors are based on data from these committees but also include data from the information management system, such as patient delays, contrast reactions, incidents, complications, time-flow analyses, film library retrieval, cancellations, missing reports, and missing clinical data. Committee data include complaints, missed diagnoses, patient identification problems, and equipment failure. The QA monitors have now been incorporated into summary reports as part of their computer networks. A systematic method for follow-up ensures corrective action and documentation. Examples of improved quality of care resulting from this approach includes reductions in delays for report signature and in repeat films

  15. Usage of super high speed computer for clarification of complex phenomena

    International Nuclear Information System (INIS)

    Sekiguchi, Tomotsugu; Sato, Mitsuhisa; Nakata, Hideki; Tatebe, Osami; Takagi, Hiromitsu

    1999-01-01

    This study aims at construction of an efficient super high speed computer system application environment in response to parallel distributed system with easy transplantation to different computer system and different number by conducting research and development on super high speed computer application technology required for elucidation of complicated phenomenon in elucidation of complicated phenomenon of nuclear power field due to computed scientific method. In order to realize such environment, the Electrotechnical Laboratory has conducted development on Ninf, a network numerical information library. This Ninf system can supply a global network infrastructure for worldwide computing with high performance on further wide range distributed network (G.K.)

  16. Closing the Gap: The Maturing of Quality Assurance in Australian University Libraries

    Science.gov (United States)

    Tang, Karen

    2012-01-01

    A benchmarking review of the quality assurance practices of the libraries of the Australian Technology Network conducted in 2006 revealed exemplars of best practice, but also sector-wide gaps. A follow-up review in 2010 indicated the best practices that remain relevant. While some gaps persist, there has been improvement across the libraries and…

  17. The Quality of Academic Library Building Improvements Has a Positive Impact on Library Usage. A review of: Shill, Harold B. and Shawn Tonner. “Does the Building Still Matter? Usage Patterns in New, Expanded, and Renovated Libraries, 1995‐2002.” College & Research Libraries 65.2 (Mar.2004: 123-150.

    Directory of Open Access Journals (Sweden)

    Julie McKenna

    2006-09-01

    point scale about facility quality and librarian satisfaction with the former and the changed facility. A further criteria requirement of the availability of pre- and post-project gate count was implemented, reducing the number of libraries to be studied to 90. Facility usage changes were calculated by subtracting the gate count total for the last complete year pre-project from the most recent year gate count post project. Main results Eighty percent of the 90 libraries reported increased gate count postproject, and 20 percent reported a decline in usage. The median increase across the libraries was 37.4 percent with 25.6 percent of libraries experiencing a post‐project increase of 100 percent or more. Renovated facilities were more likely to see usage decline, but there was no statistically significant difference in usage change between renovated and new facilities. Libraries more recently upgraded saw greater usage growth than those renovations completed earlier in the study period, although 75 percent of the facilities continued to experience higher post-project usage levels. Nearly all of the private institutions (93.1% experienced usage increases and almost half experienced growth of 100 percent or more. No statistically significant relationship was found between changes in post project usage and: • The proportion of facility space allocated for library functions • The physical location of the library on campus • The size of the library facility • The level of degrees offered at the institution • The availability of wireless access • The number of computers in the instruction lab • The number of public access workstations • A larger number of seats • The number of group study rooms • The shelving capacity, the use of compact shelving or off-site storage • The presence of coffee or snack bars • The presence of any non-library facilities There was a statistically significant correlation (Pearson’s r between increased post project usage and:

  18. LISTENing to healthcare students: the impact of new library facilities on the quality of services.

    Science.gov (United States)

    Haldane, Graham C

    2003-06-01

    Following a low assessment of 'Learning resources' provision by the Quality Assurance Agency, the librarian of Homerton College, School of Health Studies commenced the LISTEN Project, a long-term study to monitor the effects of planned interventions on the quality of library provision. Surveys of entry-to-register student nurses & midwives were conducted in 1999 and 2001 by extensive questionnaires, inviting Likert-scaled and free text responses. Following a college relocation, students made greater than expected use of a new health studies library in Cambridge, and significantly less use of the local teaching hospital library. Using both a satisfaction index and a non-parametric test of mean scores, student evaluation of library services in Cambridge significantly improved following relocation. The physical accommodation and location of library services remain important to healthcare students. Identifiable improvements to the quality of services, however, will overcome initial resistance to change. Education providers must ensure the best mix of physical and electronic services for students who spend much of their time on clinical placement.

  19. Applying GRA and QFD to Improve Library Service Quality

    Science.gov (United States)

    Chen, Yen-Ting; Chou, Tsung-Yu

    2011-01-01

    This paper applied Grey Relational Analysis (GRA) to Quality Function Deployment (QFD) to identify service improvement techniques for an academic library. First, reader needs and their importance, and satisfaction degrees were examined via questionnaires. Second, the service improvement techniques for satisfying the reader needs were developed by…

  20. TerraFERMA: Harnessing Advanced Computational Libraries in Earth Science

    Science.gov (United States)

    Wilson, C. R.; Spiegelman, M.; van Keken, P.

    2012-12-01

    Many important problems in Earth sciences can be described by non-linear coupled systems of partial differential equations. These "multi-physics" problems include thermo-chemical convection in Earth and planetary interiors, interactions of fluids and magmas with the Earth's mantle and crust and coupled flow of water and ice. These problems are of interest to a large community of researchers but are complicated to model and understand. Much of this complexity stems from the nature of multi-physics where small changes in the coupling between variables or constitutive relations can lead to radical changes in behavior, which in turn affect critical computational choices such as discretizations, solvers and preconditioners. To make progress in understanding such coupled systems requires a computational framework where multi-physics problems can be described at a high-level while maintaining the flexibility to easily modify the solution algorithm. Fortunately, recent advances in computational science provide a basis for implementing such a framework. Here we present the Transparent Finite Element Rapid Model Assembler (TerraFERMA), which leverages several advanced open-source libraries for core functionality. FEniCS (fenicsproject.org) provides a high level language for describing the weak forms of coupled systems of equations, and an automatic code generator that produces finite element assembly code. PETSc (www.mcs.anl.gov/petsc) provides a wide range of scalable linear and non-linear solvers that can be composed into effective multi-physics preconditioners. SPuD (amcg.ese.ic.ac.uk/Spud) is an application neutral options system that provides both human and machine-readable interfaces based on a single xml schema. Our software integrates these libraries and provides the user with a framework for exploring multi-physics problems. A single options file fully describes the problem, including all equations, coefficients and solver options. Custom compiled applications are

  1. compomics-utilities: an open-source Java library for computational proteomics.

    Science.gov (United States)

    Barsnes, Harald; Vaudel, Marc; Colaert, Niklaas; Helsens, Kenny; Sickmann, Albert; Berven, Frode S; Martens, Lennart

    2011-03-08

    The growing interest in the field of proteomics has increased the demand for software tools and applications that process and analyze the resulting data. And even though the purpose of these tools can vary significantly, they usually share a basic set of features, including the handling of protein and peptide sequences, the visualization of (and interaction with) spectra and chromatograms, and the parsing of results from various proteomics search engines. Developers typically spend considerable time and effort implementing these support structures, which detracts from working on the novel aspects of their tool. In order to simplify the development of proteomics tools, we have implemented an open-source support library for computational proteomics, called compomics-utilities. The library contains a broad set of features required for reading, parsing, and analyzing proteomics data. compomics-utilities is already used by a long list of existing software, ensuring library stability and continued support and development. As a user-friendly, well-documented and open-source library, compomics-utilities greatly simplifies the implementation of the basic features needed in most proteomics tools. Implemented in 100% Java, compomics-utilities is fully portable across platforms and architectures. Our library thus allows the developers to focus on the novel aspects of their tools, rather than on the basic functions, which can contribute substantially to faster development, and better tools for proteomics.

  2. compomics-utilities: an open-source Java library for computational proteomics

    Directory of Open Access Journals (Sweden)

    Helsens Kenny

    2011-03-01

    Full Text Available Abstract Background The growing interest in the field of proteomics has increased the demand for software tools and applications that process and analyze the resulting data. And even though the purpose of these tools can vary significantly, they usually share a basic set of features, including the handling of protein and peptide sequences, the visualization of (and interaction with spectra and chromatograms, and the parsing of results from various proteomics search engines. Developers typically spend considerable time and effort implementing these support structures, which detracts from working on the novel aspects of their tool. Results In order to simplify the development of proteomics tools, we have implemented an open-source support library for computational proteomics, called compomics-utilities. The library contains a broad set of features required for reading, parsing, and analyzing proteomics data. compomics-utilities is already used by a long list of existing software, ensuring library stability and continued support and development. Conclusions As a user-friendly, well-documented and open-source library, compomics-utilities greatly simplifies the implementation of the basic features needed in most proteomics tools. Implemented in 100% Java, compomics-utilities is fully portable across platforms and architectures. Our library thus allows the developers to focus on the novel aspects of their tools, rather than on the basic functions, which can contribute substantially to faster development, and better tools for proteomics.

  3. Fast High-Quality Noise

    DEFF Research Database (Denmark)

    Frisvad, Jeppe Revall; Wyvill, Geoff

    2007-01-01

    At the moment the noise functions available in a graphics programmer's toolbox are either slow to compute or they involve grid-line artifacts making them of lower quality. In this paper we present a real-time noise computation with no grid-line artifacts or other regularity problems. In other words......, we put a new tool in the box that computes fast high-quality noise. In addition to being free of artifacts, the noise we present does not rely on tabulated data (everything is computed on the fly) and it is easy to adjust quality vs. quantity for the noise. The noise is based on point rendering (like...... spot noise), but it extends to more than two dimensions. The fact that it is based on point rendering makes art direction of the noise much easier....

  4. Merging Library and Computing Services at Kenyon College: A Progress Report.

    Science.gov (United States)

    Oden, Robert A., Jr.; Temple, Daniel B.; Cottrell, Janet R.; Griggs, Ronald K.; Turney, Glen W.; Wojcik, Frank M.

    2001-01-01

    Describes the evolution and progress toward a uniquely integrated library and computer services unit at Kenyon College. Discusses its focus on constituencies; merging of the divisions; benefits for students, faculty, administrative units, and the institution; meeting challenges; and generalizing from the model. (EV)

  5. COMPUTER TRAINING AND PROFESSIONAL LIBRARY

    African Journals Online (AJOL)

    professional library activities of the subject (R = 0.28507, P < 0.05). Suggestions ... hold great value for Strategic planning. ... In Pakistan for example, the Pakistan Library ..... OECD, A report on management of marketing information. OASIS ...

  6. Architecture and Initial Development of a Digital Library Platform for Computable Knowledge Objects for Health.

    Science.gov (United States)

    Flynn, Allen J; Bahulekar, Namita; Boisvert, Peter; Lagoze, Carl; Meng, George; Rampton, James; Friedman, Charles P

    2017-01-01

    Throughout the world, biomedical knowledge is routinely generated and shared through primary and secondary scientific publications. However, there is too much latency between publication of knowledge and its routine use in practice. To address this latency, what is actionable in scientific publications can be encoded to make it computable. We have created a purpose-built digital library platform to hold, manage, and share actionable, computable knowledge for health called the Knowledge Grid Library. Here we present it with its system architecture.

  7. Chat reference service in medical libraries: part 2--Trends in medical school libraries.

    Science.gov (United States)

    Dee, Cheryl R

    2003-01-01

    An increasing number of medical school libraries offer chat service to provide immediate, high quality information at the time and point of need to students, faculty, staff, and health care professionals. Part 2 of Chat Reference Service in Medical Libraries presents a snapshot of the current trends in chat reference service in medical school libraries. In late 2002, 25 (21%) medical school libraries provided chat reference. Trends in chat reference services in medical school libraries were compiled from an exploration of medical school library Web sites and informal correspondence from medical school library personnel. Many medical libraries are actively investigating and planning new chat reference services, while others have decided not to pursue chat reference at this time. Anecdotal comments from medical school library staff provide insights into chat reference service.

  8. Computer program user's manual for FIREFINDER digital topographic data verification library dubbing system

    Science.gov (United States)

    Ceres, M.; Heselton, L. R., III

    1981-11-01

    This manual describes the computer programs for the FIREFINDER Digital Topographic Data Verification-Library-Dubbing System (FFDTDVLDS), and will assist in the maintenance of these programs. The manual contains detailed flow diagrams and associated descriptions for each computer program routine and subroutine. Complete computer program listings are also included. This information should be used when changes are made in the computer programs. The operating system has been designed to minimize operator intervention.

  9. Application Reuse Library for Software, Requirements, and Guidelines

    Science.gov (United States)

    Malin, Jane T.; Thronesbery, Carroll

    1994-01-01

    Better designs are needed for expert systems and other operations automation software, for more reliable, usable and effective human support. A prototype computer-aided Application Reuse Library shows feasibility of supporting concurrent development and improvement of advanced software by users, analysts, software developers, and human-computer interaction experts. Such a library expedites development of quality software, by providing working, documented examples, which support understanding, modification and reuse of requirements as well as code. It explicitly documents and implicitly embodies design guidelines, standards and conventions. The Application Reuse Library provides application modules with Demo-and-Tester elements. Developers and users can evaluate applicability of a library module and test modifications, by running it interactively. Sub-modules provide application code and displays and controls. The library supports software modification and reuse, by providing alternative versions of application and display functionality. Information about human support and display requirements is provided, so that modifications will conform to guidelines. The library supports entry of new application modules from developers throughout an organization. Example library modules include a timer, some buttons and special fonts, and a real-time data interface program. The library prototype is implemented in the object-oriented G2 environment for developing real-time expert systems.

  10. The ELPA library: scalable parallel eigenvalue solutions for electronic structure theory and computational science.

    Science.gov (United States)

    Marek, A; Blum, V; Johanni, R; Havu, V; Lang, B; Auckenthaler, T; Heinecke, A; Bungartz, H-J; Lederer, H

    2014-05-28

    Obtaining the eigenvalues and eigenvectors of large matrices is a key problem in electronic structure theory and many other areas of computational science. The computational effort formally scales as O(N(3)) with the size of the investigated problem, N (e.g. the electron count in electronic structure theory), and thus often defines the system size limit that practical calculations cannot overcome. In many cases, more than just a small fraction of the possible eigenvalue/eigenvector pairs is needed, so that iterative solution strategies that focus only on a few eigenvalues become ineffective. Likewise, it is not always desirable or practical to circumvent the eigenvalue solution entirely. We here review some current developments regarding dense eigenvalue solvers and then focus on the Eigenvalue soLvers for Petascale Applications (ELPA) library, which facilitates the efficient algebraic solution of symmetric and Hermitian eigenvalue problems for dense matrices that have real-valued and complex-valued matrix entries, respectively, on parallel computer platforms. ELPA addresses standard as well as generalized eigenvalue problems, relying on the well documented matrix layout of the Scalable Linear Algebra PACKage (ScaLAPACK) library but replacing all actual parallel solution steps with subroutines of its own. For these steps, ELPA significantly outperforms the corresponding ScaLAPACK routines and proprietary libraries that implement the ScaLAPACK interface (e.g. Intel's MKL). The most time-critical step is the reduction of the matrix to tridiagonal form and the corresponding backtransformation of the eigenvectors. ELPA offers both a one-step tridiagonalization (successive Householder transformations) and a two-step transformation that is more efficient especially towards larger matrices and larger numbers of CPU cores. ELPA is based on the MPI standard, with an early hybrid MPI-OpenMPI implementation available as well. Scalability beyond 10,000 CPU cores for problem

  11. The Ten Commandments revisited: the Qualities of Good Library Space

    Directory of Open Access Journals (Sweden)

    Andrew McDonald

    2006-06-01

    Full Text Available An increasing diversity of imaginative new academic libraries are being constructed around the world, successfully combining exciting architectural expression, inspiring internal spaces and good functionality. Library managers must have a strong vision for the new library and this should inspire the design and the whole building process. This paper explores the key qualities of good learning space, whether in new or refurbished buildings. It is suggested that, ideally, learning space should be functional, adaptable, accessible, varied, interactive, conducive, environmentally suitable, safe and secure, efficient and suitable for information technology. New space should also have “oomph”, capturing the minds of users and the spirit of the university. These indicative issues should be discussed in the brief and throughout the planning process, and the priority given to them will depend on the mission and culture of the library. Greater attention is being given to daylight, natural ventilation, cultural artwork, noise management, security, disabled access, information skills training and provision for e-services. Planning is increasingly centred on people or the learner, emphasising the need for social, interactive and collaborative learning spaces as well as for traditional spaces for quiet study and reflection. Some libraries are joined-up with other services. The academic library is an enduring physical ‘place’, providing a blended, hybrid environment of traditional and electronic services crucial for the future of our universities and their communities.

  12. Stampi: a message passing library for distributed parallel computing. User's guide

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Koide, Hiroshi; Takemiya, Hiroshi

    1998-11-01

    A new message passing library, Stampi, has been developed to realize a computation with different kind of parallel computers arbitrarily and making MPI (Message Passing Interface) as an unique interface for communication. Stampi is based on MPI2 specification. It realizes dynamic process creation to different machines and communication between spawned one within the scope of MPI semantics. Vender implemented MPI as a closed system in one parallel machine and did not support both functions; process creation and communication to external machines. Stampi supports both functions and enables us distributed parallel computing. Currently Stampi has been implemented on COMPACS (COMplex PArallel Computer System) introduced in CCSE, five parallel computers and one graphic workstation, and any communication on them can be processed on. (author)

  13. Organization of a library of standard relocatible programmes, or a processing measurement data module based on computers of the TRA types

    International Nuclear Information System (INIS)

    Dadi, K.; Dadi, L.; Mateeva, A.; Salamatin, I.M.

    1976-01-01

    The paper describes the organization of a library of standard programs with binary cade. The library was developed for a measurement module on the basis of a TRA-1001-i computer (Elektronika-100, PDP-8). The library is placed on a external memory (magnetic disk) and has a module structure. The external memory assigned for the library is divided into pages. When loaded into the computer internal memory, several pages are taken as one whole to represent the loading module. The magnetic disk storage capacity being 1.25 million words, the library has a total of ca. 50 10 thousand words (eight cylinders). The work provides regulations for compiling standard programs in SLANG. The library is characterized by the following main features: possibility of being used in memory dynamic distribution mode; possibility of being used for computers with internal memory capacity 4K; no need for intermediary-language coding of displaced program; and possibility of autonomous shift of standard program. The above library is compared with a comprising DES programs library

  14. Dealing with metadata quality: the legacy of digital library efforts

    OpenAIRE

    Tani, Alice; Candela, Leonardo; Castelli, Donatella

    2013-01-01

    In this work, we elaborate on the meaning of metadata quality by surveying efforts and experiences matured in the digital library domain. In particular, an overview of the frameworks developed to characterize such a multi-faceted concept is presented. Moreover, the most common quality-related problems affecting metadata both during the creation and the aggregation phase are discussed together with the approaches, technologies and tools developed to mitigate them. This survey on digital librar...

  15. Libraries for users services in academic libraries

    CERN Document Server

    Alvite, Luisa

    2010-01-01

    This book reviews the quality and evolution of academic library services. It revises service trends offered by academic libraries and the challenge of enhancing traditional ones such as: catalogues, repositories and digital collections, learning resources centres, virtual reference services, information literacy and 2.0 tools.studies the role of the university library in the new educational environment of higher educationrethinks libraries in academic contextredefines roles for academic libraries

  16. Cloud Computing and Your Library

    Science.gov (United States)

    Mitchell, Erik T.

    2010-01-01

    One of the first big shifts in how libraries manage resources was the move from print-journal purchasing models to database-subscription and electronic-journal purchasing models. Libraries found that this transition helped them scale their resources and provide better service just by thinking a bit differently about their services. Likewise,…

  17. The Presence of Web 2.0 Applications Is Associated with the Overall Service Quality of Library Websites. A Review of: Chua, A. Y. K., & Goh, D. H. (2010. A study of Web 2.0 applications in library websites. Library & Information Science Research, 32(3, 203-211.

    Directory of Open Access Journals (Sweden)

    Leslie Bussert

    2011-03-01

    Full Text Available Objective – To determine the prevalence and use of web 2.0 applications in library websites and to determine whether or not their presence enhances the quality of the sites.Design – Content analysis.Setting – Public and academic libraries in North America, Europe, and Asia.Subjects – A total of 120 academic and public library websites in English.Methods – This study identified and selected library websites from academic and public libraries using a combination of directories, ratings reports, and ranking lists from three geographic regions. Over a four-month period in 2009, three coders conducted a three-step content analysis of the selected library websites. This analysis was conducted after having established inter-coder reliability using Cohen’s Kappa and analytic procedure familiarity, using a randomly selected pilot set of 30 library websites. The remaining 90 websites were evenly distributed among the three coders for analysis. They determined whether web 2.0 applications were present, examined how those applications were used, and gave an overall appraisal of website quality. Coders inspected library website links, conducted site searches, used search engines, and searched within major social networking sites to determine the presence of web 2.0 applications. A quality framework classification scheme was used during coding to identify how the web 2.0 applications were being used within library websites. This framework was established around four dimensions of library services: information acquisition (blogs and wikis, information dissemination (Rich Site Summary, or “RSS”, information organization (social tagging, and information sharing (social networking and instant messaging. A five-point Likert scale was also used in concert with a website quality evaluation framework to assess the quality of the library websites. This framework included three aspects of website quality: system quality, information quality, and service

  18. A high-speed linear algebra library with automatic parallelism

    Science.gov (United States)

    Boucher, Michael L.

    1994-01-01

    Parallel or distributed processing is key to getting highest performance workstations. However, designing and implementing efficient parallel algorithms is difficult and error-prone. It is even more difficult to write code that is both portable to and efficient on many different computers. Finally, it is harder still to satisfy the above requirements and include the reliability and ease of use required of commercial software intended for use in a production environment. As a result, the application of parallel processing technology to commercial software has been extremely small even though there are numerous computationally demanding programs that would significantly benefit from application of parallel processing. This paper describes DSSLIB, which is a library of subroutines that perform many of the time-consuming computations in engineering and scientific software. DSSLIB combines the high efficiency and speed of parallel computation with a serial programming model that eliminates many undesirable side-effects of typical parallel code. The result is a simple way to incorporate the power of parallel processing into commercial software without compromising maintainability, reliability, or ease of use. This gives significant advantages over less powerful non-parallel entries in the market.

  19. The Library Macintosh at SCIL [Small Computers in Libraries]'88.

    Science.gov (United States)

    Valauskas, Edward J.; And Others

    1988-01-01

    The first of three papers describes the role of Macintosh workstations in a library. The second paper explains why the Macintosh was selected for end-user searching in an academic library, and the third discusses advantages and disadvantages of desktop publishing for librarians. (8 references) (MES)

  20. Quality evaluation of tandem mass spectral libraries.

    Science.gov (United States)

    Oberacher, Herbert; Weinmann, Wolfgang; Dresen, Sebastian

    2011-06-01

    Tandem mass spectral libraries are gaining more and more importance for the identification of unknowns in different fields of research, including metabolomics, forensics, toxicology, and environmental analysis. Particularly, the recent invention of reliable, robust, and transferable libraries has increased the general acceptance of these tools. Herein, we report on results obtained from thorough evaluation of the match reliabilities of two tandem mass spectral libraries: the MSforID library established by the Oberacher group in Innsbruck and the Weinmann library established by the Weinmann group in Freiburg. Three different experiments were performed: (1) Spectra of the libraries were searched against their corresponding library after excluding either this single compound-specific spectrum or all compound-specific spectra prior to searching; (2) the libraries were searched against each other using either library as reference set or sample set; (3) spectra acquired on different mass spectrometric instruments were matched to both libraries. Almost 13,000 tandem mass spectra were included in this study. The MSforID search algorithm was used for spectral matching. Statistical evaluation of the library search results revealed that principally both libraries enable the sensitive and specific identification of compounds. Due to higher mass accuracy of the QqTOF compared with the QTrap instrument, matches to the MSforID library were more reliable when comparing spectra with both libraries. Furthermore, only the MSforID library was shown to be efficiently transferable to different kinds of tandem mass spectrometers, including "tandem-in-time" instruments; this is due to the coverage of a large range of different collision energy settings-including the very low range-which is an outstanding characteristics of the MSforID library.

  1. Distributed multiscale computing with MUSCLE 2, the Multiscale Coupling Library and Environment

    NARCIS (Netherlands)

    Borgdorff, J.; Mamonski, M.; Bosak, B.; Kurowski, K.; Ben Belgacem, M.; Chopard, B.; Groen, D.; Coveney, P.V.; Hoekstra, A.G.

    2014-01-01

    We present the Multiscale Coupling Library and Environment: MUSCLE 2. This multiscale component-based execution environment has a simple to use Java, C++, C, Python and Fortran API, compatible with MPI, OpenMP and threading codes. We demonstrate its local and distributed computing capabilities and

  2. Stampi: a message passing library for distributed parallel computing. User's guide, second edition

    International Nuclear Information System (INIS)

    Imamura, Toshiyuki; Koide, Hiroshi; Takemiya, Hiroshi

    2000-02-01

    A new message passing library, Stampi, has been developed to realize a computation with different kind of parallel computers arbitrarily and making MPI (Message Passing Interface) as an unique interface for communication. Stampi is based on the MPI2 specification, and it realizes dynamic process creation to different machines and communication between spawned one within the scope of MPI semantics. Main features of Stampi are summarized as follows: (i) an automatic switch function between external- and internal communications, (ii) a message routing/relaying with a routing module, (iii) a dynamic process creation, (iv) a support of two types of connection, Master/Slave and Client/Server, (v) a support of a communication with Java applets. Indeed vendors implemented MPI libraries as a closed system in one parallel machine or their systems, and did not support both functions; process creation and communication to external machines. Stampi supports both functions and enables us distributed parallel computing. Currently Stampi has been implemented on COMPACS (COMplex PArallel Computer System) introduced in CCSE, five parallel computers and one graphic workstation, moreover on eight kinds of parallel machines, totally fourteen systems. Stampi provides us MPI communication functionality on them. This report describes mainly the usage of Stampi. (author)

  3. Automated high speed volume computed tomography for inline quality control

    International Nuclear Information System (INIS)

    Hanke, R.; Kugel, A.; Troup, P.

    2004-01-01

    Increasing complexity of innovative products as well as growing requirements on quality and reliability call for more detailed knowledge about internal structures of manufactured components rather by 100 % inspection than just by sampling test. A first-step solution, like radioscopic inline inspection machines, equipped with automated data evaluation software, have become state of the art in the production floor during the last years. However, these machines provide just ordinary two-dimensional information and deliver no volume data e.g. to evaluate exact position or shape of detected defects. One way to solve this problem is the application of X-ray computed tomography (CT). Compared to the performance of the first generation medical scanners (scanning times of many hours), today, modern Volume CT machines for industrial applications need about 5 minutes for a full object scan depending on the object size. Of course, this is still too long to introduce this powerful method into the inline production quality control. In order to gain acceptance, the scanning time including subsequent data evaluation must be decreased significantly and adapted to the manufacturing cycle times. This presentation demonstrates the new technical set up, reconstruction results and the methods for high-speed volume data evaluation of a new fully automated high-speed CT scanner with cycle times below one minute for an object size of less than 15 cm. This will directly create new opportunities in design and construction of more complex objects. (author)

  4. Statistical molecular design of balanced compound libraries for QSAR modeling.

    Science.gov (United States)

    Linusson, A; Elofsson, M; Andersson, I E; Dahlgren, M K

    2010-01-01

    A fundamental step in preclinical drug development is the computation of quantitative structure-activity relationship (QSAR) models, i.e. models that link chemical features of compounds with activities towards a target macromolecule associated with the initiation or progression of a disease. QSAR models are computed by combining information on the physicochemical and structural features of a library of congeneric compounds, typically assembled from two or more building blocks, and biological data from one or more in vitro assays. Since the models provide information on features affecting the compounds' biological activity they can be used as guides for further optimization. However, in order for a QSAR model to be relevant to the targeted disease, and drug development in general, the compound library used must contain molecules with balanced variation of the features spanning the chemical space believed to be important for interaction with the biological target. In addition, the assays used must be robust and deliver high quality data that are directly related to the function of the biological target and the associated disease state. In this review, we discuss and exemplify the concept of statistical molecular design (SMD) in the selection of building blocks and final synthetic targets (i.e. compounds to synthesize) to generate information-rich, balanced libraries for biological testing and computation of QSAR models.

  5. Close the Gate, Lock the Windows, Bolt the Doors: Securing Library Computers. Online Treasures

    Science.gov (United States)

    Balas, Janet

    2005-01-01

    This article, written by a systems librarian at the Monroeville Public Library, discusses a major issue affecting all computer users, security. It indicates that while, staying up-to-date on the latest security issues has become essential for all computer users, it's more critical for network managers who are responsible for securing computer…

  6. Ultra-High-Resolution Computed Tomography of the Lung: Image Quality of a Prototype Scanner

    OpenAIRE

    Kakinuma, Ryutaro; Moriyama, Noriyuki; Muramatsu, Yukio; Gomi, Shiho; Suzuki, Masahiro; Nagasawa, Hirobumi; Kusumoto, Masahiko; Aso, Tomohiko; Muramatsu, Yoshihisa; Tsuchida, Takaaki; Tsuta, Koji; Maeshima, Akiko Miyagi; Tochigi, Naobumi; Watanabe, Shun-ichi; Sugihara, Naoki

    2015-01-01

    Purpose: The image noise and image quality of a prototype ultra-high-resolution computed tomography (U-HRCT) scanner was evaluated and compared with those of conventional high-resolution CT (C-HRCT) scanners. Materials and Methods: This study was approved by the institutional review board. A U-HRCT scanner prototype with 0.25 mm × 4 rows and operating at 120 mAs was used. The C-HRCT images were obtained using a 0.5 mm × 16 or 0.5 mm × 64 detector-row CT scanner operating at 150 mAs. Images fr...

  7. Studying on the problem of quality, ability and adult education of the management personnel of nuclear power library and information

    International Nuclear Information System (INIS)

    Ma Hua

    2012-01-01

    Studying on how to improve the quality and ability of the nuclear power library and information management personnel has an important role in the promotion of nuclear power industry and the development of nuclear power library and information. By the view of the perspective of the development of nuclear power industry and the improvement of quality and level of nuclear power library and information. The library and information management personnel of nuclear power in the new period should have seven types of basic quality and ability. Vigorously developing adult education is an effective way to comprehensively improve the quality and ability of the nuclear power library and information management personnel. Doing well in adult education must grasp some key issues: put ideological and moral education first, regard pertinence and effectiveness and innovate ways and methods of education. (author)

  8. Computational Biology and High Performance Computing 2000

    Energy Technology Data Exchange (ETDEWEB)

    Simon, Horst D.; Zorn, Manfred D.; Spengler, Sylvia J.; Shoichet, Brian K.; Stewart, Craig; Dubchak, Inna L.; Arkin, Adam P.

    2000-10-19

    The pace of extraordinary advances in molecular biology has accelerated in the past decade due in large part to discoveries coming from genome projects on human and model organisms. The advances in the genome project so far, happening well ahead of schedule and under budget, have exceeded any dreams by its protagonists, let alone formal expectations. Biologists expect the next phase of the genome project to be even more startling in terms of dramatic breakthroughs in our understanding of human biology, the biology of health and of disease. Only today can biologists begin to envision the necessary experimental, computational and theoretical steps necessary to exploit genome sequence information for its medical impact, its contribution to biotechnology and economic competitiveness, and its ultimate contribution to environmental quality. High performance computing has become one of the critical enabling technologies, which will help to translate this vision of future advances in biology into reality. Biologists are increasingly becoming aware of the potential of high performance computing. The goal of this tutorial is to introduce the exciting new developments in computational biology and genomics to the high performance computing community.

  9. CARP: a computer code and albedo data library for use by BREESE, the MORSE albedo package

    International Nuclear Information System (INIS)

    Emmett, M.B.; Rhoades, W.A.

    1978-10-01

    The CARP computer code was written to allow processing of DOT angular flux tapes to produce albedo data for use in the MORSE computer code. An albedo data library was produced containing several materials. 3 tables

  10. Total Quality Management in Libraries: A Short Review

    Directory of Open Access Journals (Sweden)

    Erol Yılmaz

    2013-11-01

    Full Text Available Total Quality Management (TQM, is a management style that aimed at customer satisfaction. In the organisations in which TQM is applied customer is in the centre of the organization. In other words, in this kind of organisations the customer comes before everything else. Reaching the highest level of the customer satisfaction is among the targets of the mentioned organizations. In this study, primarily, definition of the TQM is going to be made together with its definition, short history, purpose, benefits and factors; besides descriptive information is going to be given about internal customer, external customer, customer centered process and customer satisfaction. In the second part of the study, TQM subject is going to be discussed within the scope of the libraries widely. In this context, after explanation of customer centeredness (placing the customer at the centre, customer satisfaction and benefits of the TQM to the libraries, the subject is going to be discussed in the context of our country.

  11. Designing a diverse high-quality library for crystallography-based FBDD screening.

    Science.gov (United States)

    Tounge, Brett A; Parker, Michael H

    2011-01-01

    A well-chosen set of fragments is able to cover a large chemical space using a small number of compounds. The actual size and makeup of the fragment set is dependent on the screening method since each technique has its own practical limits in terms of the number of compounds that can be screened and requirements for compound solubility. In this chapter, an overview of the general requirements for a fragment library is presented for different screening platforms. In the case of the FBDD work at Johnson & Johnson Pharmaceutical Research and Development, L.L.C., our main screening technology is X-ray crystallography. Since every soaked protein crystal needs to be diffracted and a protein structure determined to delineate if a fragment binds, the size of our initial screening library cannot be a rate-limiting factor. For this reason, we have chosen 900 as the appropriate primary fragment library size. To choose the best set, we have developed our own mix of simple property ("Rule of 3") and "bad" substructure filtering. While this gets one a long way in terms of limiting the fragment pool, there are still tens of thousands of compounds to choose from after this initial step. Many of the choices left at this stage are not drug-like, so we have developed an FBDD Score to help select a 900-compound set. The details of this score and the filtering are presented. Copyright © 2011 Elsevier Inc. All rights reserved.

  12. Evaluation of the quality of the college library websites in Iranian medical Universities based on the Stover model.

    Science.gov (United States)

    Nasajpour, Mohammad Reza; Ashrafi-Rizi, Hasan; Soleymani, Mohammad Reza; Shahrzadi, Leila; Hassanzadeh, Akbar

    2014-01-01

    Today, the websites of college and university libraries play an important role in providing the necessary services for clients. These websites not only allow the users to access different collections of library resources, but also provide them with the necessary guidance in order to use the information. The goal of this study is the quality evaluation of the college library websites in Iranian Medical Universities based on the Stover model. This study uses an analytical survey method and is an applied study. The data gathering tool is the standard checklist provided by Stover, which was modified by the researchers for this study. The statistical population is the college library websites of the Iranian Medical Universities (146 websites) and census method was used for investigation. The data gathering method was a direct access to each website and filling of the checklist was based on the researchers' observations. Descriptive and analytical statistics (Analysis of Variance (ANOVA)) were used for data analysis with the help of the SPSS software. The findings showed that in the dimension of the quality of contents, the highest average belonged to type one universities (46.2%) and the lowest average belonged to type three universities (24.8%). In the search and research capabilities, the highest average belonged to type one universities (48.2%) and the lowest average belonged to type three universities. In the dimension of facilities provided for the users, type one universities again had the highest average (37.2%), while type three universities had the lowest average (15%). In general the library websites of type one universities had the highest quality (44.2%), while type three universities had the lowest quality (21.1%). Also the library websites of the College of Rehabilitation and the College of Paramedics, of the Shiraz University of Medical Science, had the highest quality scores. The results showed that there was a meaningful difference between the quality

  13. Toward a TQM Paradigm: Using SERVQUAL to Measure Library Service Quality.

    Science.gov (United States)

    Coleman, Vicki; And Others

    1997-01-01

    Texas A&M University's Evans Library conducted a survey (SERVQUAL) to measure service quality in terms of tangibles (appearance of facilities, equipment, personnel, and communication materials), reliability, responsiveness, assurance, and empathy. Discrepancies between expectations and perceptions were found in reliability, responsiveness,…

  14. Information Technology Planning: Computers in the School Library--How Many Are Enough?

    Science.gov (United States)

    Simpson, Carol

    2002-01-01

    Describes the development of a formula to determine the needed quantity of computers for a school library. Four types of information technology activities (administrative, personal productive, class/group productive, online public access catalog) and several variables (age levels served, campus focus, number of staff, size of student body, average…

  15. GASPRNG: GPU accelerated scalable parallel random number generator library

    Science.gov (United States)

    Gao, Shuang; Peterson, Gregory D.

    2013-04-01

    Graphics processors represent a promising technology for accelerating computational science applications. Many computational science applications require fast and scalable random number generation with good statistical properties, so they use the Scalable Parallel Random Number Generators library (SPRNG). We present the GPU Accelerated SPRNG library (GASPRNG) to accelerate SPRNG in GPU-based high performance computing systems. GASPRNG includes code for a host CPU and CUDA code for execution on NVIDIA graphics processing units (GPUs) along with a programming interface to support various usage models for pseudorandom numbers and computational science applications executing on the CPU, GPU, or both. This paper describes the implementation approach used to produce high performance and also describes how to use the programming interface. The programming interface allows a user to be able to use GASPRNG the same way as SPRNG on traditional serial or parallel computers as well as to develop tightly coupled programs executing primarily on the GPU. We also describe how to install GASPRNG and use it. To help illustrate linking with GASPRNG, various demonstration codes are included for the different usage models. GASPRNG on a single GPU shows up to 280x speedup over SPRNG on a single CPU core and is able to scale for larger systems in the same manner as SPRNG. Because GASPRNG generates identical streams of pseudorandom numbers as SPRNG, users can be confident about the quality of GASPRNG for scalable computational science applications. Catalogue identifier: AEOI_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEOI_v1_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, N. Ireland Licensing provisions: UTK license. No. of lines in distributed program, including test data, etc.: 167900 No. of bytes in distributed program, including test data, etc.: 1422058 Distribution format: tar.gz Programming language: C and CUDA. Computer: Any PC or

  16. ALOHA: Automatic libraries of helicity amplitudes for Feynman diagram computations

    Science.gov (United States)

    de Aquino, Priscila; Link, William; Maltoni, Fabio; Mattelaer, Olivier; Stelzer, Tim

    2012-10-01

    We present an application that automatically writes the HELAS (HELicity Amplitude Subroutines) library corresponding to the Feynman rules of any quantum field theory Lagrangian. The code is written in Python and takes the Universal FeynRules Output (UFO) as an input. From this input it produces the complete set of routines, wave-functions and amplitudes, that are needed for the computation of Feynman diagrams at leading as well as at higher orders. The representation is language independent and currently it can output routines in Fortran, C++, and Python. A few sample applications implemented in the MADGRAPH 5 framework are presented. Program summary Program title: ALOHA Catalogue identifier: AEMS_v1_0 Program summary URL: http://cpc.cs.qub.ac.uk/summaries/AEMS_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: http://www.opensource.org/licenses/UoI-NCSA.php No. of lines in distributed program, including test data, etc.: 6094320 No. of bytes in distributed program, including test data, etc.: 7479819 Distribution format: tar.gz Programming language: Python2.6 Computer: 32/64 bit Operating system: Linux/Mac/Windows RAM: 512 Mbytes Classification: 4.4, 11.6 Nature of problem: An effcient numerical evaluation of a squared matrix element can be done with the help of the helicity routines implemented in the HELAS library [1]. This static library contains a limited number of helicity functions and is therefore not always able to provide the needed routine in the presence of an arbitrary interaction. This program provides a way to automatically create the corresponding routines for any given model. Solution method: ALOHA takes the Feynman rules associated to the vertex obtained from the model information (in the UFO format [2]), and multiplies it by the different wavefunctions or propagators. As a result the analytical expression of the helicity routines is obtained. Subsequently, this expression is

  17. Distribution and Availability of State and Areawide Water Quality Reports in Oklahoma Libraries.

    Science.gov (United States)

    McClure, Charles R.; Million, Anne

    This report examines the distribution and availability of water quality reports in the state of Oklahoma. Based on legislation from the Clean Water Act and regulations from the Environmental Protection Agency's "Public Participation Handbook for Water Quality Management," depository libraries must be established to provide citizen access to…

  18. High-performance computational fluid dynamics: a custom-code approach

    International Nuclear Information System (INIS)

    Fannon, James; Náraigh, Lennon Ó; Loiseau, Jean-Christophe; Valluri, Prashant; Bethune, Iain

    2016-01-01

    We introduce a modified and simplified version of the pre-existing fully parallelized three-dimensional Navier–Stokes flow solver known as TPLS. We demonstrate how the simplified version can be used as a pedagogical tool for the study of computational fluid dynamics (CFDs) and parallel computing. TPLS is at its heart a two-phase flow solver, and uses calls to a range of external libraries to accelerate its performance. However, in the present context we narrow the focus of the study to basic hydrodynamics and parallel computing techniques, and the code is therefore simplified and modified to simulate pressure-driven single-phase flow in a channel, using only relatively simple Fortran 90 code with MPI parallelization, but no calls to any other external libraries. The modified code is analysed in order to both validate its accuracy and investigate its scalability up to 1000 CPU cores. Simulations are performed for several benchmark cases in pressure-driven channel flow, including a turbulent simulation, wherein the turbulence is incorporated via the large-eddy simulation technique. The work may be of use to advanced undergraduate and graduate students as an introductory study in CFDs, while also providing insight for those interested in more general aspects of high-performance computing. (paper)

  19. High-performance computational fluid dynamics: a custom-code approach

    Science.gov (United States)

    Fannon, James; Loiseau, Jean-Christophe; Valluri, Prashant; Bethune, Iain; Náraigh, Lennon Ó.

    2016-07-01

    We introduce a modified and simplified version of the pre-existing fully parallelized three-dimensional Navier-Stokes flow solver known as TPLS. We demonstrate how the simplified version can be used as a pedagogical tool for the study of computational fluid dynamics (CFDs) and parallel computing. TPLS is at its heart a two-phase flow solver, and uses calls to a range of external libraries to accelerate its performance. However, in the present context we narrow the focus of the study to basic hydrodynamics and parallel computing techniques, and the code is therefore simplified and modified to simulate pressure-driven single-phase flow in a channel, using only relatively simple Fortran 90 code with MPI parallelization, but no calls to any other external libraries. The modified code is analysed in order to both validate its accuracy and investigate its scalability up to 1000 CPU cores. Simulations are performed for several benchmark cases in pressure-driven channel flow, including a turbulent simulation, wherein the turbulence is incorporated via the large-eddy simulation technique. The work may be of use to advanced undergraduate and graduate students as an introductory study in CFDs, while also providing insight for those interested in more general aspects of high-performance computing.

  20. Parallel Computation on Multicore Processors Using Explicit Form of the Finite Element Method and C++ Standard Libraries

    Directory of Open Access Journals (Sweden)

    Rek Václav

    2016-11-01

    Full Text Available In this paper, the form of modifications of the existing sequential code written in C or C++ programming language for the calculation of various kind of structures using the explicit form of the Finite Element Method (Dynamic Relaxation Method, Explicit Dynamics in the NEXX system is introduced. The NEXX system is the core of engineering software NEXIS, Scia Engineer, RFEM and RENEX. It has the possibilities of multithreaded running, which can now be supported at the level of native C++ programming language using standard libraries. Thanks to the high degree of abstraction that a contemporary C++ programming language provides, a respective library created in this way can be very generalized for other purposes of usage of parallelism in computational mechanics.

  1. The impact of computerisation of library operations on library ...

    African Journals Online (AJOL)

    The use of computer-based systems in libraries and information units is now a vogue. The era of manual system in library operations is on its way to extinction. Recent developments in information world tend towards a globalized information communication technology (ICT). The library as a dynamic institution cannot afford ...

  2. Computer Assisted Circulation Control at Health Sciences Library SUNYAB

    Directory of Open Access Journals (Sweden)

    Jean K. Miller

    1972-06-01

    Full Text Available A description of the circulation system which the Health Sciences Library at the State University of New York at Buffalo has been using since October 1970. Features of the system include automatic production of overdue, fine, and billing notices; notices for call-in of requested books; and book availability notices. Remote operation and processing on the IBM 360/40 and CDC 6400 computer are accomplished via the Administrative Terminal System (ATS and Terminal job Entry (T]E. The system provides information for management of the collection and improved service to the user.

  3. High Affinity, Developability and Functional Size: The Holy Grail of Combinatorial Antibody Library Generation

    Directory of Open Access Journals (Sweden)

    Kathrin Tissot

    2011-05-01

    Full Text Available Since the initial description of phage display technology for the generation of human antibodies, a variety of selection methods has been developed. The most critical parameter for all in vitro-based approaches is the quality of the antibody library. Concurrent evolution of the libraries has allowed display and selection technologies to reveal their full potential. They come in different flavors, from naïve to fully synthetic and differ in terms of size, quality, method of preparation, framework and CDR composition. Early on, the focus has mainly been on affinities and thus on library size and diversity. Subsequently, the increased awareness of developability and cost of goods as important success factors has spurred efforts to generate libraries with improved biophysical properties and favorable production characteristics. More recently a major focus on reduction of unwanted side effects through reduced immunogenicity and improved overall biophysical behavior has led to a re-evaluation of library design.

  4. SU-G-BRB-02: An Open-Source Software Analysis Library for Linear Accelerator Quality Assurance

    International Nuclear Information System (INIS)

    Kerns, J; Yaldo, D

    2016-01-01

    Purpose: Routine linac quality assurance (QA) tests have become complex enough to require automation of most test analyses. A new data analysis software library was built that allows physicists to automate routine linear accelerator quality assurance tests. The package is open source, code tested, and benchmarked. Methods: Images and data were generated on a TrueBeam linac for the following routine QA tests: VMAT, starshot, CBCT, machine logs, Winston Lutz, and picket fence. The analysis library was built using the general programming language Python. Each test was analyzed with the library algorithms and compared to manual measurements taken at the time of acquisition. Results: VMAT QA results agreed within 0.1% between the library and manual measurements. Machine logs (dynalogs & trajectory logs) were successfully parsed; mechanical axis positions were verified for accuracy and MLC fluence agreed well with EPID measurements. CBCT QA measurements were within 10 HU and 0.2mm where applicable. Winston Lutz isocenter size measurements were within 0.2mm of TrueBeam’s Machine Performance Check. Starshot analysis was within 0.2mm of the Winston Lutz results for the same conditions. Picket fence images with and without a known error showed that the library was capable of detecting MLC offsets within 0.02mm. Conclusion: A new routine QA software library has been benchmarked and is available for use by the community. The library is open-source and extensible for use in larger systems.

  5. SU-G-BRB-02: An Open-Source Software Analysis Library for Linear Accelerator Quality Assurance

    Energy Technology Data Exchange (ETDEWEB)

    Kerns, J [UT MD Anderson Cancer Center, Houston, TX (United States); Yaldo, D [Advocate Health Care, Park Ridge, IL (United States)

    2016-06-15

    Purpose: Routine linac quality assurance (QA) tests have become complex enough to require automation of most test analyses. A new data analysis software library was built that allows physicists to automate routine linear accelerator quality assurance tests. The package is open source, code tested, and benchmarked. Methods: Images and data were generated on a TrueBeam linac for the following routine QA tests: VMAT, starshot, CBCT, machine logs, Winston Lutz, and picket fence. The analysis library was built using the general programming language Python. Each test was analyzed with the library algorithms and compared to manual measurements taken at the time of acquisition. Results: VMAT QA results agreed within 0.1% between the library and manual measurements. Machine logs (dynalogs & trajectory logs) were successfully parsed; mechanical axis positions were verified for accuracy and MLC fluence agreed well with EPID measurements. CBCT QA measurements were within 10 HU and 0.2mm where applicable. Winston Lutz isocenter size measurements were within 0.2mm of TrueBeam’s Machine Performance Check. Starshot analysis was within 0.2mm of the Winston Lutz results for the same conditions. Picket fence images with and without a known error showed that the library was capable of detecting MLC offsets within 0.02mm. Conclusion: A new routine QA software library has been benchmarked and is available for use by the community. The library is open-source and extensible for use in larger systems.

  6. Shoestring Digital Library: If Existing Digital Library Software Doesn't Suit Your Needs, Create Your Own

    Science.gov (United States)

    Weber, Jonathan

    2006-01-01

    Creating a digital library might seem like a task best left to a large research collection with a vast staff and generous budget. However, tools for successfully creating digital libraries are getting easier to use all the time. The explosion of people creating content for the web has led to the availability of many high-quality applications and…

  7. phpMs: A PHP-Based Mass Spectrometry Utilities Library.

    Science.gov (United States)

    Collins, Andrew; Jones, Andrew R

    2018-03-02

    The recent establishment of cloud computing, high-throughput networking, and more versatile web standards and browsers has led to a renewed interest in web-based applications. While traditionally big data has been the domain of optimized desktop and server applications, it is now possible to store vast amounts of data and perform the necessary calculations offsite in cloud storage and computing providers, with the results visualized in a high-quality cross-platform interface via a web browser. There are number of emerging platforms for cloud-based mass spectrometry data analysis; however, there is limited pre-existing code accessible to web developers, especially for those that are constrained to a shared hosting environment where Java and C applications are often forbidden from use by the hosting provider. To remedy this, we provide an open-source mass spectrometry library for one of the most commonly used web development languages, PHP. Our new library, phpMs, provides objects for storing and manipulating spectra and identification data as well as utilities for file reading, file writing, calculations, peptide fragmentation, and protein digestion as well as a software interface for controlling search engines. We provide a working demonstration of some of the capabilities at http://pgb.liv.ac.uk/phpMs .

  8. Library design practices for success in lead generation with small molecule libraries.

    Science.gov (United States)

    Goodnow, R A; Guba, W; Haap, W

    2003-11-01

    The generation of novel structures amenable to rapid and efficient lead optimization comprises an emerging strategy for success in modern drug discovery. Small molecule libraries of sufficient size and diversity to increase the chances of discovery of novel structures make the high throughput synthesis approach the method of choice for lead generation. Despite an industry trend for smaller, more focused libraries, the need to generate novel lead structures makes larger libraries a necessary strategy. For libraries of a several thousand or more members, solid phase synthesis approaches are the most suitable. While the technology and chemistry necessary for small molecule library synthesis continue to advance, success in lead generation requires rigorous consideration in the library design process to ensure the synthesis of molecules possessing the proper characteristics for subsequent lead optimization. Without proper selection of library templates and building blocks, solid phase synthesis methods often generate molecules which are too heavy, too lipophilic and too complex to be useful for lead optimization. The appropriate filtering of virtual library designs with multiple computational tools allows the generation of information-rich libraries within a drug-like molecular property space. An understanding of the hit-to-lead process provides a practical guide to molecular design characteristics. Examples of leads generated from library approaches also provide a benchmarking of successes as well as aspects for continued development of library design practices.

  9. Combining on-chip synthesis of a focused combinatorial library with computational target prediction reveals imidazopyridine GPCR ligands.

    Science.gov (United States)

    Reutlinger, Michael; Rodrigues, Tiago; Schneider, Petra; Schneider, Gisbert

    2014-01-07

    Using the example of the Ugi three-component reaction we report a fast and efficient microfluidic-assisted entry into the imidazopyridine scaffold, where building block prioritization was coupled to a new computational method for predicting ligand-target associations. We identified an innovative GPCR-modulating combinatorial chemotype featuring ligand-efficient adenosine A1/2B and adrenergic α1A/B receptor antagonists. Our results suggest the tight integration of microfluidics-assisted synthesis with computer-based target prediction as a viable approach to rapidly generate bioactivity-focused combinatorial compound libraries with high success rates. Copyright © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  10. Starpc: a library for communication among tools on a parallel computer cluster. User's and developer's guide to Starpc

    International Nuclear Information System (INIS)

    Takemiya, Hiroshi; Yamagishi, Nobuhiro

    2000-02-01

    We report on a RPC(Remote Procedure Call)-based communication library, Starpc, for a parallel computer cluster. Starpc supports communication between Java Applets and C programs as well as between C programs. Starpc has the following three features. (1) It enables communication between Java Applets and C programs on an arbitrary computer without security violation, although Java Applets are supposed to communicate only with programs on the specific computer (Web server) in subject to a restriction on security. (2) Diverse network communication protocols are available on Starpc, because of using Nexus communication library developed at Argonne National Laboratory. (3) It works on many kinds of computers including eight parallel computers and four WS servers. In this report, the usage of Starpc and the development of applications using Starpc are described. (author)

  11. Application Portable Parallel Library

    Science.gov (United States)

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  12. HONEI: A collection of libraries for numerical computations targeting multiple processor architectures

    Science.gov (United States)

    van Dyk, Danny; Geveler, Markus; Mallach, Sven; Ribbrock, Dirk; Göddeke, Dominik; Gutwenger, Carsten

    2009-12-01

    We present HONEI, an open-source collection of libraries offering a hardware oriented approach to numerical calculations. HONEI abstracts the hardware, and applications written on top of HONEI can be executed on a wide range of computer architectures such as CPUs, GPUs and the Cell processor. We demonstrate the flexibility and performance of our approach with two test applications, a Finite Element multigrid solver for the Poisson problem and a robust and fast simulation of shallow water waves. By linking against HONEI's libraries, we achieve a two-fold speedup over straight forward C++ code using HONEI's SSE backend, and additional 3-4 and 4-16 times faster execution on the Cell and a GPU. A second important aspect of our approach is that the full performance capabilities of the hardware under consideration can be exploited by adding optimised application-specific operations to the HONEI libraries. HONEI provides all necessary infrastructure for development and evaluation of such kernels, significantly simplifying their development. Program summaryProgram title: HONEI Catalogue identifier: AEDW_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEDW_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GPLv2 No. of lines in distributed program, including test data, etc.: 216 180 No. of bytes in distributed program, including test data, etc.: 1 270 140 Distribution format: tar.gz Programming language: C++ Computer: x86, x86_64, NVIDIA CUDA GPUs, Cell blades and PlayStation 3 Operating system: Linux RAM: at least 500 MB free Classification: 4.8, 4.3, 6.1 External routines: SSE: none; [1] for GPU, [2] for Cell backend Nature of problem: Computational science in general and numerical simulation in particular have reached a turning point. The revolution developers are facing is not primarily driven by a change in (problem-specific) methodology, but rather by the fundamental paradigm shift of the

  13. The Development of E-Library at Delta State University Library, Abraka

    African Journals Online (AJOL)

    The study is on the automation of Delta University Library Abraka. The processes involved in the acquisition of information communication technology equipments to complement the traditional Library Services were stated. Major issues discussed include acquisition of computers hardware and library software, telephone ...

  14. Standardization of computer programs - basis of the Czechoslovak library of nuclear codes

    International Nuclear Information System (INIS)

    Gregor, M.

    1987-01-01

    A standardized form of computer code documentation has been established in the CSSR in the field of reactor safety. Structure and content of the documentation are described and codes already subject to this process are mentioned. The formation of a Czechoslovak nuclear code library and facilitated discussion of safety reports containing results of standardized codes are aimed at

  15. Evaluation of high-performance computing software

    Energy Technology Data Exchange (ETDEWEB)

    Browne, S.; Dongarra, J. [Univ. of Tennessee, Knoxville, TN (United States); Rowan, T. [Oak Ridge National Lab., TN (United States)

    1996-12-31

    The absence of unbiased and up to date comparative evaluations of high-performance computing software complicates a user`s search for the appropriate software package. The National HPCC Software Exchange (NHSE) is attacking this problem using an approach that includes independent evaluations of software, incorporation of author and user feedback into the evaluations, and Web access to the evaluations. We are applying this approach to the Parallel Tools Library (PTLIB), a new software repository for parallel systems software and tools, and HPC-Netlib, a high performance branch of the Netlib mathematical software repository. Updating the evaluations with feed-back and making it available via the Web helps ensure accuracy and timeliness, and using independent reviewers produces unbiased comparative evaluations difficult to find elsewhere.

  16. A Study on Education and the Organization of School Libraries in the Perspective of Total Quality Management

    Directory of Open Access Journals (Sweden)

    Cengiz Aslan

    2017-06-01

    Full Text Available The arrangement of enriched school libraries (Z Library established by the Republic of Turkey Ministry of National Education are analyzed in the perspective of total quality management in terms of institutionalization and supervision, goals and productivity.

  17. ASSESSMENT OF LIBRARY USERS’ FEEDBACK USING MODIFIED MULTILAYER PERCEPTRON NEURAL NETWORKS

    Directory of Open Access Journals (Sweden)

    K G Nandha Kumar

    2017-07-01

    Full Text Available An attempt has been made to evaluate the feedbacks of library users of four different libraries by using neural network based data mining techniques. This paper presents the results of a survey of users’ satisfactory level on four different libraries. The survey has been conducted among the users of four libraries of educational institutions of Kovai Medical Center Research and Educational Trust. Data were collected through questionnaires. Artificial neural network based data mining techniques are proposed and applied to assess the libraries in terms of level of satisfaction of users. In order to assess the users’ satisfaction level, two neural network techniques: Modified Multilayer Perceptron Network-Supervised and Modified Multilayer Perceptron Network-Unsupervised are proposed. The proposed techniques are compared with the conventional classification algorithm Multilayer Perceptron Neural Network and found better in overall performance. It is found that the quality of service provided by the libraries is highly good and users are highly satisfied with various aspects of library service. The Arts and Science College Library secured the maximum percent in terms of user satisfaction. This shows that the users’ satisfaction of ASCL is better than the other libraries. This study provides an insight into the actual quality and satisfactory level of users of libraries after proper assessment. It is strongly expected that the results will help library authorities to enhance services and quality in the near future.

  18. Curry’s Study on the Quality of Public Library Reference Service to LGBTQ Youth

    Directory of Open Access Journals (Sweden)

    Gregg A Stevens

    2018-02-01

    Full Text Available A Review of: Curry, A. (2005. If I ask, will they answer? Evaluating public library reference service to gay and lesbian youth. Reference & User Services Quarterly, 45(1, 65-75. Retrieved from https://www.jstor.org/journal/refuseserq Abstract Objective - To assess the quality of service provided by reference staff in public libraries when presented with a request for LGBTQ information by a young person. Design - Unobtrusive observation without informed consent. Setting - Public library branches in the greater Vancouver area, British Columbia, Canada. Subjects - Reference librarians. Methods - A 19-year-old posing as a high school student approached reference desk staff at 20 public library branches. The student proxy, “Angela”, was instructed to ask for books on forming a gay-straight alliance at her school and, if there was a full reference interview, to also ask for recommendations of novels that the group might read. She recorded the reactions, both verbal and nonverbal, using Reference and User Services Association guidelines as a template. Library administrators were aware of the potential visits and permitted the research, but the reference desk staff were not aware of a potential visit by the student proxy. The researcher claimed that her method, while deceptive, was necessary to obtain authentic reactions from the library staff. Main Results - Most reference librarians approached by Angela made adequate attempts to assist her, although a few library staff reacted negatively to her query. Half of the librarians reacted positively to the patron’s request, with most of the others providing neutral responses. Very few of the librarians actually taught the patron how to use the library’s catalog to search for materials, and most of the librarians were unable to find appropriate materials due to not knowing the appropriate search terms. Only three library staff showed overt disapproval of the search topic, such as frowning or rushing

  19. Improving Library Service Quality to Graduate Students: LibQual+[TM] Survey Results in a Practical Setting

    Science.gov (United States)

    Jankowska, Maria Anna; Hertel, Karen; Young, Nancy J.

    2006-01-01

    The LibQUAL+[TM] survey was conducted to determine user satisfaction and expectations concerning library service quality. The results of the "22 items and a box" constituted a rich source of information for the University of Idaho (UI) Library's strategic planning process. Focusing on graduate students, this study used three…

  20. OpenMM 4: A Reusable, Extensible, Hardware Independent Library for High Performance Molecular Simulation.

    Science.gov (United States)

    Eastman, Peter; Friedrichs, Mark S; Chodera, John D; Radmer, Randall J; Bruns, Christopher M; Ku, Joy P; Beauchamp, Kyle A; Lane, Thomas J; Wang, Lee-Ping; Shukla, Diwakar; Tye, Tony; Houston, Mike; Stich, Timo; Klein, Christoph; Shirts, Michael R; Pande, Vijay S

    2013-01-08

    OpenMM is a software toolkit for performing molecular simulations on a range of high performance computing architectures. It is based on a layered architecture: the lower layers function as a reusable library that can be invoked by any application, while the upper layers form a complete environment for running molecular simulations. The library API hides all hardware-specific dependencies and optimizations from the users and developers of simulation programs: they can be run without modification on any hardware on which the API has been implemented. The current implementations of OpenMM include support for graphics processing units using the OpenCL and CUDA frameworks. In addition, OpenMM was designed to be extensible, so new hardware architectures can be accommodated and new functionality (e.g., energy terms and integrators) can be easily added.

  1. Development of a Computational Steering Framework for High Performance Computing Environments on Blue Gene/P Systems

    KAUST Repository

    Danani, Bob K.

    2012-07-01

    Computational steering has revolutionized the traditional workflow in high performance computing (HPC) applications. The standard workflow that consists of preparation of an application’s input, running of a simulation, and visualization of simulation results in a post-processing step is now transformed into a real-time interactive workflow that significantly reduces development and testing time. Computational steering provides the capability to direct or re-direct the progress of a simulation application at run-time. It allows modification of application-defined control parameters at run-time using various user-steering applications. In this project, we propose a computational steering framework for HPC environments that provides an innovative solution and easy-to-use platform, which allows users to connect and interact with running application(s) in real-time. This framework uses RealityGrid as the underlying steering library and adds several enhancements to the library to enable steering support for Blue Gene systems. Included in the scope of this project is the development of a scalable and efficient steering relay server that supports many-to-many connectivity between multiple steered applications and multiple steering clients. Steered applications can range from intermediate simulation and physical modeling applications to complex computational fluid dynamics (CFD) applications or advanced visualization applications. The Blue Gene supercomputer presents special challenges for remote access because the compute nodes reside on private networks. This thesis presents an implemented solution and demonstrates it on representative applications. Thorough implementation details and application enablement steps are also presented in this thesis to encourage direct usage of this framework.

  2. CFDLIB05, Computational Fluid Dynamics Library

    International Nuclear Information System (INIS)

    Kashiwa, B.A.; Padial, N.T.; Rauenzahn, R.M.; VanderHeyden, W.B.

    2007-01-01

    1 - Description of program or function: CFDLib05 is the Los Alamos Computational Fluid Dynamics Library. This is a collection of hydro-codes using a common data structure and a common numerical method, for problems ranging from single-field, incompressible flow, to multi-species, multi-field, compressible flow. The data structure is multi-block, with a so-called structured grid in each block. The numerical method is a Finite-Volume scheme employing a state vector that is fully cell-centered. This means that the integral form of the conversation laws is solved on the physical domain that is represented by a mesh of control volumes. The typical control volume is an arbitrary quadrilateral in 2D and an arbitrary hexahedron in 3D. The Finite-Volume scheme is for time-unsteady flow and remains well coupled by means of time and space centered fluxes; if a steady state solution is required, the problem is integrated forward in time until the user is satisfied that the state is stationary. 2 - Methods: Cells-centered Implicit Continuous-fluid Eulerian (ICE) method

  3. zipHMMlib: a highly optimised HMM library exploiting repetitions in the input to speed up the forward algorithm.

    Science.gov (United States)

    Sand, Andreas; Kristiansen, Martin; Pedersen, Christian N S; Mailund, Thomas

    2013-11-22

    Hidden Markov models are widely used for genome analysis as they combine ease of modelling with efficient analysis algorithms. Calculating the likelihood of a model using the forward algorithm has worst case time complexity linear in the length of the sequence and quadratic in the number of states in the model. For genome analysis, however, the length runs to millions or billions of observations, and when maximising the likelihood hundreds of evaluations are often needed. A time efficient forward algorithm is therefore a key ingredient in an efficient hidden Markov model library. We have built a software library for efficiently computing the likelihood of a hidden Markov model. The library exploits commonly occurring substrings in the input to reuse computations in the forward algorithm. In a pre-processing step our library identifies common substrings and builds a structure over the computations in the forward algorithm which can be reused. This analysis can be saved between uses of the library and is independent of concrete hidden Markov models so one preprocessing can be used to run a number of different models.Using this library, we achieve up to 78 times shorter wall-clock time for realistic whole-genome analyses with a real and reasonably complex hidden Markov model. In one particular case the analysis was performed in less than 8 minutes compared to 9.6 hours for the previously fastest library. We have implemented the preprocessing procedure and forward algorithm as a C++ library, zipHMM, with Python bindings for use in scripts. The library is available at http://birc.au.dk/software/ziphmm/.

  4. Use of ETOG and ETOT computer codes for preparating the Library of LEOPARD with data from ENDFIB-IV

    International Nuclear Information System (INIS)

    Cunha Menezes Filho, A. da.

    1983-01-01

    The modifications carried out in the ETOT-3 and ETOG-3 computer codes used for preparating the thermal (172 energy groups) and epithermal (54 energy groups) libraries, respectivelly, of LEOPARD computer code, are presented. (M.C.K.) [pt

  5. GBParsy: A GenBank flatfile parser library with high speed

    Directory of Open Access Journals (Sweden)

    Kim Yeon-Ki

    2008-07-01

    Full Text Available Abstract Background GenBank flatfile (GBF format is one of the most popular sequence file formats because of its detailed sequence features and ease of readability. To use the data in the file by a computer, a parsing process is required and is performed according to a given grammar for the sequence and the description in a GBF. Currently, several parser libraries for the GBF have been developed. However, with the accumulation of DNA sequence information from eukaryotic chromosomes, parsing a eukaryotic genome sequence with these libraries inevitably takes a long time, due to the large GBF file and its correspondingly large genomic nucleotide sequence and related feature information. Thus, there is significant need to develop a parsing program with high speed and efficient use of system memory. Results We developed a library, GBParsy, which was C language-based and parses GBF files. The parsing speed was maximized by using content-specified functions in place of regular expressions that are flexible but slow. In addition, we optimized an algorithm related to memory usage so that it also increased parsing performance and efficiency of memory usage. GBParsy is at least 5 - 100× faster than current parsers in benchmark tests. Conclusion GBParsy is estimated to extract annotated information from almost 100 Mb of a GenBank flatfile for chromosomal sequence information within a second. Thus, it should be used for a variety of applications such as on-time visualization of a genome at a web site.

  6. Benchmarking high performance computing architectures with CMS’ skeleton framework

    Science.gov (United States)

    Sexton-Kennedy, E.; Gartung, P.; Jones, C. D.

    2017-10-01

    In 2012 CMS evaluated which underlying concurrency technology would be the best to use for its multi-threaded framework. The available technologies were evaluated on the high throughput computing systems dominating the resources in use at that time. A skeleton framework benchmarking suite that emulates the tasks performed within a CMSSW application was used to select Intel’s Thread Building Block library, based on the measured overheads in both memory and CPU on the different technologies benchmarked. In 2016 CMS will get access to high performance computing resources that use new many core architectures; machines such as Cori Phase 1&2, Theta, Mira. Because of this we have revived the 2012 benchmark to test it’s performance and conclusions on these new architectures. This talk will discuss the results of this exercise.

  7. FENDL multigroup libraries

    International Nuclear Information System (INIS)

    Ganesan, S.; Muir, D.W.

    1992-01-01

    Selected neutron reaction nuclear data libraries and photon-atomic interaction cross section libraries for elements of interest to the IAEA's program on Fusion Evaluated Nuclear Data Library (FENDL) have been processed into MATXSR format using the NJOY system on the VAX4000 computer of the IAEA. This document lists the resulting multigroup data libraries. All the multigroup data generated are available cost-free upon request from the IAEA Nuclear Data Section. (author). 9 refs

  8. High-performance computing on GPUs for resistivity logging of oil and gas wells

    Science.gov (United States)

    Glinskikh, V.; Dudaev, A.; Nechaev, O.; Surodina, I.

    2017-10-01

    We developed and implemented into software an algorithm for high-performance simulation of electrical logs from oil and gas wells using high-performance heterogeneous computing. The numerical solution of the 2D forward problem is based on the finite-element method and the Cholesky decomposition for solving a system of linear algebraic equations (SLAE). Software implementations of the algorithm used the NVIDIA CUDA technology and computing libraries are made, allowing us to perform decomposition of SLAE and find its solution on central processor unit (CPU) and graphics processor unit (GPU). The calculation time is analyzed depending on the matrix size and number of its non-zero elements. We estimated the computing speed on CPU and GPU, including high-performance heterogeneous CPU-GPU computing. Using the developed algorithm, we simulated resistivity data in realistic models.

  9. [cDNA library construction from panicle meristem of finger millet].

    Science.gov (United States)

    Radchuk, V; Pirko, Ia V; Isaenkov, S V; Emets, A I; Blium, Ia B

    2014-01-01

    The protocol for production of full-size cDNA using SuperScript Full-Length cDNA Library Construction Kit II (Invitrogen) was tested and high quality cDNA library from meristematic tissue of finger millet panicle (Eleusine coracana (L.) Gaertn) was created. The titer of obtained cDNA library comprised 3.01 x 10(5) CFU/ml in avarage. In average the length of cDNA insertion consisted about 1070 base pairs, the effectivity of cDNA fragment insertions--99.5%. The selective sequencing of cDNA clones from created library was performed. The sequences of cDNA clones were identified with usage of BLAST-search. The results of cDNA library analysis and selective sequencing represents prove good functionality and full length character of inserted cDNA clones. Obtained cDNA library from meristematic tissue of finger millet panicle represents good and valuable source for isolation and identification of key genes regulating metabolism and meristematic development and for mining of new molecular markers to conduct out high quality genetic investigations and molecular breeding as well.

  10. Wavelet subband coding of computer simulation output using the A++ array class library

    Energy Technology Data Exchange (ETDEWEB)

    Bradley, J.N.; Brislawn, C.M.; Quinlan, D.J.; Zhang, H.D. [Los Alamos National Lab., NM (United States); Nuri, V. [Washington State Univ., Pullman, WA (United States). School of EECS

    1995-07-01

    The goal of the project is to produce utility software for off-line compression of existing data and library code that can be called from a simulation program for on-line compression of data dumps as the simulation proceeds. Naturally, we would like the amount of CPU time required by the compression algorithm to be small in comparison to the requirements of typical simulation codes. We also want the algorithm to accomodate a wide variety of smooth, multidimensional data types. For these reasons, the subband vector quantization (VQ) approach employed in has been replaced by a scalar quantization (SQ) strategy using a bank of almost-uniform scalar subband quantizers in a scheme similar to that used in the FBI fingerprint image compression standard. This eliminates the considerable computational burdens of training VQ codebooks for each new type of data and performing nearest-vector searches to encode the data. The comparison of subband VQ and SQ algorithms in indicated that, in practice, there is relatively little additional gain from using vector as opposed to scalar quantization on DWT subbands, even when the source imagery is from a very homogeneous population, and our subjective experience with synthetic computer-generated data supports this stance. It appears that a careful study is needed of the tradeoffs involved in selecting scalar vs. vector subband quantization, but such an analysis is beyond the scope of this paper. Our present work is focused on the problem of generating wavelet transform/scalar quantization (WSQ) implementations that can be ported easily between different hardware environments. This is an extremely important consideration given the great profusion of different high-performance computing architectures available, the high cost associated with learning how to map algorithms effectively onto a new architecture, and the rapid rate of evolution in the world of high-performance computing.

  11. The Automation of Stochastization Algorithm with Use of SymPy Computer Algebra Library

    Science.gov (United States)

    Demidova, Anastasya; Gevorkyan, Migran; Kulyabov, Dmitry; Korolkova, Anna; Sevastianov, Leonid

    2018-02-01

    SymPy computer algebra library is used for automatic generation of ordinary and stochastic systems of differential equations from the schemes of kinetic interaction. Schemes of this type are used not only in chemical kinetics but also in biological, ecological and technical models. This paper describes the automatic generation algorithm with an emphasis on application details.

  12. Construction of CRISPR Libraries for Functional Screening.

    Science.gov (United States)

    Carstens, Carsten P; Felts, Katherine A; Johns, Sarah E

    2018-01-01

    Identification of gene function has been aided by the ability to generate targeted gene knockouts or transcriptional repression using the CRISPR/CAS9 system. Using pooled libraries of guide RNA expression vectors that direct CAS9 to a specific genomic site allows identification of genes that are either enriched or depleted in response to a selection scheme, thus linking the affected gene to the chosen phenotype. The quality of the data generated by the screening is dependent on the quality of the guide RNA delivery library with regards to error rates and especially evenness of distribution of the guides. Here, we describe a method for constructing complex plasmid libraries based on pooled designed oligomers with high representation and tight distributions. The procedure allows construction of plasmid libraries of >60,000 members with a 95th/5th percentile ratio of less than 3.5.

  13. The Steacie Library Dungeon Hackfest: Hackers in the Library Coding, Collaborating and Creating

    Directory of Open Access Journals (Sweden)

    Sarah Shujah

    2013-12-01

    Full Text Available The Steacie Science and Engineering Library at York University hosted its first annual Steacie Library Dungeon Hackfest in February 2013. The purpose of a hackfest is to spend a day (or longer using computer programming skills to collaborate on a particular software project and, hopefully, bring the project to reality. The project may be an app, widget, or website. It is evident that hackfests, as a form of engaged learning, help to reinforce the relevance of the library since it has implications for information literacy, open access, faculty liaison, and the changing perception of library as place. Twenty-five participants that included students, faculty, and staff hacked on Open York Data: York University’s openly accessible records and data such as course codes, lab stats of computers available in various libraries, subject headings, and research interests of faculty. The Hackfest produced several innovative ideas including a “Class Optimizer” program to help students generate a course schedule, and a mobile android app “YU Labs” that informs students of the computers available in a specific library. Library instruction was informally imparted during the Hackfest. There were discussions about open source, developing a research idea, and recommended tools to help students complete their projects. This article encapsulates the inspiration behind the Hackfest, discusses the library as the host, the resources used, the theme, achievements, and assessment. To summarize, the library Hackfest provided a space for literacy, life-long learning, and support of the open source pedagogy.

  14. Status of Medical Library Resources and Services in Teaching Hospitals in Enugu State, Nigeria: implications for quality health care services

    Directory of Open Access Journals (Sweden)

    Oluchi C. Okeke

    2017-06-01

    Full Text Available Considering the need for quality health information and resultant health care services in any society, this study was carried out to look into the status of library and information resources and services provided by medical libraries in Enugu State, Nigeria. The main objective of the study was to find out the information resources and services available for medical library users towards quality health care provision. Five (5 medical libraries of major teaching hospitals were used for the study with 980 registered users as the study population from where 245 users were sampled. Observation checklist was used to collect data on resources while questionnaire was used to collect data from the respondents on the services provided. The Medical Library Association Standard guided the analysis of data from observation while frequency counts and mean scores were used to analyze data from the questionnaire. Major findings showed that even though some of the required resources and services are available and provided the medical libraries, most of the required resources and services are not adequately provided by these libraries.

  15. library use instruction and the pattern of utilization of library services ...

    African Journals Online (AJOL)

    Global Journal

    the University of Calabar library, the vision of the parent institution is a .... prevalence of a user category devoid of basic information literacy ... because they might have been computer literate. 90 ..... Nigerian medical libraries: Challenges for ...

  16. Preparation and comparitive analysis of MCNP thermal libraries for liquid hydrogen and deuterium using NJOY97 on 32 bit and 64 bit computers

    International Nuclear Information System (INIS)

    Jo, Y. S.; Kim, J. D.; Kil, C. S.; Jang, J. H.

    1999-01-01

    The scattering laws and MCNP thermal libraries for liquid hydrogen and deuterium are comparatively calculated on HP715 (32-bit computer) and SGI IP27 (64-bit computer) using NJOY97. The results are also compared with the experimental data. In addition, MCNP calculations for the nuclear design of a cold neutron source at HANARO are performed with the newly generated MCNP thermal libraries from two different computers and the results are compared

  17. Electronic Journals in Academic Libraries: A Comparison of ARL and Non-ARL Libraries.

    Science.gov (United States)

    Shemberg, Marian; Grossman, Cheryl

    1999-01-01

    Describes a survey dealing with academic library provision of electronic journals and other electronic resources that compared ARL (Association of Research Libraries) members to non-ARL members. Highlights include full-text electronic journals; computers in libraries; online public access catalogs; interlibrary loan and electronic reserves; access…

  18. WIMSTAR-4: a computer program for generating WIMS library data from ENDF/B

    International Nuclear Information System (INIS)

    Wilkin, G.B.

    1981-08-01

    WIMSTAR (Version 4) is a FORTRAN-IV computer program developed to generate data files for the WIMS lattice code library from the ENDF/B data base. The program must be used in conjunction with the AMPX-II system and has been designed for implementation as a module of that system. This report describes the structure, implementation and use of the AMPX/WIMSTAR system

  19. Computer program CDCID: an automated quality control program using CDC update

    International Nuclear Information System (INIS)

    Singer, G.L.; Aguilar, F.

    1984-04-01

    A computer program, CDCID, has been developed in coordination with a quality control program to provide a highly automated method of documenting changes to computer codes at EG and G Idaho, Inc. The method uses the standard CDC UPDATE program in such a manner that updates and their associated documentation are easily made and retrieved in various formats. The method allows each card image of a source program to point to the document which describes it, who created the card, and when it was created. The method described is applicable to the quality control of computer programs in general. The computer program described is executable only on CDC computing systems, but the program could be modified and applied to any computing system with an adequate updating program

  20. Assertiveness and communication in the library environment: the case study of the City and University Library in Osijek

    Directory of Open Access Journals (Sweden)

    Srđan Lukačević

    2016-07-01

    Full Text Available The paper presents the findings of the City and University Library in Osijek research on the communication knowledge and skills required for the good communication between the library and its customers. Assertive communication is described as a means that can contribute to the positive perception of a library as an institution. The paper discusses the need for the librarians to acquire good communication skills in order to present the library and the array of its services (projects, departments, etc. in the best possible way to the public. This argument is supported by the findings of the web survey conducted via the library's Facebook page in 2012. The survey evaluated the communication skills of the librarians working in the City and University Library with the main goal of detecting problems in communication with library customers and ensure the high quality communication and service in the future.

  1. Using focus groups to investigate service quality determinants for customer satisfaction in selected university libraries in Sri Lanka

    Directory of Open Access Journals (Sweden)

    Chaminda Jayasundara

    2010-01-01

    Full Text Available This study aimed at establishing service quality determinants which may affect customer satisfaction in university libraries in Sri Lanka. Using the literature, 113 service quality determinants were identified. These were then reviewed by eight focus groups in four different universities. Forty of the determinants were perceived to be applicable to their context. The participants also added 14 quality requirements which they thought were not provided for in the list. Finally, the content and face validity of the 54 determinants were evaluated by a panel of experts who ultimately reduced them to 50. This study recommends the use of the identified quality determinants by library administrators and policymakers in the higher education sector in Sri Lanka to gauge the levels of customer satisfaction and assure quality of service.

  2. Development of automatic editing system for MCNP library 'autonj'

    International Nuclear Information System (INIS)

    Maekawa, Fujio; Sakurai, Kiyoshi; Kume, Etsuo; Nomura, Yasushi; Kosako, Kazuaki; Kawasaki, Nobuo; Naito, Yoshitaka

    1999-12-01

    As an activity of the MCNP High-Temperature Library Production Working Group under the Nuclear Code Evaluation Special Committee of Nuclear Code Committee, the automatic editing system for MCNP library 'autonj' was developed. The autonj includes the NJOY-97 code as its main body, and is a system that enables us to easily produce cross section libraries for MCNP from evaluated nuclear data files such as JENDL-3.2. A temperature dependent library at six temperature points based on JENDL-3.2 was produced by using autonj. The autonj system and the temperature dependent library were installed on the JAERI AP3000 computer. (author)

  3. MACKLIB-IV: a library of nuclear response functions generated with the MACK-IV computer program from ENDF/B-IV

    International Nuclear Information System (INIS)

    Gohar, Y.; Abdou, M.A.

    1978-03-01

    MACKLIB-IV employs the CTR energy group structure of 171 neutron groups and 36 gamma groups. A retrieval computer program is included with the library to permit collapsing into any other energy group structure. The library is in the new format of the ''MACK-Activity Table'' which uses a fixed position for each specific response function. This permits the user when employing the library with present transport codes to obtain directly the nuclear responses (e.g. the total nuclear heating) summed for all isotopes and integrated over any geometrical volume. The response functions included in the library are neutron kerma factor, gamma kerma factor, gas production and tritium-breeding functions, and all important reaction cross sections. Pertinent information about the library and a graphical display of six response functions for all materials in the library are given

  4. Assessment of Service Desk Quality at an Academic Health Sciences Library.

    Science.gov (United States)

    Blevins, Amy E; DeBerg, Jennifer; Kiscaden, Elizabeth

    2016-01-01

    Due to an identified need for formal assessment, a small team of librarians designed and administered a survey to gauge the quality of customer service at their academic health sciences library. Though results did not drive major changes to services, several important improvements were implemented and a process was established to serve as a foundation for future use. This article details the assessment process used and lessons learned during the project.

  5. Positive Correlation Between Academic Library Services and High-Impact Practices for

    Directory of Open Access Journals (Sweden)

    Saori Wendy Herman, MLIS, AHIP

    2016-03-01

    Full Text Available Objective – To investigate the perceived alignment between academic library services and high-impact practices (HIPs that affect student retention. Design – Survey questionnaire. Setting – Public comprehensive universities in the United States of America with a Carnegie classification of master’s level as of January 2013. Subjects – 68 library deans or directors out of the 271 who were originally contacted. Methods – The author used Qualtrics software to create a survey based on the HIPs, tested the survey for reliability, and then distributed it to 271 universities. Library services were grouped into 1 of 3 library scales: library collection, library instruction, or library facilities. The survey consisted of a matrix of 10 Likert-style questions addressing the perceived level of alignment between the library scales and the HIPs. Each question provided an opportunity for the respondent to enter a “brief description of support practices” (p 477. Additional demographic questions addressed the years of experience of the respondent, undergraduate student enrollment of the university, and whether librarians held faculty rank. Main Results – The author measured Pearson correlation coefficients and found a positive correlation between the library scales and the HIPs. All three library scales displayed a moderately strong positive correlation between first-year seminars and experiences (HIP 1, common intellectual experiences (HIP 2, writing-intensive courses (HIP 4, undergraduate research (HIP 6, diversity and global learning (HIP 7, service learning and community-based learning (HIP 8, internships (HIP 9, and capstone courses and projects (HIP 10. The library collections scale and library facilities scale displayed a moderately strong correlation with learning communities (HIP 3 and collaborative assignments and projects (HIP 5. The library instruction scale displayed a strong positive correlation with HIP 3 and a very strong

  6. Wikis in Libraries

    Directory of Open Access Journals (Sweden)

    Matthew Bejune

    2007-09-01

    Full Text Available Wikis have recently been adopted to support a variety of collaborative activities within libraries. This article and its companion wiki, LibraryWikis (http://librarywikis.pbwiki.com/, seek to document the phenomenon of wikis in libraries. This subject is considered within the framework of computer-supported cooperative work (CSCW. The author identified thirty-three library wikis and developed a classification schema with four categories: (1 collaboration among libraries (45.7 percent; (2 collaboration among library staff (31.4 percent; (3 collaboration among library staff and patrons (14.3 percent; and (4 collaboration among patrons (8.6 percent. Examples of library wikis are presented within the article, as is a discussion for why wikis are primarily utilized within categories I and II and not within categories III and IV. It is clear that wikis have great utility within libraries, and the author urges further application of wikis in libraries.

  7. Damsel: A Data Model Storage Library for Exascale Science

    Energy Technology Data Exchange (ETDEWEB)

    Choudhary, Alok [Northwestern Univ., Evanston, IL (United States); Liao, Wei-keng [Northwestern Univ., Evanston, IL (United States)

    2014-07-11

    Computational science applications have been described as having one of seven motifs (the “seven dwarfs”), each having a particular pattern of computation and communication. From a storage and I/O perspective, these applications can also be grouped into a number of data model motifs describing the way data is organized and accessed during simulation, analysis, and visualization. Major storage data models developed in the 1990s, such as Network Common Data Format (netCDF) and Hierarchical Data Format (HDF) projects, created support for more complex data models. Development of both netCDF and HDF5 was influenced by multi-dimensional dataset storage requirements, but their access models and formats were designed with sequential storage in mind (e.g., a POSIX I/O model). Although these and other high-level I/O libraries have had a beneficial impact on large parallel applications, they do not always attain a high percentage of peak I/O performance due to fundamental design limitations, and they do not address the full range of current and future computational science data models. The goal of this project is to enable exascale computational science applications to interact conveniently and efficiently with storage through abstractions that match their data models. The project consists of three major activities: (1) identifying major data model motifs in computational science applications and developing representative benchmarks; (2) developing a data model storage library, called Damsel, that supports these motifs, provides efficient storage data layouts, incorporates optimizations to enable exascale operation, and is tolerant to failures; and (3) productizing Damsel and working with computational scientists to encourage adoption of this library by the scientific community. The product of this project, Damsel library, is openly available for download from http://cucis.ece.northwestern.edu/projects/DAMSEL. Several case studies and application programming interface

  8. High-Performance Java Codes for Computational Fluid Dynamics

    Science.gov (United States)

    Riley, Christopher; Chatterjee, Siddhartha; Biswas, Rupak; Biegel, Bryan (Technical Monitor)

    2001-01-01

    The computational science community is reluctant to write large-scale computationally -intensive applications in Java due to concerns over Java's poor performance, despite the claimed software engineering advantages of its object-oriented features. Naive Java implementations of numerical algorithms can perform poorly compared to corresponding Fortran or C implementations. To achieve high performance, Java applications must be designed with good performance as a primary goal. This paper presents the object-oriented design and implementation of two real-world applications from the field of Computational Fluid Dynamics (CFD): a finite-volume fluid flow solver (LAURA, from NASA Langley Research Center), and an unstructured mesh adaptation algorithm (2D_TAG, from NASA Ames Research Center). This work builds on our previous experience with the design of high-performance numerical libraries in Java. We examine the performance of the applications using the currently available Java infrastructure and show that the Java version of the flow solver LAURA performs almost within a factor of 2 of the original procedural version. Our Java version of the mesh adaptation algorithm 2D_TAG performs within a factor of 1.5 of its original procedural version on certain platforms. Our results demonstrate that object-oriented software design principles are not necessarily inimical to high performance.

  9. Computer controlled quality of analytical measurements

    International Nuclear Information System (INIS)

    Clark, J.P.; Huff, G.A.

    1979-01-01

    A PDP 11/35 computer system is used in evaluating analytical chemistry measurements quality control data at the Barnwell Nuclear Fuel Plant. This computerized measurement quality control system has several features which are not available in manual systems, such as real-time measurement control, computer calculated bias corrections and standard deviation estimates, surveillance applications, evaluaton of measurement system variables, records storage, immediate analyst recertificaton, and the elimination of routine analysis of known bench standards. The effectiveness of the Barnwell computer system has been demonstrated in gathering and assimilating the measurements of over 1100 quality control samples obtained during a recent plant demonstration run. These data were used to determine equaitons for predicting measurement reliability estimates (bias and precision); to evaluate the measurement system; and to provide direction for modification of chemistry methods. The analytical chemistry measurement quality control activities represented 10% of the total analytical chemistry effort

  10. A standard library for modeling satellite orbits on a microcomputer

    Science.gov (United States)

    Beutel, Kenneth L.

    1988-03-01

    Introductory students of astrodynamics and the space environment are required to have a fundamental understanding of the kinematic behavior of satellite orbits. This thesis develops a standard library that contains the basic formulas for modeling earth orbiting satellites. This library is used as a basis for implementing a satellite motion simulator that can be used to demonstrate orbital phenomena in the classroom. Surveyed are the equations of orbital elements, coordinate systems and analytic formulas, which are made into a standard method for modeling earth orbiting satellites. The standard library is written in the C programming language and is designed to be highly portable between a variety of computer environments. The simulation draws heavily on the standards established by the library to produce a graphics-based orbit simulation program written for the Apple Macintosh computer. The simulation demonstrates the utility of the standard library functions but, because of its extensive use of the Macintosh user interface, is not portable to other operating systems.

  11. Analysis of Environmental Friendly Library Based on the Satisfaction and Service Quality: study at Library “X”

    Science.gov (United States)

    Herdiansyah, Herdis; Satriya Utama, Andre; Safruddin; Hidayat, Heri; Gema Zuliana Irawan, Angga; Immanuel Tjandra Muliawan, R.; Mutia Pratiwi, Diana

    2017-10-01

    One of the factor that influenced the development of science is the existence of the library, which in this case is the college libraries. Library, which is located in the college environment, aims to supply collections of literatures to support research activities as well as educational for students of the college. Conceptually, every library now starts to practice environmental principles. For example, “X” library as a central library claims to be an environmental friendly library for practicing environmental friendly management, but the X library has not inserted the satisfaction and service aspect to the users, including whether it is true that environmental friendly process is perceived by library users. Satisfaction can be seen from the comparison between expectations and reality of library users. This paper analyzes the level of library user satisfaction with library services in the campus area and the gap between expectations and reality felt by the library users. The result of the research shows that there is a disparity between the hope of library management, which is sustainable and environmentally friendly with the reality in the management of the library, so that it has not given satisfaction to the users yet. The gap value of satisfaction that has the biggest difference is in the library collection with the value of 1.57; while for the smallest gap value is in the same service to all students with a value of 0.67.

  12. Ground-glass opacity: High-resolution computed tomography and 64-multi-slice computed tomography findings comparison

    International Nuclear Information System (INIS)

    Sergiacomi, Gianluigi; Ciccio, Carmelo; Boi, Luca; Velari, Luca; Crusco, Sonia; Orlacchio, Antonio; Simonetti, Giovanni

    2010-01-01

    Objective: Comparative evaluation of ground-glass opacity using conventional high-resolution computed tomography technique and volumetric computed tomography by 64-row multi-slice scanner, verifying advantage of volumetric acquisition and post-processing technique allowed by 64-row CT scanner. Methods: Thirty-four patients, in which was assessed ground-glass opacity pattern by previous high-resolution computed tomography during a clinical-radiological follow-up for their lung disease, were studied by means of 64-row multi-slice computed tomography. Comparative evaluation of image quality was done by both CT modalities. Results: It was reported good inter-observer agreement (k value 0.78-0.90) in detection of ground-glass opacity with high-resolution computed tomography technique and volumetric Computed Tomography acquisition with moderate increasing of intra-observer agreement (k value 0.46) using volumetric computed tomography than high-resolution computed tomography. Conclusions: In our experience, volumetric computed tomography with 64-row scanner shows good accuracy in detection of ground-glass opacity, providing a better spatial and temporal resolution and advanced post-processing technique than high-resolution computed tomography.

  13. Evaluation of a transposase protocol for rapid generation of shotgun high-throughput sequencing libraries from nanogram quantities of DNA.

    Science.gov (United States)

    Marine, Rachel; Polson, Shawn W; Ravel, Jacques; Hatfull, Graham; Russell, Daniel; Sullivan, Matthew; Syed, Fraz; Dumas, Michael; Wommack, K Eric

    2011-11-01

    Construction of DNA fragment libraries for next-generation sequencing can prove challenging, especially for samples with low DNA yield. Protocols devised to circumvent the problems associated with low starting quantities of DNA can result in amplification biases that skew the distribution of genomes in metagenomic data. Moreover, sample throughput can be slow, as current library construction techniques are time-consuming. This study evaluated Nextera, a new transposon-based method that is designed for quick production of DNA fragment libraries from a small quantity of DNA. The sequence read distribution across nine phage genomes in a mock viral assemblage met predictions for six of the least-abundant phages; however, the rank order of the most abundant phages differed slightly from predictions. De novo genome assemblies from Nextera libraries provided long contigs spanning over half of the phage genome; in four cases where full-length genome sequences were available for comparison, consensus sequences were found to match over 99% of the genome with near-perfect identity. Analysis of areas of low and high sequence coverage within phage genomes indicated that GC content may influence coverage of sequences from Nextera libraries. Comparisons of phage genomes prepared using both Nextera and a standard 454 FLX Titanium library preparation protocol suggested that the coverage biases according to GC content observed within the Nextera libraries were largely attributable to bias in the Nextera protocol rather than to the 454 sequencing technology. Nevertheless, given suitable sequence coverage, the Nextera protocol produced high-quality data for genomic studies. For metagenomics analyses, effects of GC amplification bias would need to be considered; however, the library preparation standardization that Nextera provides should benefit comparative metagenomic analyses.

  14. Academic Libraries and High-Impact Practices for Student Retention: Library Deans' Perspectives

    Science.gov (United States)

    Murray, Adam

    2015-01-01

    Numerous studies on retention have highlighted the role of student engagement in influencing students' withdrawal decisions. This study seeks to address how academic libraries affect student retention by examining the perception of academic library deans or directors on the alignment between library services and resources with ten nationally…

  15. Overview of Parallel Platforms for Common High Performance Computing

    Directory of Open Access Journals (Sweden)

    T. Fryza

    2012-04-01

    Full Text Available The paper deals with various parallel platforms used for high performance computing in the signal processing domain. More precisely, the methods exploiting the multicores central processing units such as message passing interface and OpenMP are taken into account. The properties of the programming methods are experimentally proved in the application of a fast Fourier transform and a discrete cosine transform and they are compared with the possibilities of MATLAB's built-in functions and Texas Instruments digital signal processors with very long instruction word architectures. New FFT and DCT implementations were proposed and tested. The implementation phase was compared with CPU based computing methods and with possibilities of the Texas Instruments digital signal processing library on C6747 floating-point DSPs. The optimal combination of computing methods in the signal processing domain and new, fast routines' implementation is proposed as well.

  16. Efficient preparation of shuffled DNA libraries through recombination (Gateway) cloning.

    Science.gov (United States)

    Lehtonen, Soili I; Taskinen, Barbara; Ojala, Elina; Kukkurainen, Sampo; Rahikainen, Rolle; Riihimäki, Tiina A; Laitinen, Olli H; Kulomaa, Markku S; Hytönen, Vesa P

    2015-01-01

    Efficient and robust subcloning is essential for the construction of high-diversity DNA libraries in the field of directed evolution. We have developed a more efficient method for the subcloning of DNA-shuffled libraries by employing recombination cloning (Gateway). The Gateway cloning procedure was performed directly after the gene reassembly reaction, without additional purification and amplification steps, thus simplifying the conventional DNA shuffling protocols. Recombination-based cloning, directly from the heterologous reassembly reaction, conserved the high quality of the library and reduced the time required for the library construction. The described method is generally compatible for the construction of DNA-shuffled gene libraries. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  17. Customer-Centered Structuring in University Libraries and Personnel Satisfaction

    Directory of Open Access Journals (Sweden)

    Erol Yılmaz

    2013-11-01

    Full Text Available Total Quality Management (TQM aims at creating a client-centered organizati- on and providing customer satisfaction. This study reviews TQM and its applica- tions in university libraries and examines employees' satisfaction with their work respectively. It was hypothesized that library personnel is not satisfıed because university libraries lack customer-centered (users and personnel structuring. Survey method was used to gather data. The questionnaire was administered to the 66 employees of Hacettepe University (HU libraries. High-level administra- tors of HU libraries were also interviewed. Data w ere analyzed using “t test” and “one-way analysis ofvariance (ANOVA”.

  18. A computer program with graphical user interface to plot the multigroup cross sections of WIMS-D library

    International Nuclear Information System (INIS)

    Thiyagarajan, T.K.; Ganesan, S.; Jagannathan, V.; Karthikeyan, R.

    2002-01-01

    As a result of the IAEA Co-ordinated Research Programme entitled 'Final Stage of the WIMS Library Update Project', new and updated WIMS-D libraries based upon ENDF/B-VI.5, JENDL-3.2 and JEF-2.2 have become available. A project to prepare an exhaustive handbook of WIMS-D cross sections from old and new libraries has been taken up by the authors. As part of this project, we have developed a computer program XnWlup with user-friendly graphical interface to help the users of WIMS-D library to enable quick visualization of the plots of the energy dependence of the multigroup cross sections of any nuclide of interest. This software enables the user to generate and view the histogram of 69 multi-group cross sections as a function of neutron energy under Microsoft Windows environment. This software is designed using Microsoft Visual C++ and Microsoft Foundation Classes Library. The current features of the software, on-line help manual and future plans for further development are described in this paper

  19. High-pitch spiral computed tomography: effect on image quality and radiation dose in pediatric chest computed tomography.

    Science.gov (United States)

    Lell, Michael M; May, Matthias; Deak, Paul; Alibek, Sedat; Kuefner, Michael; Kuettner, Axel; Köhler, Henrik; Achenbach, Stephan; Uder, Michael; Radkow, Tanja

    2011-02-01

    computed tomography (CT) is considered the method of choice in thoracic imaging for a variety of indications. Sedation is usually necessary to enable CT and to avoid deterioration of image quality because of patient movement in small children. We evaluated a new, subsecond high-pitch scan mode (HPM), which obviates the need of sedation and to hold the breath. a total of 60 patients were included in this study. 30 patients (mean age, 14 ± 17 month; range, 0-55 month) were examined with a dual source CT system in an HPM. Scan parameters were as follows: pitch = 3.0, 128 × 0.6 mm slice acquisition, 0.28 seconds gantry rotation time, ref. mAs adapted to the body weight (50-100 mAs) at 80 kV. Images were reconstructed with a slice thickness of 0.75 mm. None of the children was sedated for the CT examination and no breathing instructions were given. Image quality was assessed focusing on motion artifacts and delineation of the vascular structures and lung parenchyma. Thirty patients (mean age, 15 ± 17 month; range, 0-55 month) were examined under sedation on 2 different CT systems (10-slice CT, n = 18; 64-slice CT, n = 13 patients) in conventional pitch mode (CPM). Dose values were calculated from the dose length product provided in the patient protocol/dose reports, Monte Carlo simulations were performed to assess dose distribution for CPM and HPM. all scans were performed without complications. Image quality was superior with HPM, because of a significant reduction in motion artifacts, as compared to CPM with 10- and 64-slice CT. In the control group, artifacts were encountered at the level of the diaphragm (n = 30; 100%), the borders of the heart (n = 30; 100%), and the ribs (n = 20; 67%) and spine (n = 6; 20%), whereas motion artifacts were detected in the HPM-group only in 6 patients in the lung parenchyma next to the diaphragm or the heart (P detector width and pitch-value. high-pitch chest CT is a robust method to provide highest image quality making sedation

  20. A platform independent communication library for distributed computing

    NARCIS (Netherlands)

    Groen, D.; Rieder, S.; Grosso, P.; de Laat, C.; Portegies Zwart, S.

    2010-01-01

    We present MPWide, a platform independent communication library for performing message passing between supercomputers. Our library couples several local MPI applications through a long distance network using, for example, optical links. The implementation is deliberately kept light-weight, platform

  1. The survey of academic libraries

    CERN Document Server

    2014-01-01

    The Survey of Academic Libraries, 2014-15 Edition looks closely at key benchmarks for academic libraries in areas such as spending for books and e-books, deployment and pay rates for student workers, use of tablet computers, cloud computing and other new technologies, database licensing practices, and much more. The study includes detailed data on overall budgets, capital budgets, salaries and materials spending, and much more of interest to academic librarians and their suppliers. Data in this 200+ page report is broken out by size and type of library for easy benchmarking.

  2. Microcomputers in the Anesthesia Library.

    Science.gov (United States)

    Wright, A. J.

    The combination of computer technology and library operation is helping to alleviate such library problems as escalating costs, increasing collection size, deteriorating materials, unwieldy arrangement schemes, poor subject control, and the acquisition and processing of large numbers of rarely used documents. Small special libraries such as…

  3. Academic Library Department Experience Fosters the Development of Leadership Skills Relevant to Academic Library Directorship

    OpenAIRE

    Joanne M. Muellenbach

    2017-01-01

    A Review of: Harris-Keith, Colleen S. (2015). The Relationship Between Academic Library Department Experience and Perceptions of Leadership Skill Development Relevant to Academic Library Directorship. The Journal of Academic Librarianship, 41(3), 246-263. doi:10.1016/j.acalib.2015.03.017 Objective – This study sought to identify if the perception of library leadership skill and quality development is equal across departmental experience, and what are the leadership skills and qualities...

  4. MARC and the Library Service Center: Automation at Bargain Rates.

    Science.gov (United States)

    Pearson, Karl M.

    Despite recent research and development in the field of library automation, libraries have been unable to reap the benefits promised by technology due to the high cost of building and maintaining their own computer-based systems. Time-sharing and disc mass storage devices will bring automation costs, if spread over a number of users, within the…

  5. Short Communication: Toward Improving Library and Information ...

    African Journals Online (AJOL)

    Nigerian libraries operate in an environment characterised by weak institutional and social support for libraries and the demanding challenges of the digital society. At the same time, the libraries need to embrace philosophies and strategies for developing quality systems and providing quality services. The paper discusses ...

  6. SCHOOL COMMUNITY PERCEPTION OF LIBRARY APPS AGAINTS LIBRARY EMPOWERMENT

    Directory of Open Access Journals (Sweden)

    Achmad Riyadi Alberto

    2017-07-01

    Full Text Available Abstract. This research is motivated by the development of information and communication technology (ICT in the library world so rapidly that allows libraries in the present to develop its services into digital-based services. This study aims to find out the school community’s perception of library apps developed by Riche Cynthia Johan, Hana Silvana, and Holin Sulistyo and its influence on library empowerment at the library of SD Laboratorium Percontohan UPI Bandung. Library apps in this research belong to the context of m-libraries, which is a library that meets the needs of its users by using mobile platforms such as smartphones,computers, and other mobile devices. Empowerment of library is the utilization of all aspects of the implementation of libraries to the best in order to achieve the expected goals. An analysis of the schoolcommunity’s perception of library apps using the Technology Acceptance Model (TAM includes: ease of use, usefulness, usability, usage trends, and real-use conditions. While the empowerment of the library includes aspects: information empowerment, empowerment of learning resources, empowerment of human resources, empowerment of library facilities, and library promotion. The research method used in this research is descriptive method with quantitative approach. Population and sample in this research is school community at SD Laboratorium Percontohan UPI Bandung. Determination of sample criteria by using disproportionate stratified random sampling with the number of samples of 83 respondents. Data analysis using simple linear regression to measure the influence of school community perception about library apps to library empowerment. The result of data analysis shows that there is influence between school community perception about library apps to library empowerment at library of SD Laboratorium Percontohan UPI Bandung which is proved by library acceptance level and library empowerment improvement.

  7. Chinese computerized nuclear data library

    International Nuclear Information System (INIS)

    Liang Qichang; Cai Dunjiu

    1996-01-01

    The Second Version of Chinese Evaluated Nuclear Data Library (CENDL-2) includes the complete neutron nuclear data sets of 54 important elements and isotopes used for nuclear science and engineering with the incident neutron energy from 10 -5 eV to 20 MeV, the international universal format ENDF/B-6 was adopted. Now, the Chinese Computerized nuclear data library has been developed and put into operation. That is, the users can make on-line use of the main data libraries for evaluated neutron reaction data in the world of EXFOR experimental nuclear data library on the terminal of computer via the perfect computer software system, carry out directly the nuclear engineering calculation or nuclear data evaluation, enjoy the use of the resource of our nuclear data libraries for their development of nuclear energy and nuclear technology applications

  8. Female Public Library Patrons Value the Library for Services, Programs, and Technology. A Review of: Fidishun, Dolores. “Women and the Public Library: Using Technology, Using the Library.” Library Trends 56.2 (2007: 328-43.

    Directory of Open Access Journals (Sweden)

    Virginia Wilson

    2009-03-01

    respondents who reported ages between 18 and 79 years. Seventy-one percent of these reported having a bachelor’s degree or higher. The study uses some contrasting data from the men’s responses where appropriate. In terms of library use, out of the 184 respondents, 42% came to the library monthly, while 36% visited the library weekly. Sixty-two percent of respondents knew they could email the library and 72% knew that they could call the library with questions. As for reasons for library use, the most prominent response was to borrow books rather than buying them. The second most common reason for using the library related to children’s books and programming for children. Other common reasons for library use included research activities, using public computers, reading, use of services such as photocopying and tax forms, and to volunteer or tutor. The library was also used as a place of solitude, where women could find a place and time for themselves. The author compared the men’s results to the women’s responses, and found that coming to the library for books was lower on the list, and very few men mentioned children’s library services. Men came to the library more often than women to study or read. In terms of using technology, the female respondents were fairly tech-savvy. Seventy-four percent of respondents felt comfortable using computers. Only 5% replied that using computers meant more work for them. Eighty-two percent said they used a computer on a regular basis, and 98% reported that they had used the Internet. Out of those who use the Internet, 91% used it at home, 64% used it at work, and 34% used it at the public library. Ninety-eight percent of women who used the Internet used a search engine such as Google or Yahoo to find information. Topics frequently mentioned were medical and travel information, information for their children, and shopping. Men, by contrast, listed shopping and finding medical information as their second reason for using the

  9. Applying Machine Learning and High Performance Computing to Water Quality Assessment and Prediction

    OpenAIRE

    Ruijian Zhang; Deren Li

    2017-01-01

    Water quality assessment and prediction is a more and more important issue. Traditional ways either take lots of time or they can only do assessments. In this research, by applying machine learning algorithm to a long period time of water attributes’ data; we can generate a decision tree so that it can predict the future day’s water quality in an easy and efficient way. The idea is to combine the traditional ways and the computer algorithms together. Using machine learning algorithms, the ass...

  10. Afghanistan Digital Library Initiative: Revitalizing an Integrated Library System

    Directory of Open Access Journals (Sweden)

    Yan HAN

    2007-12-01

    Full Text Available This paper describes an Afghanistan digital library initiative of building an integrated library system (ILS for Afghanistan universities and colleges based on open-source software. As one of the goals of the Afghan eQuality Digital Libraries Alliance, the authors applied systems analysis approach, evaluated different open-source ILSs, and customized the selected software to accommodate users’ needs. Improvements include Arabic and Persian language support, user interface changes, call number label printing, and ISBN-13 support. To our knowledge, this ILS is the first at a large academic library running on open-source software.

  11. Acoustic Sample Deposition MALDI-MS (ASD-MALDI-MS): A Novel Process Flow for Quality Control Screening of Compound Libraries.

    Science.gov (United States)

    Chin, Jefferson; Wood, Elizabeth; Peters, Grace S; Drexler, Dieter M

    2016-02-01

    In the early stages of drug discovery, high-throughput screening (HTS) of compound libraries against pharmaceutical targets is a common method to identify potential lead molecules. For these HTS campaigns to be efficient and successful, continuous quality control of the compound collection is necessary and crucial. However, the large number of compound samples and the limited sample amount pose unique challenges. Presented here is a proof-of-concept study for a novel process flow for the quality control screening of small-molecule compound libraries that consumes only minimal amounts of samples and affords compound-specific molecular data. This process employs an acoustic sample deposition (ASD) technique for the offline sample preparation by depositing nanoliter volumes in an array format onto microscope glass slides followed by matrix-assisted laser desorption/ionization mass spectrometric (MALDI-MS) analysis. An initial study of a 384-compound array employing the ASD-MALDI-MS workflow resulted in a 75% first-pass positive identification rate with an analysis time of <1 s per sample. © 2015 Society for Laboratory Automation and Screening.

  12. NINJA: Java for High Performance Numerical Computing

    Directory of Open Access Journals (Sweden)

    José E. Moreira

    2002-01-01

    Full Text Available When Java was first introduced, there was a perception that its many benefits came at a significant performance cost. In the particularly performance-sensitive field of numerical computing, initial measurements indicated a hundred-fold performance disadvantage between Java and more established languages such as Fortran and C. Although much progress has been made, and Java now can be competitive with C/C++ in many important situations, significant performance challenges remain. Existing Java virtual machines are not yet capable of performing the advanced loop transformations and automatic parallelization that are now common in state-of-the-art Fortran compilers. Java also has difficulties in implementing complex arithmetic efficiently. These performance deficiencies can be attacked with a combination of class libraries (packages, in Java that implement truly multidimensional arrays and complex numbers, and new compiler techniques that exploit the properties of these class libraries to enable other, more conventional, optimizations. Two compiler techniques, versioning and semantic expansion, can be leveraged to allow fully automatic optimization and parallelization of Java code. Our measurements with the NINJA prototype Java environment show that Java can be competitive in performance with highly optimized and tuned Fortran code.

  13. Quality control of next-generation sequencing library through an integrative digital microfluidic platform.

    Science.gov (United States)

    Thaitrong, Numrin; Kim, Hanyoup; Renzi, Ronald F; Bartsch, Michael S; Meagher, Robert J; Patel, Kamlesh D

    2012-12-01

    We have developed an automated quality control (QC) platform for next-generation sequencing (NGS) library characterization by integrating a droplet-based digital microfluidic (DMF) system with a capillary-based reagent delivery unit and a quantitative CE module. Using an in-plane capillary-DMF interface, a prepared sample droplet was actuated into position between the ground electrode and the inlet of the separation capillary to complete the circuit for an electrokinetic injection. Using a DNA ladder as an internal standard, the CE module with a compact LIF detector was capable of detecting dsDNA in the range of 5-100 pg/μL, suitable for the amount of DNA required by the Illumina Genome Analyzer sequencing platform. This DMF-CE platform consumes tenfold less sample volume than the current Agilent BioAnalyzer QC technique, preserving precious sample while providing necessary sensitivity and accuracy for optimal sequencing performance. The ability of this microfluidic system to validate NGS library preparation was demonstrated by examining the effects of limited-cycle PCR amplification on the size distribution and the yield of Illumina-compatible libraries, demonstrating that as few as ten cycles of PCR bias the size distribution of the library toward undesirable larger fragments. © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  14. Computer-assisted instruction: a library service for the community teaching hospital.

    Science.gov (United States)

    McCorkel, J; Cook, V

    1986-04-01

    This paper reports on five years of experience with computer-assisted instruction (CAI) at Winthrop-University Hospital, a major affiliate of the SUNY at Stony Brook School of Medicine. It compares CAI programs available from Ohio State University and Massachusetts General Hospital (accessed by telephone and modem), and software packages purchased from the Health Sciences Consortium (MED-CAPS) and Scientific American (DISCOTEST). The comparison documents one library's experience of the cost of these programs and the use made of them by medical students, house staff, and attending physicians. It describes the space allocated for necessary equipment, as well as the marketing of CAI. Finally, in view of the decision of the National Board of Medical Examiners to administer the Part III examination on computer (the so-called CBX) starting in 1988, the paper speculates on the future importance of CAI in the community teaching hospital.

  15. What makes computational open source software libraries successful?

    International Nuclear Information System (INIS)

    Bangerth, Wolfgang; Heister, Timo

    2013-01-01

    Software is the backbone of scientific computing. Yet, while we regularly publish detailed accounts about the results of scientific software, and while there is a general sense of which numerical methods work well, our community is largely unaware of best practices in writing the large-scale, open source scientific software upon which our discipline rests. This is particularly apparent in the commonly held view that writing successful software packages is largely the result of simply ‘being a good programmer’ when in fact there are many other factors involved, for example the social skill of community building. In this paper, we consider what we have found to be the necessary ingredients for successful scientific software projects and, in particular, for software libraries upon which the vast majority of scientific codes are built today. In particular, we discuss the roles of code, documentation, communities, project management and licenses. We also briefly comment on the impact on academic careers of engaging in software projects. (paper)

  16. What makes computational open source software libraries successful?

    Science.gov (United States)

    Bangerth, Wolfgang; Heister, Timo

    2013-01-01

    Software is the backbone of scientific computing. Yet, while we regularly publish detailed accounts about the results of scientific software, and while there is a general sense of which numerical methods work well, our community is largely unaware of best practices in writing the large-scale, open source scientific software upon which our discipline rests. This is particularly apparent in the commonly held view that writing successful software packages is largely the result of simply ‘being a good programmer’ when in fact there are many other factors involved, for example the social skill of community building. In this paper, we consider what we have found to be the necessary ingredients for successful scientific software projects and, in particular, for software libraries upon which the vast majority of scientific codes are built today. In particular, we discuss the roles of code, documentation, communities, project management and licenses. We also briefly comment on the impact on academic careers of engaging in software projects.

  17. BrAD-seq: Breath Adapter Directional sequencing: a streamlined, ultra-simple and fast library preparation protocol for strand specific mRNA library construction.

    Directory of Open Access Journals (Sweden)

    Brad Thomas Townsley

    2015-05-01

    Full Text Available Next Generation Sequencing (NGS is driving rapid advancement in biological understanding and RNA-sequencing (RNA-seq has become an indispensable tool for biology and medicine. There is a growing need for access to these technologies although preparation of NGS libraries remains a bottleneck to wider adoption. Here we report a novel method for the production of strand specific RNA-seq libraries utilizing inherent properties of double-stranded cDNA to capture and incorporate a sequencing adapter. Breath Adapter Directional sequencing (BrAD-seq reduces sample handling and requires far fewer enzymatic steps than most available methods to produce high quality strand-specific RNA-seq libraries. The method we present is optimized for 3-prime Digital Gene Expression (DGE libraries and can easily extend to full transcript coverage shotgun (SHO type strand-specific libraries and is modularized to accommodate a diversity of RNA and DNA input materials. BrAD-seq offers a highly streamlined and inexpensive option for RNA-seq libraries.

  18. Public Library Training Program for Older Adults Addresses Their Computer and Health Literacy Needs. A Review of: Xie, B. (2011. Improving older adults’ e-health literacy through computer training using NIH online resources. Library & Information Science Research, 34, 63-71. doi: /10.1016/j.lisr.2011.07.006

    Directory of Open Access Journals (Sweden)

    Cari Merkley

    2012-12-01

    Full Text Available Objective – To evaluate the efficacy of an ehealthliteracy educational intervention aimedat older adults.Design – Pre and post interventionquestionnaires administered in anexperimental study.Setting – Two public library branches inMaryland.Subjects – 218 adults between 60 and 89 yearsof age.Methods – A convenience sample of olderadults was recruited to participate in a fourweek training program structured around theNational Institutes of Health toolkit HelpingOlder Adults Search for Health InformationOnline. During the program, classes met at theparticipating libraries twice a week. Sessionswere two hours in length, and employedhands on exercises led by Master of LibraryScience students. The training included anintroduction to the Internet, as well as in depthtraining in the use of the NIHSeniorHealth andMedlinePlus websites. In the first class,participants were asked to complete a pretrainingquestionnaire that included questionsrelating to demographics and previouscomputer and Internet experience, as well asmeasures from the Computer Anxiety Scaleand two subscales of the Attitudes towardComputers Questionnaire. Participantsbetween September 2008 and June 2009 alsocompleted pre-training computer and web knowledge tests that asked individuals to label the parts of a computer and of a website using a provided list of terms. At the end of the program, participants were asked to complete post-training questionnaires that included the previously employed questions from the Computer Anxiety Scale and Attitudes towards Computer Questionnaire. New questions were added relating to the participants’ satisfaction with the training, its impact on their health decision making, their perceptions of public libraries, and the perceived usability and utility of the two websites highlighted during the training program. Those who completed pre-training knowledge tests were also asked to complete the same exercises at the end of the program.Main Results

  19. Evaluation of a Digital Library by Means of Quality Function Deployment (QFD) and the Kano Model

    Science.gov (United States)

    Garibay, Cecilia; Gutierrez, Humberto; Figueroa, Arturo

    2010-01-01

    This paper proposes utilizing a combination of the Quality Function Deployment (QFD)-Kano model as a useful tool to evaluate service quality. The digital library of the University of Guadalajara (Mexico) is presented as a case study. Data to feed the QFD-Kano model was gathered by an online questionnaire that was made available to users on the…

  20. High-content screening of yeast mutant libraries by shotgun lipidomics

    DEFF Research Database (Denmark)

    Tarasov, Kirill; Stefanko, Adam; Casanovas, Albert

    2014-01-01

    To identify proteins with a functional role in lipid metabolism and homeostasis we designed a high-throughput platform for high-content lipidomic screening of yeast mutant libraries. To this end, we combined culturing and lipid extraction in 96-well format, automated direct infusion...... factor KAR4 precipitated distinct lipid metabolic phenotypes. These results demonstrate that the high-throughput shotgun lipidomics platform is a valid and complementary proxy for high-content screening of yeast mutant libraries....... nanoelectrospray ionization, high-resolution Orbitrap mass spectrometry, and a dedicated data processing framework to support lipid phenotyping across hundreds of Saccharomyces cerevisiae mutants. Our novel approach revealed that the absence of genes with unknown function YBR141C and YJR015W, and the transcription...

  1. From Computer-interpretable Guidelines to Computer-interpretable Quality Indicators: A Case for an Ontology.

    Science.gov (United States)

    White, Pam; Roudsari, Abdul

    2014-01-01

    In the United Kingdom's National Health Service, quality indicators are generally measured electronically by using queries and data extraction, resulting in overlap and duplication of query components. Electronic measurement of health care quality indicators could be improved through an ontology intended to reduce duplication of effort during healthcare quality monitoring. While much research has been published on ontologies for computer-interpretable guidelines, quality indicators have lagged behind. We aimed to determine progress on the use of ontologies to facilitate computer-interpretable healthcare quality indicators. We assessed potential for improvements to computer-interpretable healthcare quality indicators in England. We concluded that an ontology for a large, diverse set of healthcare quality indicators could benefit the NHS and reduce workload, with potential lessons for other countries.

  2. Unified, Cross-Platform, Open-Source Library Package for High-Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Kozacik, Stephen [EM Photonics, Inc., Newark, DE (United States)

    2017-05-15

    Compute power is continually increasing, but this increased performance is largely found in sophisticated computing devices and supercomputer resources that are difficult to use, resulting in under-utilization. We developed a unified set of programming tools that will allow users to take full advantage of the new technology by allowing them to work at a level abstracted away from the platform specifics, encouraging the use of modern computing systems, including government-funded supercomputer facilities.

  3. Using Cloud Services for Library IT Infrastructure

    OpenAIRE

    Erik Mitchell

    2010-01-01

    Cloud computing comes in several different forms and this article documents how service, platform, and infrastructure forms of cloud computing have been used to serve library needs. Following an overview of these uses the article discusses the experience of one library in migrating IT infrastructure to a cloud environment and concludes with a model for assessing cloud computing.

  4. Biomathematical Description of Synthetic Peptide Libraries

    Science.gov (United States)

    Trepel, Martin

    2015-01-01

    Libraries of randomised peptides displayed on phages or viral particles are essential tools in a wide spectrum of applications. However, there is only limited understanding of a library's fundamental dynamics and the influences of encoding schemes and sizes on their quality. Numeric properties of libraries, such as the expected number of different peptides and the library's coverage, have long been in use as measures of a library's quality. Here, we present a graphical framework of these measures together with a library's relative efficiency to help to describe libraries in enough detail for researchers to plan new experiments in a more informed manner. In particular, these values allow us to answer-in a probabilistic fashion-the question of whether a specific library does indeed contain one of the "best" possible peptides. The framework is implemented in a web-interface based on two packages, discreteRV and peptider, to the statistical software environment R. We further provide a user-friendly web-interface called PeLiCa (Peptide Library Calculator, http://www.pelica.org), allowing scientists to plan and analyse their peptide libraries. PMID:26042419

  5. Computer Aided Reference Services in the Academic Library: Experiences in Organizing and Operating an Online Reference Service.

    Science.gov (United States)

    Hoover, Ryan E.

    1979-01-01

    Summarizes the development of the Computer-Aided Reference Services (CARS) division of the University of Utah Libraries' reference department. Development, organizational structure, site selection, equipment, management, staffing and training considerations, promotion and marketing, budget and pricing, record keeping, statistics, and evaluation…

  6. The Gender and Science Digital Library: Affecting Student Achievement in Science.

    Science.gov (United States)

    Nair, Sarita

    2003-01-01

    Describes the Gender and Science Digital Library (GSDL), an online collection of high-quality, interactive science resources that are gender-fair, inclusive, and engaging to students. Considers use by teachers and school library media specialists to encourage girls to enter careers in science, technology, engineering, and math (STEM). (LRW)

  7. Shelving: shelf reading and upkeep of library stock in high school ...

    African Journals Online (AJOL)

    Research findings and experience in Library Service has greatly shown shelve management as a tool for measuring performance, satisfaction and realization of set goal. This paper examines the state of shelving activities in five high school libraries at different local government areas of Lagos State. This includes statistics ...

  8. Development and influence of European and American university libraries

    Directory of Open Access Journals (Sweden)

    Irena Sapač

    2000-01-01

    Full Text Available The author compares the development of university libraries in Europe and in the United States of America. She finds that the university libraries in the United States of America have developed for three centuries under the influence of the European libraries, but now in the last century the European libraries have developed under the influence of the American ones. In times when there were no professional librarians, the American university libraries were managed by university professors, who were educated at European universities. The European management patterns were consequently applied also to the American libraries. The first books were also first brought from Europe. The Humboldt university also had a strong influence on the development of the American university libraries. Not until the second half of the 19th and especially the 20th century did the American university libraries achieve such high levels of cataloguing, classification, co-operation, organisation, computer networks, information holders, education and constructing library buildings that the European libraries started assuming their methods.

  9. AutoCNet: A Python library for sparse multi-image correspondence identification for planetary data

    Science.gov (United States)

    Laura, Jason; Rodriguez, Kelvin; Paquette, Adam C.; Dunn, Evin

    2018-01-01

    In this work we describe the AutoCNet library, written in Python, to support the application of computer vision techniques for n-image correspondence identification in remotely sensed planetary images and subsequent bundle adjustment. The library is designed to support exploratory data analysis, algorithm and processing pipeline development, and application at scale in High Performance Computing (HPC) environments for processing large data sets and generating foundational data products. We also present a brief case study illustrating high level usage for the Apollo 15 Metric camera.

  10. Dynamically-Loaded Hardware Libraries (HLL) Technology for Audio Applications

    DEFF Research Database (Denmark)

    Esposito, A.; Lomuscio, A.; Nunzio, L. Di

    2016-01-01

    In this work, we apply hardware acceleration to embedded systems running audio applications. We present a new framework, Dynamically-Loaded Hardware Libraries or HLL, to dynamically load hardware libraries on reconfigurable platforms (FPGAs). Provided a library of application-specific processors......, we load on-the-fly the specific processor in the FPGA, and we transfer the execution from the CPU to the FPGA-based accelerator. The proposed architecture provides excellent flexibility with respect to the different audio applications implemented, high quality audio, and an energy efficient solution....

  11. Computer Cataloging of Electronic Journals in Unstable Aggregator Databases: The Hong Kong Baptist University Library Experience.

    Science.gov (United States)

    Li, Yiu-On; Leung, Shirley W.

    2001-01-01

    Discussion of aggregator databases focuses on a project at the Hong Kong Baptist University library to integrate full-text electronic journal titles from three unstable aggregator databases into its online public access catalog (OPAC). Explains the development of the electronic journal computer program (EJCOP) to generate MARC records for…

  12. Academic Library Department Experience Fosters the Development of Leadership Skills Relevant to Academic Library Directorship

    Directory of Open Access Journals (Sweden)

    Joanne M. Muellenbach

    2017-03-01

    Full Text Available A Review of: Harris-Keith, Colleen S. (2015. The Relationship Between Academic Library Department Experience and Perceptions of Leadership Skill Development Relevant to Academic Library Directorship. The Journal of Academic Librarianship, 41(3, 246-263. doi:10.1016/j.acalib.2015.03.017 Objective – This study sought to identify if the perception of library leadership skill and quality development is equal across departmental experience, and what are the leadership skills and qualities most commonly perceived to be used in each department. Design – Quantitative online survey instrument. Setting – Master’s colleges and universities from 728 institutions in the United States of America, as classified by the Carnegie Foundation. Subjects – 666 academic library directors. Methods – Selected participants, representing academic library administrative leadership, were contacted by email a maximum of four times and were invited to complete an online survey instrument composed of six sections. The first three sections contained the purpose and confidentiality statements, demographic information, and data on the past five positions held by respondents prior to their current directorship. The next two sections each had 25 statements on a 5-point Likert scale, to collect data on perceived leadership skills and qualities exercised by respondents in their most recent three positions. The final section had four open-ended questions to help explain the academic library directors’ responses and provide context for the ratings in previous sections of the instrument. Main results – A total of 296 responses were received, for a 40.66% response rate, which was representative of the institution type demographics, including private non-profit, public, and private for-profit. The first research question asked: is the perception of library leadership skill and quality development equal across departmental experience? The data used for this question

  13. Reliability and Validity of SERVQUAL Scores Used To Evaluate Perceptions of Library Service Quality.

    Science.gov (United States)

    Thompson, Bruce; Cook, Colleen

    Research libraries are increasingly supplementing collection counts with perceptions of service quality as indices of status and productivity. The present study was undertaken to explore the reliability and validity of scores from the SERVQUAL measurement protocol (A. Parasuraman and others, 1991), which has previously been used in this type of…

  14. Library and Education

    Directory of Open Access Journals (Sweden)

    Gheorghe Buluţă

    2011-01-01

    Full Text Available The psycho-social phenomena generated by mass-media and the new information and communication technologies at the level of the young generations have led to new communication practices that bypass libraries and revolutionized the intellectual labor practices, with texts being rather used than read. In this context, our article examines the need to increase the library's role in developing the quality of education and research and brings to attention a few possible solutions which include a partnership between various types of libraries and between librarians' associations and NGOs to facilitate education through library and safeguard reading.

  15. Big Data and High-Performance Computing in Global Seismology

    Science.gov (United States)

    Bozdag, Ebru; Lefebvre, Matthieu; Lei, Wenjie; Peter, Daniel; Smith, James; Komatitsch, Dimitri; Tromp, Jeroen

    2014-05-01

    Much of our knowledge of Earth's interior is based on seismic observations and measurements. Adjoint methods provide an efficient way of incorporating 3D full wave propagation in iterative seismic inversions to enhance tomographic images and thus our understanding of processes taking place inside the Earth. Our aim is to take adjoint tomography, which has been successfully applied to regional and continental scale problems, further to image the entire planet. This is one of the extreme imaging challenges in seismology, mainly due to the intense computational requirements and vast amount of high-quality seismic data that can potentially be assimilated. We have started low-resolution inversions (T > 30 s and T > 60 s for body and surface waves, respectively) with a limited data set (253 carefully selected earthquakes and seismic data from permanent and temporary networks) on Oak Ridge National Laboratory's Cray XK7 "Titan" system. Recent improvements in our 3D global wave propagation solvers, such as a GPU version of the SPECFEM3D_GLOBE package, will enable us perform higher-resolution (T > 9 s) and longer duration (~180 m) simulations to take the advantage of high-frequency body waves and major-arc surface waves, thereby improving imbalanced ray coverage as a result of the uneven global distribution of sources and receivers. Our ultimate goal is to use all earthquakes in the global CMT catalogue within the magnitude range of our interest and data from all available seismic networks. To take the full advantage of computational resources, we need a solid framework to manage big data sets during numerical simulations, pre-processing (i.e., data requests and quality checks, processing data, window selection, etc.) and post-processing (i.e., pre-conditioning and smoothing kernels, etc.). We address the bottlenecks in our global seismic workflow, which are mainly coming from heavy I/O traffic during simulations and the pre- and post-processing stages, by defining new data

  16. Fundamentals of Library Automation and Technology. Participant Workbook.

    Science.gov (United States)

    Bridge, Frank; Walton, Robert

    This workbook presents outlines of topics to be covered during a two-day workshop on the fundamentals for library automation. Topics for the first day include: (1) Introduction; (2) Computer Technology--A Historical Overview; (3) Evolution of Library Automation; (4) Computer Hardware Technology--An Introduction; (5) Computer Software…

  17. Use Of Computer Among Library Staff In Four Universities Of ...

    African Journals Online (AJOL)

    4) selected Universities of Technology Libraries in Northern Nigeria. Survey research was adopted with population of 151 Library staff and a random sample size of 120 staff in four (4) selected Universities of Technology Libraries in Northern ...

  18. UMPHE: A Library for Effective Computing On Encrypted Data

    Directory of Open Access Journals (Sweden)

    Philipp Borisovich Burtyka

    2016-03-01

    Full Text Available The paper describes the design and implementation of a new software library that implements fully homomorphic encryption schemes based on unilateral matrix polynomials. The library is written in C++ using the NTL mathematical library and has multilayer structure. The main focus is on optimizations and batching techniques. The paper presents novel encryption and key generation algorithms for matrix polynomial based cryptosystems and techniques for data movement between the slots of the ciphertext.

  19. DNA-encoded chemical libraries: advancing beyond conventional small-molecule libraries.

    Science.gov (United States)

    Franzini, Raphael M; Neri, Dario; Scheuermann, Jörg

    2014-04-15

    DNA-encoded chemical libraries (DECLs) represent a promising tool in drug discovery. DECL technology allows the synthesis and screening of chemical libraries of unprecedented size at moderate costs. In analogy to phage-display technology, where large antibody libraries are displayed on the surface of filamentous phage and are genetically encoded in the phage genome, DECLs feature the display of individual small organic chemical moieties on DNA fragments serving as amplifiable identification barcodes. The DNA-tag facilitates the synthesis and allows the simultaneous screening of very large sets of compounds (up to billions of molecules), because the hit compounds can easily be identified and quantified by PCR-amplification of the DNA-barcode followed by high-throughput DNA sequencing. Several approaches have been used to generate DECLs, differing both in the methods used for library encoding and for the combinatorial assembly of chemical moieties. For example, DECLs can be used for fragment-based drug discovery, displaying a single molecule on DNA or two chemical moieties at the extremities of complementary DNA strands. DECLs can vary substantially in the chemical structures and the library size. While ultralarge libraries containing billions of compounds have been reported containing four or more sets of building blocks, also smaller libraries have been shown to be efficient for ligand discovery. In general, it has been found that the overall library size is a poor predictor for library performance and that the number and diversity of the building blocks are rather important indicators. Smaller libraries consisting of two to three sets of building blocks better fulfill the criteria of drug-likeness and often have higher quality. In this Account, we present advances in the DECL field from proof-of-principle studies to practical applications for drug discovery, both in industry and in academia. DECL technology can yield specific binders to a variety of target

  20. TJ-II Library Manual

    International Nuclear Information System (INIS)

    Tribaldos, V.; Milligen, B.Ph., van; Lopez-Fraguas, A.

    1996-01-01

    This report contains a detailed description of the TJ-II library and its routines. The library is written in FORTRAN 77 language and is available in the CRAY J916 and DEC Alpha 8400 computers at CIEMAT. This document also contains some examples of its use. (Author)

  1. Single-tube library preparation for degraded DNA

    DEFF Research Database (Denmark)

    Carøe, Christian; Gopalakrishnan, Shyam; Vinner, Lasse

    2018-01-01

    these obstacles and enable higher throughput are therefore of interest to researchers working with degraded DNA. 2.In this study, we compare four Illumina library preparation protocols, including two “single-tube” methods developed for this study with the explicit aim of improving data quality and reducing...... of chemically damaged and highly fragmented DNA molecules. In particular, the enzymatic reactions and DNA purification steps during library preparation can result in DNA template loss and sequencing biases, affecting downstream analyses. The development of library preparation methods that circumvent...... preparation time and expenses. The methods are tested on grey wolf (Canis lupus) museum specimens. 3.We found single-tube protocols increase library complexity, yield more reads that map uniquely to the reference genome, reduce processing time, and may decrease laboratory costs by 90%. 4.Given the advantages...

  2. Google vs. the Library (Part III): Assessing the Quality of Sources Found by Undergraduates

    Science.gov (United States)

    Georgas, Helen

    2015-01-01

    This study assesses and compares the quality of sources found by undergraduate students when doing research using both Google and a library (federated) search tool. Thirty undergraduates were asked to find four sources (one book, two articles, and one additional source of their choosing) related to a selected research topic. Students used both…

  3. Quantitative and quality test of cross section library ENDF/B-b2

    International Nuclear Information System (INIS)

    Zajac, R.; Necas, V.

    2006-01-01

    This article includes a test or in other words data verification of neutron ENDF/B-VIIb2 sub library. The first part consists from the process of preparation ACE files by NJOY 99.90. The starting point of data verification describes needed patches in NJOY 99.90, which are necessary to do for correctly production of ACE files. After the obtaining ACE files follow the test of all ACE files through GODIVA - input file for MCNP. GODIVA is high enrichment sphere of U-235, where every material is added as impurity. The aim of GODIVA test is to obtain a certainty if produced ACE files are able to run through MCNP. The second part of this article begins with choose of benchmarks from 'International Handbook of Evaluated Criticality Safety Benchmark Experiments, 2005'. From this source of criticality experiments were separated some benchmarks for quality verification of ACE files by MCNP (Authors)

  4. Hybrid Direct and Iterative Solver with Library of Multi-criteria Optimal Orderings for h Adaptive Finite Element Method Computations

    KAUST Repository

    AbouEisha, Hassan M.

    2016-06-02

    In this paper we present a multi-criteria optimization of element partition trees and resulting orderings for multi-frontal solver algorithms executed for two dimensional h adaptive finite element method. In particular, the problem of optimal ordering of elimination of rows in the sparse matrices resulting from adaptive finite element method computations is reduced to the problem of finding of optimal element partition trees. Given a two dimensional h refined mesh, we find all optimal element partition trees by using the dynamic programming approach. An element partition tree defines a prescribed order of elimination of degrees of freedom over the mesh. We utilize three different metrics to estimate the quality of the element partition tree. As the first criterion we consider the number of floating point operations(FLOPs) performed by the multi-frontal solver. As the second criterion we consider the number of memory transfers (MEMOPS) performed by the multi-frontal solver algorithm. As the third criterion we consider memory usage (NONZEROS) of the multi-frontal direct solver. We show the optimization results for FLOPs vs MEMOPS as well as for the execution time estimated as FLOPs+100MEMOPS vs NONZEROS. We obtain Pareto fronts with multiple optimal trees, for each mesh, and for each refinement level. We generate a library of optimal elimination trees for small grids with local singularities. We also propose an algorithm that for a given large mesh with identified local sub-grids, each one with local singularity. We compute Schur complements over the sub-grids using the optimal trees from the library, and we submit the sequence of Schur complements into the iterative solver ILUPCG.

  5. A novel heavy domain antibody library with functionally optimized complementarity determining regions.

    Directory of Open Access Journals (Sweden)

    Ole Aalund Mandrup

    Full Text Available Today a number of synthetic antibody libraries of different formats have been created and used for the selection of a large number of recombinant antibodies. One of the determining factors for successful isolation of recombinant antibodies from libraries lies in the quality of the libraries i.e. the number of correctly folded, functional antibodies contained in the library. Here, we describe the construction of a novel, high quality, synthetic single domain antibody library dubbed Predator. The library is based on the HEL4 domain antibody with the addition of recently reported mutations concerning the amino acid composition at positions critical for the folding characteristics and aggregation propensities of domain antibodies. As a unique feature, the CDR3 of the library was designed to mimic the natural human immune response by designating amino acids known to be prevalent in functional antibodies to the diversity in CDR3. CDR randomizations were performed using trinucleotide synthesis to avoid the presence of stop codons. Furthermore a novel cycle free elongation method was used for the conversion of the synthesized single stranded DNA containing the randomized CDRs into double stranded DNA of the library. In addition a modular approach has been adopted for the scaffold in which each CDR region is flanked by unique restrictions sites, allowing easy affinity maturation of selected clones by CDR shuffling. To validate the quality of the library, one round phage display selections were performed on purified antigens and highly complex antigen mixtures such as cultured eukaryotic cells resulting in several specific binders. The further characterization of some of the selected clones, however, indicates a reduction in thermodynamic stability caused by the inclusion the additional mutations to the HEL4 scaffold.

  6. MATH77 - A LIBRARY OF MATHEMATICAL SUBPROGRAMS FOR FORTRAN 77, RELEASE 4.0

    Science.gov (United States)

    Lawson, C. L.

    1994-01-01

    MATH77 is a high quality library of ANSI FORTRAN 77 subprograms implementing contemporary algorithms for the basic computational processes of science and engineering. The portability of MATH77 meets the needs of present-day scientists and engineers who typically use a variety of computing environments. Release 4.0 of MATH77 contains 454 user-callable and 136 lower-level subprograms. Usage of the user-callable subprograms is described in 69 sections of the 416 page users' manual. The topics covered by MATH77 are indicated by the following list of chapter titles in the users' manual: Mathematical Functions, Pseudo-random Number Generation, Linear Systems of Equations and Linear Least Squares, Matrix Eigenvalues and Eigenvectors, Matrix Vector Utilities, Nonlinear Equation Solving, Curve Fitting, Table Look-Up and Interpolation, Definite Integrals (Quadrature), Ordinary Differential Equations, Minimization, Polynomial Rootfinding, Finite Fourier Transforms, Special Arithmetic , Sorting, Library Utilities, Character-based Graphics, and Statistics. Besides subprograms that are adaptations of public domain software, MATH77 contains a number of unique packages developed by the authors of MATH77. Instances of the latter type include (1) adaptive quadrature, allowing for exceptional generality in multidimensional cases, (2) the ordinary differential equations solver used in spacecraft trajectory computation for JPL missions, (3) univariate and multivariate table look-up and interpolation, allowing for "ragged" tables, and providing error estimates, and (4) univariate and multivariate derivative-propagation arithmetic. MATH77 release 4.0 is a subroutine library which has been carefully designed to be usable on any computer system that supports the full ANSI standard FORTRAN 77 language. It has been successfully implemented on a CRAY Y/MP computer running UNICOS, a UNISYS 1100 computer running EXEC 8, a DEC VAX series computer running VMS, a Sun4 series computer running Sun

  7. Computer-supported quality control in X-ray diagnosis

    International Nuclear Information System (INIS)

    Maier, W.; Klotz, E.

    1989-01-01

    Quality control of X-ray facilities in radiological departments of large hospitals is possible only if the instrumentation used for measurements is interfaced to a computer. The central computer helps to organize the measurements as well as analyse and record the results. It can also be connected to a densitometer and camera for evaluating radiographs of test devices. Other quality control tests are supported by a mobile station with equipment for non-invasive dosimetry measurements. Experience with a computer-supported system in quality control of film and film processing is described and the evaluation methods of ANSI and the German industrial standard DIN are compared. The disadvantage of these methods is the exclusion of film quality parameters, which can make processing control almost worthless. (author)

  8. Libraries as a venue for exciting education technology, both high tech and low

    Science.gov (United States)

    Harold, J. B.; Dusenbery, P.; Holland, A.

    2016-12-01

    Public libraries provide a broad range of possibilities for reaching diverse audiences with NASA and STEM related content and activities, from hands-on activities, to interactive kiosks incorporating science based games, simulations, and real-time data. NCIL/SSI has been developing STEM-based exhibits and program activities for public libraries since 2007, and is currently managing 7 national tours in partnership with the American Library Association and other organizations. Past and current exhibitions will reach over 100 libraries and an estimated 1.5 million patrons. In this paper we will discuss a range of findings from almost a decade of deploying both high and low tech STEM learning strategies into libraries, including usage and engagement by library patrons, and challenges (and solutions) for deploying technologically sophisticated components into libraries which may or may not have dedicated technical staff.

  9. A Framework for Evaluating Digital Library Services; Interdisciplinarity: The Road Ahead for Education in Digital Libraries; Federated Digital Rights Management: A Proposed DRM Solution for Research and Education; Learning Lessons Holistically in the Glasgow Digital Library.

    Science.gov (United States)

    Choudhury, Sayeed; Hobbs, Benjamin; Lorie, Mark; Flores, Nicholas; Coleman, Anita; Martin, Mairead; Kuhlman, David L.; McNair, John H.; Rhodes, William A.; Tipton, Ron; Agnew, Grace; Nicholson, Dennis; Macgregor, George

    2002-01-01

    Includes four articles that address issues related to digital libraries. Highlights include a framework for evaluating digital library services, particularly academic research libraries; interdisciplinary approaches to education about digital libraries that includes library and information science and computing; digital rights management; and the…

  10. What Physicists Should Know About High Performance Computing - Circa 2002

    Science.gov (United States)

    Frederick, Donald

    2002-08-01

    High Performance Computing (HPC) is a dynamic, cross-disciplinary field that traditionally has involved applied mathematicians, computer scientists, and others primarily from the various disciplines that have been major users of HPC resources - physics, chemistry, engineering, with increasing use by those in the life sciences. There is a technological dynamic that is powered by economic as well as by technical innovations and developments. This talk will discuss practical ideas to be considered when developing numerical applications for research purposes. Even with the rapid pace of development in the field, the author believes that these concepts will not become obsolete for a while, and will be of use to scientists who either are considering, or who have already started down the HPC path. These principles will be applied in particular to current parallel HPC systems, but there will also be references of value to desktop users. The talk will cover such topics as: computing hardware basics, single-cpu optimization, compilers, timing, numerical libraries, debugging and profiling tools and the emergence of Computational Grids.

  11. Ultra-High-Resolution Computed Tomography of the Lung: Image Quality of a Prototype Scanner

    Science.gov (United States)

    Kakinuma, Ryutaro; Moriyama, Noriyuki; Muramatsu, Yukio; Gomi, Shiho; Suzuki, Masahiro; Nagasawa, Hirobumi; Kusumoto, Masahiko; Aso, Tomohiko; Muramatsu, Yoshihisa; Tsuchida, Takaaki; Tsuta, Koji; Maeshima, Akiko Miyagi; Tochigi, Naobumi; Watanabe, Shun-ichi; Sugihara, Naoki; Tsukagoshi, Shinsuke; Saito, Yasuo; Kazama, Masahiro; Ashizawa, Kazuto; Awai, Kazuo; Honda, Osamu; Ishikawa, Hiroyuki; Koizumi, Naoya; Komoto, Daisuke; Moriya, Hiroshi; Oda, Seitaro; Oshiro, Yasuji; Yanagawa, Masahiro; Tomiyama, Noriyuki; Asamura, Hisao

    2015-01-01

    Purpose The image noise and image quality of a prototype ultra-high-resolution computed tomography (U-HRCT) scanner was evaluated and compared with those of conventional high-resolution CT (C-HRCT) scanners. Materials and Methods This study was approved by the institutional review board. A U-HRCT scanner prototype with 0.25 mm x 4 rows and operating at 120 mAs was used. The C-HRCT images were obtained using a 0.5 mm x 16 or 0.5 mm x 64 detector-row CT scanner operating at 150 mAs. Images from both scanners were reconstructed at 0.1-mm intervals; the slice thickness was 0.25 mm for the U-HRCT scanner and 0.5 mm for the C-HRCT scanners. For both scanners, the display field of view was 80 mm. The image noise of each scanner was evaluated using a phantom. U-HRCT and C-HRCT images of 53 images selected from 37 lung nodules were then observed and graded using a 5-point score by 10 board-certified thoracic radiologists. The images were presented to the observers randomly and in a blinded manner. Results The image noise for U-HRCT (100.87 ± 0.51 Hounsfield units [HU]) was greater than that for C-HRCT (40.41 ± 0.52 HU; P < .0001). The image quality of U-HRCT was graded as superior to that of C-HRCT (P < .0001) for all of the following parameters that were examined: margins of subsolid and solid nodules, edges of solid components and pulmonary vessels in subsolid nodules, air bronchograms, pleural indentations, margins of pulmonary vessels, edges of bronchi, and interlobar fissures. Conclusion Despite a larger image noise, the prototype U-HRCT scanner had a significantly better image quality than the C-HRCT scanners. PMID:26352144

  12. Ultra-High-Resolution Computed Tomography of the Lung: Image Quality of a Prototype Scanner.

    Directory of Open Access Journals (Sweden)

    Ryutaro Kakinuma

    Full Text Available The image noise and image quality of a prototype ultra-high-resolution computed tomography (U-HRCT scanner was evaluated and compared with those of conventional high-resolution CT (C-HRCT scanners.This study was approved by the institutional review board. A U-HRCT scanner prototype with 0.25 mm x 4 rows and operating at 120 mAs was used. The C-HRCT images were obtained using a 0.5 mm x 16 or 0.5 mm x 64 detector-row CT scanner operating at 150 mAs. Images from both scanners were reconstructed at 0.1-mm intervals; the slice thickness was 0.25 mm for the U-HRCT scanner and 0.5 mm for the C-HRCT scanners. For both scanners, the display field of view was 80 mm. The image noise of each scanner was evaluated using a phantom. U-HRCT and C-HRCT images of 53 images selected from 37 lung nodules were then observed and graded using a 5-point score by 10 board-certified thoracic radiologists. The images were presented to the observers randomly and in a blinded manner.The image noise for U-HRCT (100.87 ± 0.51 Hounsfield units [HU] was greater than that for C-HRCT (40.41 ± 0.52 HU; P < .0001. The image quality of U-HRCT was graded as superior to that of C-HRCT (P < .0001 for all of the following parameters that were examined: margins of subsolid and solid nodules, edges of solid components and pulmonary vessels in subsolid nodules, air bronchograms, pleural indentations, margins of pulmonary vessels, edges of bronchi, and interlobar fissures.Despite a larger image noise, the prototype U-HRCT scanner had a significantly better image quality than the C-HRCT scanners.

  13. Simultaneous digital quantification and fluorescence-based size characterization of massively parallel sequencing libraries.

    Science.gov (United States)

    Laurie, Matthew T; Bertout, Jessica A; Taylor, Sean D; Burton, Joshua N; Shendure, Jay A; Bielas, Jason H

    2013-08-01

    Due to the high cost of failed runs and suboptimal data yields, quantification and determination of fragment size range are crucial steps in the library preparation process for massively parallel sequencing (or next-generation sequencing). Current library quality control methods commonly involve quantification using real-time quantitative PCR and size determination using gel or capillary electrophoresis. These methods are laborious and subject to a number of significant limitations that can make library calibration unreliable. Herein, we propose and test an alternative method for quality control of sequencing libraries using droplet digital PCR (ddPCR). By exploiting a correlation we have discovered between droplet fluorescence and amplicon size, we achieve the joint quantification and size determination of target DNA with a single ddPCR assay. We demonstrate the accuracy and precision of applying this method to the preparation of sequencing libraries.

  14. Modular high-throughput test stand for versatile screening of thin-film materials libraries

    International Nuclear Information System (INIS)

    Thienhaus, Sigurd; Hamann, Sven; Ludwig, Alfred

    2011-01-01

    Versatile high-throughput characterization tools are required for the development of new materials using combinatorial techniques. Here, we describe a modular, high-throughput test stand for the screening of thin-film materials libraries, which can carry out automated electrical, magnetic and magnetoresistance measurements in the temperature range of −40 to 300 °C. As a proof of concept, we measured the temperature-dependent resistance of Fe–Pd–Mn ferromagnetic shape-memory alloy materials libraries, revealing reversible martensitic transformations and the associated transformation temperatures. Magneto-optical screening measurements of a materials library identify ferromagnetic samples, whereas resistivity maps support the discovery of new phases. A distance sensor in the same setup allows stress measurements in materials libraries deposited on cantilever arrays. A combination of these methods offers a fast and reliable high-throughput characterization technology for searching for new materials. Using this approach, a composition region has been identified in the Fe–Pd–Mn system that combines ferromagnetism and martensitic transformation.

  15. School Libraries...Unfinished Business: A Report on New York City's Elementary School Libraries.

    Science.gov (United States)

    Fund for New York City Public Education.

    Effective school library media centers are a cost-efficient way for schools to provide children with the sophisticated knowledge, research, and computer skills they will need for the growing demands of work and citizenship in the 21st century. The libraries' cultivation of literacy, research, and thinking skills is particularly crucial in…

  16. An Exploratory Study of the Relationship between Academic Library Work Experience and Perceptions of Leadership Skill Development Relevant to Academic Library Directorship

    Science.gov (United States)

    Harris-Keith, Colleen Susan

    2015-01-01

    Though research into academic library director leadership has established leadership skills and qualities required for success, little research has been done to establish where in their career library directors were most likely to acquire those skills and qualities. This research project surveyed academic library directors at Carnegie-designated…

  17. Science photo library

    CERN Document Server

    1999-01-01

    SPL [Science Photo Library] holds a wide range of pictures on all aspects of science, medicine and technology. The pictures come with detailed captions and are available as high quality transparencies in medium or 35mm format. Digital files can be made available on request. Our website provides low resolution files of the pictures in this catalogue, which you can search and download for layout presentation use once you have registered. High resolution files or reproduction are available on request and can be delivered to you by disk or ISDN. Visit the online catalog: www.sciencephoto.com

  18. Lab-on-a-chip platform for high throughput drug discovery with DNA-encoded chemical libraries

    Science.gov (United States)

    Grünzner, S.; Reddavide, F. V.; Steinfelder, C.; Cui, M.; Busek, M.; Klotzbach, U.; Zhang, Y.; Sonntag, F.

    2017-02-01

    The fast development of DNA-encoded chemical libraries (DECL) in the past 10 years has received great attention from pharmaceutical industries. It applies the selection approach for small molecular drug discovery. Because of the limited choices of DNA-compatible chemical reactions, most DNA-encoded chemical libraries have a narrow structural diversity and low synthetic yield. There is also a poor correlation between the ranking of compounds resulted from analyzing the sequencing data and the affinity measured through biochemical assays. By combining DECL with dynamical chemical library, the resulting DNA-encoded dynamic library (EDCCL) explores the thermodynamic equilibrium of reversible reactions as well as the advantages of DNA encoded compounds for manipulation/detection, thus leads to enhanced signal-to-noise ratio of the selection process and higher library quality. However, the library dynamics are caused by the weak interactions between the DNA strands, which also result in relatively low affinity of the bidentate interaction, as compared to a stable DNA duplex. To take advantage of both stably assembled dual-pharmacophore libraries and EDCCLs, we extended the concept of EDCCLs to heat-induced EDCCLs (hi-EDCCLs), in which the heat-induced recombination process of stable DNA duplexes and affinity capture are carried out separately. To replace the extremely laborious and repetitive manual process, a fully automated device will facilitate the use of DECL in drug discovery. Herein we describe a novel lab-on-a-chip platform for high throughput drug discovery with hi-EDCCL. A microfluidic system with integrated actuation was designed which is able to provide a continuous sample circulation by reducing the volume to a minimum. It consists of a cooled and a heated chamber for constant circulation. The system is capable to generate stable temperatures above 75 °C in the heated chamber to melt the double strands of the DNA and less than 15 °C in the cooled chamber

  19. Amino acid substitutions in random mutagenesis libraries: lessons from analyzing 3000 mutations.

    Science.gov (United States)

    Zhao, Jing; Frauenkron-Machedjou, Victorine Josiane; Kardashliev, Tsvetan; Ruff, Anna Joëlle; Zhu, Leilei; Bocola, Marco; Schwaneberg, Ulrich

    2017-04-01

    The quality of amino acid substitution patterns in random mutagenesis libraries is decisive for the success in directed evolution campaigns. In this manuscript, we provide a detailed analysis of the amino acid substitutions by analyzing 3000 mutations of three random mutagenesis libraries (1000 mutations each; epPCR with a low-mutation and a high-mutation frequency and SeSaM-Tv P/P) employing lipase A from Bacillus subtilis (bsla). A comparison of the obtained numbers of beneficial variants in the mentioned three random mutagenesis libraries with a site saturation mutagenesis (SSM) (covering the natural diversity at each amino acid position of BSLA) concludes the diversity analysis. Seventy-six percent of the SeSaM-Tv P/P-generated substitutions yield chemically different amino acid substitutions compared to 64% (epPCR-low) and 69% (epPCR-high). Unique substitutions from one amino acid to others are termed distinct amino acid substitutions. In the SeSaM-Tv P/P library, 35% of all theoretical distinct amino acid substitutions were found in the 1000 mutation library compared to 25% (epPCR-low) and 26% (epPCR-high). Thirty-six percent of distinct amino acid substitutions found in SeSaM-Tv P/P were unobtainable by epPCR-low. Comparison with the SSM library showed that epPCR-low covers 15%, epPCR-high 18%, and SeSaM-Tv P/P 21% of obtainable beneficial amino acid positions. In essence, this study provides first insights on the quality of epPCR and SeSaM-Tv P/P libraries in terms of amino acid substitutions, their chemical differences, and the number of obtainable beneficial amino acid positions.

  20. Integrated library system in the library of Japan Atomic Energy Research Institute

    International Nuclear Information System (INIS)

    Yonezawa, Minoru; Mineo, Yukinobu; Itabashi, Keizo

    1987-01-01

    Integrated library system has been developed using a stand-alone mini-computer in the Japan Atomic Energy Research Institute library. This system consists of three subsystems for serials control, books acquisition and circulation control. Serials control subsystem deals with subscription, acquisition, claiming and inquiry of journals. This has been operating since the beginning of 1985. Book acquisition sub-system, which has been started since April 1986, deals with accounting and cataloguing of books. Circulation control sub-system deals with circulation, statistics compilation, book inventory and retrieval, which has been operating since April 1987. This system contributes greatly not only to the reduction of the circulation work load but also to the promotion of the library services. However, the convenience in circulation processing should be improved for materials without catalogue information stored in the computer. The pertinence for maximum number of books retrieved has to be also reconsidered. (author)

  1. Booth Library On-Line Circulation System (BLOC

    Directory of Open Access Journals (Sweden)

    Paladugu V. Rao

    1971-06-01

    Full Text Available An on-line circulation system developed at a relatively small university library demonstrates that academic libraries with limited funds can develop automated systems utilizing parent institution's computer facilities in a time-sharing mode. In operation since September 1968, using an IBM 360/50 computer and associated peripheral equipment, it provides control over all stack books.

  2. Design issues for numerical libraries on scalable multicore architectures

    International Nuclear Information System (INIS)

    Heroux, M A

    2008-01-01

    Future generations of scalable computers will rely on multicore nodes for a significant portion of overall system performance. At present, most applications and libraries cannot exploit multiple cores beyond running addition MPI processes per node. In this paper we discuss important multicore architecture issues, programming models, algorithms requirements and software design related to effective use of scalable multicore computers. In particular, we focus on important issues for library research and development, making recommendations for how to effectively develop libraries for future scalable computer systems

  3. Python Materials Genomics (pymatgen): A robust, open-source python library for materials analysis

    OpenAIRE

    Ong, Shyue Ping; Richards, William Davidson; Jain, Anubhav; Hautier, Geoffroy; Kocher, Michael; Cholia, Shreyas; Gunter, Dan; Chevrier, Vincent L.; Persson, Kristin A.; Ceder, Gerbrand

    2012-01-01

    We present the Python Materials Genomics (pymatgen) library, a robust, open-source Python library for materials analysis. A key enabler in high-throughput computational materials science efforts is a robust set of software tools to perform initial setup for the calculations (e.g., generation of structures and necessary input files) and post-calculation analysis to derive useful material properties from raw calculated data. The pymatgen library aims to meet these needs by (1) defining core Pyt...

  4. Interconnectivity and the Electronic Academic Library

    Directory of Open Access Journals (Sweden)

    Donald E. Riggs

    1988-03-01

    Full Text Available 無Due to the emphasis on the use of computing networks on campuses and to the very nature of more information being accessible to library users only via electronic means, we are witnessing a migration to electronic academic libraries. this new type of library is being required to have interconnections with the campus' other online information/data systems. Arizona State University libraries have been provided the opportunity to develop an electronic library that will be the focal point of a campus-wide information/data network.

  5. Myanmar Library Association Newsletter No. 2

    International Nuclear Information System (INIS)

    1995-04-01

    The news and articles related to the Myanmar Library Association (MLA) are contained in this issue in English language namely. 1. National library statistics, activities of the Sarpay Beikman public library and the list of books translated from English to Myanmar in the year 1994. Compiled by Daw Nu-Nu. 2. The Myanmar Library Association by U Tin - Maung - Lwin. 3. Librarians conference to be huge (extract from China Daily November 18, 1994). 4. Magic of computers in libraries by U Maung - Maung. 5. The two union catalogues of Myanmar by Hla-Win (MSTRD)

  6. Advanced Transport Operating System (ATOPS) utility library software description

    Science.gov (United States)

    Clinedinst, Winston C.; Slominski, Christopher J.; Dickson, Richard W.; Wolverton, David A.

    1993-01-01

    The individual software processes used in the flight computers on-board the Advanced Transport Operating System (ATOPS) aircraft have many common functional elements. A library of commonly used software modules was created for general uses among the processes. The library includes modules for mathematical computations, data formatting, system database interfacing, and condition handling. The modules available in the library and their associated calling requirements are described.

  7. Bayesian approach to peak deconvolution and library search for high resolution gas chromatography - Mass spectrometry.

    Science.gov (United States)

    Barcaru, A; Mol, H G J; Tienstra, M; Vivó-Truyols, G

    2017-08-29

    A novel probabilistic Bayesian strategy is proposed to resolve highly coeluting peaks in high-resolution GC-MS (Orbitrap) data. Opposed to a deterministic approach, we propose to solve the problem probabilistically, using a complete pipeline. First, the retention time(s) for a (probabilistic) number of compounds for each mass channel are estimated. The statistical dependency between m/z channels was implied by including penalties in the model objective function. Second, Bayesian Information Criterion (BIC) is used as Occam's razor for the probabilistic assessment of the number of components. Third, a probabilistic set of resolved spectra, and their associated retention times are estimated. Finally, a probabilistic library search is proposed, computing the spectral match with a high resolution library. More specifically, a correlative measure was used that included the uncertainties in the least square fitting, as well as the probability for different proposals for the number of compounds in the mixture. The method was tested on simulated high resolution data, as well as on a set of pesticides injected in a GC-Orbitrap with high coelution. The proposed pipeline was able to detect accurately the retention times and the spectra of the peaks. For our case, with extremely high coelution situation, 5 out of the 7 existing compounds under the selected region of interest, were correctly assessed. Finally, the comparison with the classical methods of deconvolution (i.e., MCR and AMDIS) indicates a better performance of the proposed algorithm in terms of the number of correctly resolved compounds. Copyright © 2017 Elsevier B.V. All rights reserved.

  8. Common Graphics Library (CGL). Volume 2: Low-level user's guide

    Science.gov (United States)

    Taylor, Nancy L.; Hammond, Dana P.; Theophilos, Pauline M.

    1989-01-01

    The intent is to instruct the users of the Low-Level routines of the Common Graphics Library (CGL). The Low-Level routines form an application-independent graphics package enabling the user community to construct and design scientific charts conforming to the publication and/or viewgraph process. The Low-Level routines allow the user to design unique or unusual report-quality charts from a set of graphics utilities. The features of these routines can be used stand-alone or in conjunction with other packages to enhance or augment their capabilities. This library is written in ANSI FORTRAN 77, and currently uses a CORE-based underlying graphics package, and is therefore machine-independent, providing support for centralized and/or distributed computer systems.

  9. Technostress in Libraries: Causes, Effects and Solutions.

    Science.gov (United States)

    Bichteler, Julie

    1987-01-01

    Examines some of the fears, frustrations, and misconceptions of library staff and patrons that hamper the effective use of computers in libraries. Strategies that library administrators could use to alleviate stress are outlined, including staff participation in the automation process, well-designed workstations, and adequate training for staff…

  10. Get It? Got It. Good!: Utilizing Get It Now Article Delivery Service at a Health Sciences Library

    Science.gov (United States)

    Jarvis, Christy; Gregory, Joan M.

    2016-01-01

    With journal price increases continuing to outpace inflation and library collection funds remaining stagnant or shrinking, libraries are seeking innovative ways to control spending while continuing to provide patrons with high-quality content. The Spencer S. Eccles Health Sciences Library reports on the evaluation, implementation, and use of…

  11. TRAC Searchable Research Library

    Science.gov (United States)

    2016-05-01

    Relational Data Modeling (VRDM) computational paradigm. VRDM has the key attributes of being cloud available, using domain semantics for configured...Figure 1. Methodology for TRAC Searchable Research Library Development. ........................... 5 Figure 2. The conceptual model for the cloud ...TRAC Searchable Research Library project was initiated by TRAC- HQ to address a current capability gap in the TRAC organization. Currently TRAC does not

  12. Y2K Resources for Public Libraries.

    Science.gov (United States)

    Foster, Janet

    1999-01-01

    Presents information for public libraries on computer-related vulnerabilities as the century turns from 1999 to 2000. Highlights include: general Y2K information; the Y2K Bug and PCs; Y2K sites for librarians; Online Computer Library Center (OCLC) and USMARC; technological developments in cyberspace; and a list of Web sites and Y2K resources. (AEF)

  13. Expert Systems for Libraries at SCIL [Small Computers in Libraries]'88.

    Science.gov (United States)

    Kochtanek, Thomas R.; And Others

    1988-01-01

    Six brief papers on expert systems for libraries cover (1) a knowledge-based approach to database design; (2) getting started in expert systems; (3) using public domain software to develop a business reference system; (4) a music cataloging inquiry system; (5) linguistic analysis of reference transactions; and (6) a model of a reference librarian.…

  14. MARKETING COMMUNICATIONS IN THE LIBRARY OF KRYVYI RIH STATE PEDAGOGICAL UNIVERSITY

    Directory of Open Access Journals (Sweden)

    Г. М. Віняр

    2017-10-01

    Full Text Available The subject of research is the examination of marketing techniques aimed at creating a favorable image of the university library, engaging its users in high use of information resources, and modernization of library processes. Purpose of the article is to analyze the feasibility and efficiency of different types of marketing communications in the library ofKryvyiRihStatePedagogicalUniversity. The article is based on modern theoretical developments in the area of library marketing, without which successful work of an information institution is impossible. In the investigation are used statistical data from annual reports of the library, quantitative data on site traffic and users’ attendance of cultural and educational events, information from survey among students and teachers. The article describes forms of marketing interaction between the library institution, its users, and the public. University libraries, including the library websites, are the main application area of developments in information marketing communications. Promising directions of activity in the Kryvyi Rih State Pedagogical University library remain the following: further improvement of library and information services with maximum involvement of new information technologies and innovative methodologies, development of library-based powerful, extensive information system designed to organize the most complete, high-quality and quick support for all areas of the university work based on current information needs of all categories of library users.

  15. Robotic tape library system level testing at NSA: Present and planned

    Science.gov (United States)

    Shields, Michael F.

    1994-01-01

    In the present of declining Defense budgets, increased pressure has been placed on the DOD to utilize Commercial Off the Shelf (COTS) solutions to incrementally solve a wide variety of our computer processing requirements. With the rapid growth in processing power, significant expansion of high performance networking, and the increased complexity of applications data sets, the requirement for high performance, large capacity, reliable and secure, and most of all affordable robotic tape storage libraries has greatly increased. Additionally, the migration to a heterogeneous, distributed computing environment has further complicated the problem. With today's open system compute servers approaching yesterday's supercomputer capabilities, the need for affordable, reliable secure Mass Storage Systems (MSS) has taken on an ever increasing importance to our processing center's ability to satisfy operational mission requirements. To that end, NSA has established an in-house capability to acquire, test, and evaluate COTS products. Its goal is to qualify a set of COTS MSS libraries, thereby achieving a modicum of standardization for robotic tape libraries which can satisfy our low, medium, and high performance file and volume serving requirements. In addition, NSA has established relations with other Government Agencies to complete this in-house effort and to maximize our research, testing, and evaluation work. While the preponderance of the effort is focused at the high end of the storage ladder, considerable effort will be extended this year and next at the server class or mid range storage systems.

  16. 圖書館事業與交流/What Cloud Computing Means to Libraries and Information Services/Xiaocan (Lucy Wang, Jie Huang

    Directory of Open Access Journals (Sweden)

    Xiaocan (Lucy Wang, Jie Huang Xiaocan (Lucy Wang, Jie Huang

    2011-10-01

    Full Text Available Cloud computing, as an emerging style of computing that involves various technologies and provides virtual and scalable IT-related services over the Internet, has become a popular topic in the past few years. Much has been written about this concept and its applications in the IT and business fields, but little has been done in the library setting. This paper first introduces cloud computing with respect to its concepts, types, and general applications. The focus is then shifted to what cloud computing might mean to libraries and information services. The impact of cloud computing on the storage of libraries’ digital assets, integrated library systems, libraries’ IT infrastructures, as well as library services and internal work, has been investigated. The paper also explores the potential benefits of incorporating cloud computing in the library environment and points out some concerns and challenges libraries may have. 雲端計算作為一種新興的、利用多元技術來提供基於網路的IT 服務計算模式,在近幾年來已成為熱門話題。許多文章論述了有關雲端計算的概念及其在IT 和商業中的應用模式,但卻少見對其在圖書館領域中運用的闡述。本文首先介紹有關雲端計算的概念、類型,以及一些基本的應用模式。隨後,本文著眼於雲端計算對圖書館和資訊服務的意義,探討雲端計算在圖書館數位資源儲存、整合圖書館系統、圖書館資訊科技的設施建設、圖書館服務和內部作業等方面產生的影響以及圖書館運用雲端計算的潛在效益。本文剖析圖書館採用雲端計算的益處之時,也同時探討可能面臨的一些問題和挑戰。 頁次:166-174

  17. The Principals and Practice of Distributed High Throughput Computing

    CERN Multimedia

    CERN. Geneva

    2016-01-01

    The potential of Distributed Processing Systems to deliver computing capabilities with qualities ranging from high availability and reliability to easy expansion in functionality and capacity were recognized and formalized in the 1970’s. For more three decade these principals Distributed Computing guided the development of the HTCondor resource and job management system. The widely adopted suite of software tools offered by HTCondor are based on novel distributed computing technologies and are driven by the evolving needs of High Throughput scientific applications. We will review the principals that underpin our work, the distributed computing frameworks and technologies we developed and the lessons we learned from delivering effective and dependable software tools in an ever changing landscape computing technologies and needs that range today from a desktop computer to tens of thousands of cores offered by commercial clouds. About the speaker Miron Livny received a B.Sc. degree in Physics and Mat...

  18. On an efficient general mathematical library

    International Nuclear Information System (INIS)

    Li Xiaomei; Yan Baoyong

    1992-01-01

    In this paper, the architecture of vector pipeline computer YH-1 is briefly described, and the characteristics of an efficient vector general mathematical library are discussed. Some approaches to vectorization, adopted in developing the vector mathematical library, are presented

  19. High performance computing and communications: Advancing the frontiers of information technology

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-12-31

    This report, which supplements the President`s Fiscal Year 1997 Budget, describes the interagency High Performance Computing and Communications (HPCC) Program. The HPCC Program will celebrate its fifth anniversary in October 1996 with an impressive array of accomplishments to its credit. Over its five-year history, the HPCC Program has focused on developing high performance computing and communications technologies that can be applied to computation-intensive applications. Major highlights for FY 1996: (1) High performance computing systems enable practical solutions to complex problems with accuracies not possible five years ago; (2) HPCC-funded research in very large scale networking techniques has been instrumental in the evolution of the Internet, which continues exponential growth in size, speed, and availability of information; (3) The combination of hardware capability measured in gigaflop/s, networking technology measured in gigabit/s, and new computational science techniques for modeling phenomena has demonstrated that very large scale accurate scientific calculations can be executed across heterogeneous parallel processing systems located thousands of miles apart; (4) Federal investments in HPCC software R and D support researchers who pioneered the development of parallel languages and compilers, high performance mathematical, engineering, and scientific libraries, and software tools--technologies that allow scientists to use powerful parallel systems to focus on Federal agency mission applications; and (5) HPCC support for virtual environments has enabled the development of immersive technologies, where researchers can explore and manipulate multi-dimensional scientific and engineering problems. Educational programs fostered by the HPCC Program have brought into classrooms new science and engineering curricula designed to teach computational science. This document contains a small sample of the significant HPCC Program accomplishments in FY 1996.

  20. Leadership In The Public Library : Case Study In Tenteram City Public Library

    Directory of Open Access Journals (Sweden)

    Muhammad Irsyad Al Fatih

    2018-01-01

    Full Text Available In the formal and informal organization, it is always found someone who is considered more prominent than others, who then appointed or designated as a leader. The Leader of an organization is expected to have leadership qualities, which showed his skills in leading organizations. A leader in the library requires synergy between skills in library management and leadership in general in order to create a good working atmosphere and to encourage the library staff in providing the best services to the patron community. Tenteram City Public library since year of 2018 has never experienced a change of its leader. Research using qualitative approach aimed to identify the leadership of Tenteram City public library leader in developing human resources and library services. Data of this study is gathered through observation and interviews. It is found that the head of the library serves as a role model in imbedding working discipline, he never involved the staff in decision making, the fear of making mistake is the working atmosphere created in the library. The Research found that the head of Tenteram City public library adopts autocratic leadership.

  1. Leadership In The Public Library : Case Study In Tenteram City Public Library

    Directory of Open Access Journals (Sweden)

    Muhammad Irsyad Alfatih

    2017-06-01

    In the formal and informal organization, it is always found someone who is considered more prominent than others, who then appointed or designated as a leader. The Leader of an organization is expected to have leadership qualities, which showed his skills in leading organizations. A leader in the library requires synergy between skills in library management and leadership in general in order to create a good working atmosphere and to encourage the library staff in providing the best services to the patron community. Tenteram City Public library since year of 2018 has never experienced a change of its leader. Research using qualitative approach aimed to identify the leadership of Tenteram City public library leader in developing human resources and library services. Data of this study is gathered through observation and interviews. It is found that the head of the library serves as a role model in imbedding working discipline, he never involved the staff in decision making, the fear of making mistake is the working atmosphere created in the library. The Research found that the head of Tenteram City public library adopts autocratic leadership.

  2. Possibilities for Proactive Library Services.

    Science.gov (United States)

    Morgan, Eric Lease

    1999-01-01

    Considers ways in which library services can be more proactive in today's networked-computer environment. Discusses how to find and use patterns of behavior, such as borrowing behavior and profiles of patrons' interests; making a CD-ROM with information describing the library's services and products; and reviving telephone reference. (LRW)

  3. Assessing Tax Form Distribution Costs: A Proposed Method for Computing the Dollar Value of Tax Form Distribution in a Public Library.

    Science.gov (United States)

    Casey, James B.

    1998-01-01

    Explains how a public library can compute the actual cost of distributing tax forms to the public by listing all direct and indirect costs and demonstrating the formulae and necessary computations. Supplies directions for calculating costs involved for all levels of staff as well as associated public relations efforts, space, and utility costs.…

  4. Development and benchmark of high energy continuous-energy neutron cross Section library HENDL-ADS/MC

    International Nuclear Information System (INIS)

    Chen Chong; Wang Minghuang; Zou Jun; Xu Dezheng; Zeng Qin

    2012-01-01

    The ADS (accelerator driven sub-critical system) has great energy spans, complex energy spectrum structures and strong physical effects. Hence, the existing nuclear data libraries can't fully meet the needs of nuclear analysis in ADS. In order to do nuclear analysis for ADS system, a point-wise data library HENDL-ADS/MC (hybrid evaluated nuclear data library) was produced by FDS team. Meanwhile, to test the availability and reliability of the HENDL-ADS/MC data library, a series of shielding and critical safety benchmarks were performed. To validate and qualify the reliability of the high-energy cross section for HENDL-ADS/MC library further, a series of high neutronics integral experiments have been performed. The testing results confirm the accuracy and reliability of HENDL-ADS/MC. (authors)

  5. Library Website Usability Test Project

    KAUST Repository

    Ramli, Rindra M.; Bukhari, Duaa

    2013-01-01

    This usability testing project was conducted to elicit an understanding of our community use of the library website. The researchers wanted to know how our users are interacting with the library website and the ease of obtaining relevant information from the website. The methodology deployed was computer user testing where participants are made to answer several questions and executing the actions on the library website. Their actions are recorded via Techsmith Camtasia software for later analysis by the researchers.

  6. Library Website Usability Test Project

    KAUST Repository

    Ramli, Rindra M.

    2013-06-01

    This usability testing project was conducted to elicit an understanding of our community use of the library website. The researchers wanted to know how our users are interacting with the library website and the ease of obtaining relevant information from the website. The methodology deployed was computer user testing where participants are made to answer several questions and executing the actions on the library website. Their actions are recorded via Techsmith Camtasia software for later analysis by the researchers.

  7. Recent advances in PC-Linux systems for electronic structure computations by optimized compilers and numerical libraries.

    Science.gov (United States)

    Yu, Jen-Shiang K; Yu, Chin-Hui

    2002-01-01

    One of the most frequently used packages for electronic structure research, GAUSSIAN 98, is compiled on Linux systems with various hardware configurations, including AMD Athlon (with the "Thunderbird" core), AthlonMP, and AthlonXP (with the "Palomino" core) systems as well as the Intel Pentium 4 (with the "Willamette" core) machines. The default PGI FORTRAN compiler (pgf77) and the Intel FORTRAN compiler (ifc) are respectively employed with different architectural optimization options to compile GAUSSIAN 98 and test the performance improvement. In addition to the BLAS library included in revision A.11 of this package, the Automatically Tuned Linear Algebra Software (ATLAS) library is linked against the binary executables to improve the performance. Various Hartree-Fock, density-functional theories, and the MP2 calculations are done for benchmarking purposes. It is found that the combination of ifc with ATLAS library gives the best performance for GAUSSIAN 98 on all of these PC-Linux computers, including AMD and Intel CPUs. Even on AMD systems, the Intel FORTRAN compiler invariably produces binaries with better performance than pgf77. The enhancement provided by the ATLAS library is more significant for post-Hartree-Fock calculations. The performance on one single CPU is potentially as good as that on an Alpha 21264A workstation or an SGI supercomputer. The floating-point marks by SpecFP2000 have similar trends to the results of GAUSSIAN 98 package.

  8. A theoretical model to predict customer satisfaction in relation to service quality in selected university libraries in Sri Lanka

    Directory of Open Access Journals (Sweden)

    Chaminda Jayasundara

    2009-01-01

    Full Text Available University library administrators in Sri Lanka have begun to search for alternative ways to satisfy their clientele on the basis of service quality. This article aims at providing a theoretical model to facilitate the identification of service quality attributes and domains that may be used to predict customer satisfaction from a service quality perspective. The effectiveness of existing service quality models such as LibQUAL, SERVQUAL and SERVPREF have been questioned. In that regard, this study developed a theoretical model for academic libraries in Sri Lanka based on the disconfirmation and performance-only paradigms. These perspectives were considered by researchers to be the core mechanism to develop service quality/customer satisfaction models. The attributes and domain identification of service quality was carried out with a stratified sample of 263 participants selected from postgraduate and undergraduate students and academic staff members from the faculties of Arts in four universities in Sri Lanka. The study established that responsiveness, supportiveness, building environment, collection and access, furniture and facilities, technology, Web services and service delivery were quality domains which can be used to predict customer satisfaction. The theoretical model is unique in its domain structure compared to the existing models. The model needs to be statistically tested to make it valid and parsimonious.

  9. Library Automation.

    Science.gov (United States)

    Husby, Ole

    1990-01-01

    The challenges and potential benefits of automating university libraries are reviewed, with special attention given to cooperative systems. Aspects discussed include database size, the role of the university computer center, storage modes, multi-institutional systems, resource sharing, cooperative system management, networking, and intelligent…

  10. Manual on JSSL (JAERI scientific subroutine library)

    International Nuclear Information System (INIS)

    Fujimura, Toichiro; Nishida, Takahiko; Asai, Kiyoshi

    1977-05-01

    A manual on the revised JAERI scientific subroutine library is presented. The library is a collection of subroutines developed or modified in JAERI which complements the library installed for FACOM 230-75 computer. It is subject to further extension in the future, since the present one is still insufficient for scientific calculations. (auth.)

  11. libRoadRunner: a high performance SBML simulation and analysis library.

    Science.gov (United States)

    Somogyi, Endre T; Bouteiller, Jean-Marie; Glazier, James A; König, Matthias; Medley, J Kyle; Swat, Maciej H; Sauro, Herbert M

    2015-10-15

    This article presents libRoadRunner, an extensible, high-performance, cross-platform, open-source software library for the simulation and analysis of models expressed using Systems Biology Markup Language (SBML). SBML is the most widely used standard for representing dynamic networks, especially biochemical networks. libRoadRunner is fast enough to support large-scale problems such as tissue models, studies that require large numbers of repeated runs and interactive simulations. libRoadRunner is a self-contained library, able to run both as a component inside other tools via its C++ and C bindings, and interactively through its Python interface. Its Python Application Programming Interface (API) is similar to the APIs of MATLAB ( WWWMATHWORKSCOM: ) and SciPy ( HTTP//WWWSCIPYORG/: ), making it fast and easy to learn. libRoadRunner uses a custom Just-In-Time (JIT) compiler built on the widely used LLVM JIT compiler framework. It compiles SBML-specified models directly into native machine code for a variety of processors, making it appropriate for solving extremely large models or repeated runs. libRoadRunner is flexible, supporting the bulk of the SBML specification (except for delay and non-linear algebraic equations) including several SBML extensions (composition and distributions). It offers multiple deterministic and stochastic integrators, as well as tools for steady-state analysis, stability analysis and structural analysis of the stoichiometric matrix. libRoadRunner binary distributions are available for Mac OS X, Linux and Windows. The library is licensed under Apache License Version 2.0. libRoadRunner is also available for ARM-based computers such as the Raspberry Pi. http://www.libroadrunner.org provides online documentation, full build instructions, binaries and a git source repository. hsauro@u.washington.edu or somogyie@indiana.edu Supplementary data are available at Bioinformatics online. Published by Oxford University Press 2015. This work is written

  12. Construction of bacterial artificial chromosome libraries for Zhikong Scallop Chlamys farreri

    Institute of Scientific and Technical Information of China (English)

    ZHANG Yang; ZHANG Xiaojun; Chantel F.SCHEURING; ZHANG Hongbin; LI Fuhua; XIANG Jianhai

    2008-01-01

    Two Large-insert genomic bacterial artificial chromosome (BAC) libraries of Zhikong scallop Chlamys farreri were constructed to promote our genetic and genomic research.High-quality megabase-sized DNA was isolated from the adductor muscle of the scallop and partially digested by BamH I and Mbo I,respectively.The BamH I library consisted of 53760 clones while the Mbo I library consisted of 7680 clones.Approximately 96% of the clones in BamH I library contained nuclear DNA inserts in average size of 100 kb,providing a coverage of 5.3 haploid genome equivalents.Similarly,the Mbo I library with an average insert of 145 kb and no insert-empty clones,thus providing a genome coverage of 1.1 haploid genome equivalents.

  13. Law Firms and Public Libraries.

    Science.gov (United States)

    Fong, Colin

    1997-01-01

    Summarizes sources of legal information in New South Wales and outlines the libraries' driving motivations, access policies, and practices. Also highlights the important value-added service professional library staff contribute to both the quality and quantity of information delivery in this discipline. (Author/AEF)

  14. Methods for computing water-quality loads at sites in the U.S. Geological Survey National Water Quality Network

    Science.gov (United States)

    Lee, Casey J.; Murphy, Jennifer C.; Crawford, Charles G.; Deacon, Jeffrey R.

    2017-10-24

    The U.S. Geological Survey publishes information on concentrations and loads of water-quality constituents at 111 sites across the United States as part of the U.S. Geological Survey National Water Quality Network (NWQN). This report details historical and updated methods for computing water-quality loads at NWQN sites. The primary updates to historical load estimation methods include (1) an adaptation to methods for computing loads to the Gulf of Mexico; (2) the inclusion of loads computed using the Weighted Regressions on Time, Discharge, and Season (WRTDS) method; and (3) the inclusion of loads computed using continuous water-quality data. Loads computed using WRTDS and continuous water-quality data are provided along with those computed using historical methods. Various aspects of method updates are evaluated in this report to help users of water-quality loading data determine which estimation methods best suit their particular application.

  15. Construction of human antibody gene libraries and selection of antibodies by phage display.

    Science.gov (United States)

    Frenzel, André; Kügler, Jonas; Wilke, Sonja; Schirrmann, Thomas; Hust, Michael

    2014-01-01

    Antibody phage display is the most commonly used in vitro selection technology and has yielded thousands of useful antibodies for research, diagnostics, and therapy.The prerequisite for successful generation and development of human recombinant antibodies using phage display is the construction of a high-quality antibody gene library. Here, we describe the methods for the construction of human immune and naive scFv gene libraries.The success also depends on the panning strategy for the selection of binders from these libraries. In this article, we describe a panning strategy that is high-throughput compatible and allows parallel selection in microtiter plates.

  16. Library designs for generic C++ sparse matrix computations of iterative methods

    Energy Technology Data Exchange (ETDEWEB)

    Pozo, R.

    1996-12-31

    A new library design is presented for generic sparse matrix C++ objects for use in iterative algorithms and preconditioners. This design extends previous work on C++ numerical libraries by providing a framework in which efficient algorithms can be written *independent* of the matrix layout or format. That is, rather than supporting different codes for each (element type) / (matrix format) combination, only one version of the algorithm need be maintained. This not only reduces the effort for library developers, but also simplifies the calling interface seen by library users. Furthermore, the underlying matrix library can be naturally extended to support user-defined objects, such as hierarchical block-structured matrices, or application-specific preconditioners. Utilizing optimized kernels whenever possible, the resulting performance of such framework can be shown to be competitive with optimized Fortran programs.

  17. Digital Libraries--Methods and Applications

    Science.gov (United States)

    Huang, Kuo Hung, Ed.

    2011-01-01

    Digital library is commonly seen as a type of information retrieval system which stores and accesses digital content remotely via computer networks. However, the vision of digital libraries is not limited to technology or management, but user experience. This book is an attempt to share the practical experiences of solutions to the operation of…

  18. MCNP4c JEFF-3.1 Based Libraries. Eccolib-Jeff-3.1 libraries

    International Nuclear Information System (INIS)

    Sublet, J.Ch.

    2006-01-01

    Continuous-energy and multi-temperatures MCNP Ace types libraries, derived from the Joint European Fusion-Fission JEFF-3.1 evaluations, have been generated using the NJOY-99.111 processing code system. They include the continuous-energy neutron JEFF-3.1/General Purpose, JEFF-3.1/Activation-Dosimetry and thermal S(α,β) JEFF-3.1/Thermal libraries and data tables. The processing steps and features are explained together with the Quality Assurance processes and records linked to the generation of such multipurpose libraries. (author)

  19. Initial draft of CSE-UCLA evaluation model based on weighted product in order to optimize digital library services in computer college in Bali

    Science.gov (United States)

    Divayana, D. G. H.; Adiarta, A.; Abadi, I. B. G. S.

    2018-01-01

    The aim of this research was to create initial design of CSE-UCLA evaluation model modified with Weighted Product in evaluating digital library service at Computer College in Bali. The method used in this research was developmental research method and developed by Borg and Gall model design. The results obtained from the research that conducted earlier this month was a rough sketch of Weighted Product based CSE-UCLA evaluation model that the design had been able to provide a general overview of the stages of weighted product based CSE-UCLA evaluation model used in order to optimize the digital library services at the Computer Colleges in Bali.

  20. Library-Based Learning in an Information Society.

    Science.gov (United States)

    Breivik, Patricia Senn

    1986-01-01

    The average academic library has great potential for quality nonclassroom learning benefiting students, faculty, alumni, and the local business community. The major detriments are the limited perceptions about libraries and librarians among campus administrators and faculty. Library-based learning should be planned to be assimilated into overall…

  1. An Analysis of User Satisfaction of K University’s Library Service

    Directory of Open Access Journals (Sweden)

    Younghee Noh

    2011-06-01

    Full Text Available This study purposed to discover whether or not academic libraries reflect these changing roles. We selected K University as the research target and surveyed user satisfaction of materials, staff services, facilities, electronic devices, media, and so on. The research findings are as follows: 1 the frequency of library visits of University K was on the high side, 2 the primary purpose of using the academic library was associated with learning or reading, therefore, the most used library spaces were related to that, 3 the most used library materials were 'general books', the most unused were 'reference books', 4 the most preferred way to obtain needed materials when failing to find wanted materials was 'Contact librarian'. A similar phenomenon occurred in terms of facility use, 5 university K's users were usually satisfied with the loan policy, 6 the rate of users who don't know whether there is user education was very high, the rate of users who have no experience with user education was extremely low. These research findings can be referenced by library management to improve libraries' service quality and take advantage of complex spatial configurations.

  2. New data libraries and physics data management tools

    International Nuclear Information System (INIS)

    Han, M; Kim, C H; Seo, H; Pia, M G; Saracco, P; Augelli, M; Hauf, S; Kuster, M; Moneta, L; Quintieri, L

    2011-01-01

    A number of physics data libraries for Monte Carlo simulation are reviewed. The development of a package for the management of physics data is described: its design, implementation and computational benchmarks. This package improves the data management tools originally developed for Geant4 electromagnetic physics models based on data libraries. The implementation exploits recent evolutions of the C++ libraries appearing in the C++0x draft, which are intended for inclusion in the next C++ ISO Standard. The new tools improve the computational performance of physics data management.

  3. Designing Syntax Embeddings and Assimilations for Language Libraries

    NARCIS (Netherlands)

    Bravenboer, M.; Visser, E.

    2009-01-01

    Preprint of paper published in: Models in Software Engineering, Lecture Notes in Computer Science 5002, 2008; doi:10.1007/978-3-540-69073-3_5 Language libraries extend regular libraries with domain-specific notation. More precisely, a language library is a combination of a domain-specific language

  4. Recent developments in the Sesame equation-of-state library

    International Nuclear Information System (INIS)

    Bennett, B.I.; Johnson, J.D.; Kerley, G.I.; Rood, G.T.

    1978-02-01

    The Sesame Library is a computer file of tabular equations of state and other material properties, developed in Group T-4 of the Los Alamos Scientific Laboratory. This report describes some of the theoretical models used in the library, the structure of the data storage and acquisition system, and the computer routines that were developed to access and use the tables in practical applications. It also discusses how non-LASL users can obtain the Sesame data and subroutine libraries

  5. Beating Bias in the Directed Evolution of Proteins: Combining High-Fidelity on-Chip Solid-Phase Gene Synthesis with Efficient Gene Assembly for Combinatorial Library Construction.

    Science.gov (United States)

    Li, Aitao; Acevedo-Rocha, Carlos G; Sun, Zhoutong; Cox, Tony; Xu, Jia Lucy; Reetz, Manfred T

    2018-02-02

    Saturation mutagenesis (SM) constitutes a widely used technique in the directed evolution of selective enzymes as catalysts in organic chemistry and in the manipulation of metabolic paths and genomes, but the quality of the libraries is far from optimal due to the inherent amino acid bias. Herein, it is shown how this fundamental problem can be solved by applying high-fidelity solid-phase chemical gene synthesis on silicon chips followed by efficient gene assembly. Limonene epoxide hydrolase was chosen as the catalyst in the model desymmetrization of cyclohexene oxide with the stereoselective formation of (R,R)- and (S,S)-cyclohexane-1,2-diol. A traditional combinatorial PCR-based SM library, produced by simultaneous randomization at several residues by using a reduced amino acid alphabet, and the respective synthetic library were constructed and compared. Statistical analysis at the DNA level with massive sequencing demonstrates that, in the synthetic approach, 97 % of the theoretically possible DNA mutants are formed, whereas the traditional SM library contained only about 50 %. Screening at the protein level also showed the superiority of the synthetic library; many highly (R,R)- and (S,S)-selective variants being discovered are not found in the traditional SM library. With the prices of synthetic genes decreasing, this approach may point the way to future directed evolution. © 2018 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. Mixture-based combinatorial libraries from small individual peptide libraries: a case study on α1-antitrypsin deficiency.

    Science.gov (United States)

    Chang, Yi-Pin; Chu, Yen-Ho

    2014-05-16

    The design, synthesis and screening of diversity-oriented peptide libraries using a "libraries from libraries" strategy for the development of inhibitors of α1-antitrypsin deficiency are described. The major buttress of the biochemical approach presented here is the use of well-established solid-phase split-and-mix method for the generation of mixture-based libraries. The combinatorial technique iterative deconvolution was employed for library screening. While molecular diversity is the general consideration of combinatorial libraries, exquisite design through systematic screening of small individual libraries is a prerequisite for effective library screening and can avoid potential problems in some cases. This review will also illustrate how large peptide libraries were designed, as well as how a conformation-sensitive assay was developed based on the mechanism of the conformational disease. Finally, the combinatorially selected peptide inhibitor capable of blocking abnormal protein aggregation will be characterized by biophysical, cellular and computational methods.

  7. Touring the Campus Library from the World Wide Web.

    Science.gov (United States)

    Mosley, Pixey Anne; Xiao, Daniel

    1996-01-01

    The philosophy, design, implementation and evaluation of a World Wide Web-accessible Virtual Library Tour of Texas A & M University's Evans Library is presented. Its design combined technical computer issues and library instruction expertise. The tour can be used to simulate a typical walking tour through the library or heading directly to a…

  8. Fast parallel tandem mass spectral library searching using GPU hardware acceleration.

    Science.gov (United States)

    Baumgardner, Lydia Ashleigh; Shanmugam, Avinash Kumar; Lam, Henry; Eng, Jimmy K; Martin, Daniel B

    2011-06-03

    Mass spectrometry-based proteomics is a maturing discipline of biologic research that is experiencing substantial growth. Instrumentation has steadily improved over time with the advent of faster and more sensitive instruments collecting ever larger data files. Consequently, the computational process of matching a peptide fragmentation pattern to its sequence, traditionally accomplished by sequence database searching and more recently also by spectral library searching, has become a bottleneck in many mass spectrometry experiments. In both of these methods, the main rate-limiting step is the comparison of an acquired spectrum with all potential matches from a spectral library or sequence database. This is a highly parallelizable process because the core computational element can be represented as a simple but arithmetically intense multiplication of two vectors. In this paper, we present a proof of concept project taking advantage of the massively parallel computing available on graphics processing units (GPUs) to distribute and accelerate the process of spectral assignment using spectral library searching. This program, which we have named FastPaSS (for Fast Parallelized Spectral Searching), is implemented in CUDA (Compute Unified Device Architecture) from NVIDIA, which allows direct access to the processors in an NVIDIA GPU. Our efforts demonstrate the feasibility of GPU computing for spectral assignment, through implementation of the validated spectral searching algorithm SpectraST in the CUDA environment.

  9. Total Quality Management in Academic Libraries: Initial Implementation Efforts. Proceedings from the International Conference on TQM and Academic Libraries (1st, Washington, D.C., April 20-22, 1994).

    Science.gov (United States)

    Rounds, Laura, Ed.; Matthews, Michael, Ed.

    This document contains conference papers, other speeches, and supplementary material from the first International Conference on Total Quality Management (TQM) and Academic Libraries, held in 1994. The conference was comprised of four sessions, and the introductory remarks of each are included, along with transcriptions of each session's…

  10. Computational hybrid anthropometric paediatric phantom library for internal radiation dosimetry

    Science.gov (United States)

    Xie, Tianwu; Kuster, Niels; Zaidi, Habib

    2017-04-01

    Hybrid computational phantoms combine voxel-based and simplified equation-based modelling approaches to provide unique advantages and more realism for the construction of anthropomorphic models. In this work, a methodology and C++ code are developed to generate hybrid computational phantoms covering statistical distributions of body morphometry in the paediatric population. The paediatric phantoms of the Virtual Population Series (IT’IS Foundation, Switzerland) were modified to match target anthropometric parameters, including body mass, body length, standing height and sitting height/stature ratio, determined from reference databases of the National Centre for Health Statistics and the National Health and Nutrition Examination Survey. The phantoms were selected as representative anchor phantoms for the newborn, 1, 2, 5, 10 and 15 years-old children, and were subsequently remodelled to create 1100 female and male phantoms with 10th, 25th, 50th, 75th and 90th body morphometries. Evaluation was performed qualitatively using 3D visualization and quantitatively by analysing internal organ masses. Overall, the newly generated phantoms appear very reasonable and representative of the main characteristics of the paediatric population at various ages and for different genders, body sizes and sitting stature ratios. The mass of internal organs increases with height and body mass. The comparison of organ masses of the heart, kidney, liver, lung and spleen with published autopsy and ICRP reference data for children demonstrated that they follow the same trend when correlated with age. The constructed hybrid computational phantom library opens up the prospect of comprehensive radiation dosimetry calculations and risk assessment for the paediatric population of different age groups and diverse anthropometric parameters.

  11. Virtual in-service training from the librarians' point of view in libraries of medical sciences universities in Tehran

    Science.gov (United States)

    Mohaghegh, Niloofar; Raiesi Dehkordi, Puran; Alibeik, MohammadReza; Ghashghaee, Ahmad; Janbozorgi, Mojgan

    2016-01-01

    Background: In-service training courses are one of the most available programs that are used to improve the quantity and quality level of the staff services in various organizations, including libraries and information centers. With the advent of new technologies in the field of education, the problems and shortcomings of traditional in-service training courses were replaced with virtual ones. This study aimed to evaluate the virtual in-service training courses from the librarians' point of view in libraries of state universities of medical sciences in Tehran. Methods: This was a descriptive- analytical study. The statistical population consisted of all librarians at libraries of universities of medical sciences in Tehran. Out of 103 librarians working in the libraries under the study, 93 (90%) participated in this study. Data were collected, using a questionnaire. Results: The results revealed that 94/6% of librarians were satisfied to participate in virtual in-service training courses. In this study, only 45 out of 93 participants said that the virtual in-service courses were held in their libraries. Of the participants, 75.6% were satisfied with the length of training courses, and one month seemed to be adequate time duration for the librarians to be more satisfied. The satisfaction level of the individuals who participated in in-service courses of the National Library was moderate to high. A total of 84.4% participants announced that the productivity level of the training courses was moderate to high. The most important problem with which the librarians were confronted in virtual in-service training was the "low speed of the internet and inadequate computer substructures". Conclusion: Effectiveness of in-service training courses from librarians’ point of view was at an optimal level in the studied libraries. PMID:28491833

  12. Virtual in-service training from the librarians' point of view in libraries of medical sciences universities in Tehran.

    Science.gov (United States)

    Mohaghegh, Niloofar; Raiesi Dehkordi, Puran; Alibeik, MohammadReza; Ghashghaee, Ahmad; Janbozorgi, Mojgan

    2016-01-01

    Background: In-service training courses are one of the most available programs that are used to improve the quantity and quality level of the staff services in various organizations, including libraries and information centers. With the advent of new technologies in the field of education, the problems and shortcomings of traditional in-service training courses were replaced with virtual ones. This study aimed to evaluate the virtual in-service training courses from the librarians' point of view in libraries of state universities of medical sciences in Tehran. Methods: This was a descriptive- analytical study. The statistical population consisted of all librarians at libraries of universities of medical sciences in Tehran. Out of 103 librarians working in the libraries under the study, 93 (90%) participated in this study. Data were collected, using a questionnaire. Results: The results revealed that 94/6% of librarians were satisfied to participate in virtual in-service training courses. In this study, only 45 out of 93 participants said that the virtual in-service courses were held in their libraries. Of the participants, 75.6% were satisfied with the length of training courses, and one month seemed to be adequate time duration for the librarians to be more satisfied. The satisfaction level of the individuals who participated in in-service courses of the National Library was moderate to high. A total of 84.4% participants announced that the productivity level of the training courses was moderate to high. The most important problem with which the librarians were confronted in virtual in-service training was the "low speed of the internet and inadequate computer substructures". Conclusion: Effectiveness of in-service training courses from librarians' point of view was at an optimal level in the studied libraries.

  13. Outsourcing in libraries

    Directory of Open Access Journals (Sweden)

    Matjaž Žaucer

    1999-01-01

    Full Text Available Like other organisations more flexible libraries tend to conform to the changing environment as this is the only way to be successful and effective. They are expected to offer "more for less" and they are reorganising and searching the ways to reduce the costs. Outsourcing is one of possible solutions. The article deals with the possibilities of outsourcing in libraries, higher quality of their work eoneentrated on principal activities and gives some experienees in this field.

  14. Paper Cuts Don't Hurt at the Gerstein Library

    Science.gov (United States)

    Cunningham, Heather; Feder, Elah; Muise, Isaac

    2010-01-01

    The Gerstein Science Information Centre (Gerstein Library) is one of 40 libraries within the University of Toronto (U of T) and is the largest academic science and health science library in Canada. It offers 109 computers and two networked printers for student, staff, and faculty use. In addition, the library provides patrons' laptops with…

  15. An Integrated Library System: Preliminary Considerations.

    Science.gov (United States)

    Neroda, Edward

    Noting difficulties experienced by small to medium sized colleges in acquiring integrated library computer systems, this position paper outlines issues related to the subject with the intention of increasing familiarity and interest in integrated library systems. The report includes: a brief review of technological advances as they relate to…

  16. Use of library statistics to support library and advisory services and ...

    African Journals Online (AJOL)

    Statistical information is a vital tool for management and development of organizations. Keeping statistics of activities is basic to the survival and progress of a library and enables the library to measure its performance periodically. The National Library of Nigeria (NLN) places high premium on the library statistics that it ...

  17. Evaluating Library Staff: A Performance Appraisal System.

    Science.gov (United States)

    Belcastro, Patricia

    This manual provides librarians and library managers with a performance appraisal system that measures staff fairly and objectively and links performance to the goals of the library. The following topics are addressed: (1) identifying expectations for quality service or standards of performance; (2) the importance of a library's code of service,…

  18. High Performance Computing Facility Operational Assessment 2015: Oak Ridge Leadership Computing Facility

    Energy Technology Data Exchange (ETDEWEB)

    Barker, Ashley D. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Bernholdt, David E. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Bland, Arthur S. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Gary, Jeff D. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Hack, James J. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; McNally, Stephen T. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Rogers, James H. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Smith, Brian E. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Straatsma, T. P. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Sukumar, Sreenivas Rangan [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Thach, Kevin G. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Tichenor, Suzy [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Vazhkudai, Sudharshan S. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility; Wells, Jack C. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility

    2016-03-01

    California to date. The Titan system provides the largest extant heterogeneous architecture for computing and computational science. Usage is high, delivering on the promise of a system well-suited for capability simulations for science. This success is due in part to innovations in tracking and reporting the activity on the compute nodes, and using this information to further enable and optimize applications, extending and balancing workload across the entire node. The OLCF continues to invest in innovative processes, tools, and resources necessary to meet continuing user demand. The facility’s leadership in data analysis and workflows was featured at the Department of Energy (DOE) booth at SC15, for the second year in a row, highlighting work with researchers from the National Library of Medicine coupled with unique computational and data resources serving experimental and observational data across facilities. Effective operations of the OLCF play a key role in the scientific missions and accomplishments of its users. Building on the exemplary year of 2014, as shown by the 2014 Operational Assessment Report (OAR) review committee response in Appendix A, this OAR delineates the policies, procedures, and innovations implemented by the OLCF to continue delivering a multi-petaflop resource for cutting-edge research. This report covers CY 2015, which, unless otherwise specified, denotes January 1, 2015, through December 31, 2015.

  19. VO-compliant libraries of high resolution spectra of cool stars

    Science.gov (United States)

    Montes, D.

    2008-10-01

    In this contribution we describe a Virtual Observatory (VO) compliant version of the libraries of high resolution spectra of cool stars described by Montes et al. (1997; 1998; and 1999). Since their publication the fully reduced spectra in FITS format have been available via ftp and in the World Wide Web. However, in the VO all the spectra will be accessible using a common web interface following the standards of the International Virtual Observatory Alliance (IVOA). These libraries include F, G, K and M field stars, from dwarfs to giants. The spectral coverage is from 3800 to 10000 Å, with spectral resolution ranging from 0.09 to 3.0 Å.

  20. Factors Related Management Skills of High School Library Directors in the Republic of China

    Directory of Open Access Journals (Sweden)

    Li-ling Kuo

    1999-04-01

    Full Text Available

    頁次:12-34

    This study attempted to determine the factors related to the management skills of high school library directors in Taiwan, R. o. C. There were five dimensions of the Level of Management Skills (LMS, namely, professionalism, communication, library knowledge and skills, administration, and instructional leadership. The sample size was 201 randomly selected high school library directors in Taiwan. Data were collected by mail questionnaire from July to September 1996. Correlation analyses were conducted to examine the relationships between the factors of the library directors' sum of the LMS scores and independent variables. The regression equation was drawn. In the full model of the regression analysis, derived independent variables, "Support, circulation volumes, and the type of school," "Education and effort," and "Continuing professional education activities in library science" explained the greatest amount of unique variance in the dependent variable, the sum of the LMS scores. This study suggested establishing the standards of basic abilities for school library directors, arranging local unions of high school library directors in the country, re-arranging the core courses of the library majors, and emphasizing the role of "instructional consultant"

  1. mosaicQA - A General Approach to Facilitate Basic Data Quality Assurance for Epidemiological Research.

    Science.gov (United States)

    Bialke, Martin; Rau, Henriette; Schwaneberg, Thea; Walk, Rene; Bahls, Thomas; Hoffmann, Wolfgang

    2017-05-29

    Epidemiological studies are based on a considerable amount of personal, medical and socio-economic data. To answer research questions with reliable results, epidemiological research projects face the challenge of providing high quality data. Consequently, gathered data has to be reviewed continuously during the data collection period. This article describes the development of the mosaicQA-library for non-statistical experts consisting of a set of reusable R functions to provide support for a basic data quality assurance for a wide range of application scenarios in epidemiological research. To generate valid quality reports for various scenarios and data sets, a general and flexible development approach was needed. As a first step, a set of quality-related questions, targeting quality aspects on a more general level, was identified. The next step included the design of specific R-scripts to produce proper reports for metric and categorical data. For more flexibility, the third development step focussed on the generalization of the developed R-scripts, e.g. extracting characteristics and parameters. As a last step the generic characteristics of the developed R functionalities and generated reports have been evaluated using different metric and categorical datasets. The developed mosaicQA-library generates basic data quality reports for multivariate input data. If needed, more detailed results for single-variable data, including definition of units, variables, descriptions, code lists and categories of qualified missings, can easily be produced. The mosaicQA-library enables researchers to generate reports for various kinds of metric and categorical data without the need for computational or scripting knowledge. At the moment, the library focusses on the data structure quality and supports the assessment of several quality indicators, including frequency, distribution and plausibility of research variables as well as the occurrence of missing and extreme values. To

  2. Performance Analysis of Parallel Mathematical Subroutine library PARCEL

    International Nuclear Information System (INIS)

    Yamada, Susumu; Shimizu, Futoshi; Kobayashi, Kenichi; Kaburaki, Hideo; Kishida, Norio

    2000-01-01

    The parallel mathematical subroutine library PARCEL (Parallel Computing Elements) has been developed by Japan Atomic Energy Research Institute for easy use of typical parallelized mathematical codes in any application problems on distributed parallel computers. The PARCEL includes routines for linear equations, eigenvalue problems, pseudo-random number generation, and fast Fourier transforms. It is shown that the results of performance for linear equations routines exhibit good parallelization efficiency on vector, as well as scalar, parallel computers. A comparison of the efficiency results with the PETSc (Portable Extensible Tool kit for Scientific Computations) library has been reported. (author)

  3. A highly redundant BAC library of Atlantic salmon (Salmo salar: an important tool for salmon projects

    Directory of Open Access Journals (Sweden)

    Koop Ben F

    2005-04-01

    Full Text Available Abstract Background As farming of Atlantic salmon is growing as an aquaculture enterprise, the need to identify the genomic mechanisms for specific traits is becoming more important in breeding and management of the animal. Traits of importance might be related to growth, disease resistance, food conversion efficiency, color or taste. To identify genomic regions responsible for specific traits, genomic large insert libraries have previously proven to be of crucial importance. These large insert libraries can be screened using gene or genetic markers in order to identify and map regions of interest. Furthermore, large-scale mapping can utilize highly redundant libraries in genome projects, and hence provide valuable data on the genome structure. Results Here we report the construction and characterization of a highly redundant bacterial artificial chromosome (BAC library constructed from a Norwegian aquaculture strain male of Atlantic salmon (Salmo salar. The library consists of a total number of 305 557 clones, in which approximately 299 000 are recombinants. The average insert size of the library is 188 kbp, representing 18-fold genome coverage. High-density filters each consisting of 18 432 clones spotted in duplicates have been produced for hybridization screening, and are publicly available 1. To characterize the library, 15 expressed sequence tags (ESTs derived overgos and 12 oligo sequences derived from microsatellite markers were used in hybridization screening of the complete BAC library. Secondary hybridizations with individual probes were performed for the clones detected. The BACs positive for the EST probes were fingerprinted and mapped into contigs, yielding an average of 3 contigs for each probe. Clones identified using genomic probes were PCR verified using microsatellite specific primers. Conclusion Identification of genes and genomic regions of interest is greatly aided by the availability of the CHORI-214 Atlantic salmon BAC

  4. Promoting the Effect of the Qing Dynasty Imperial Garden Architectural Component Library on the Digitalization of Cultural Heritage

    Science.gov (United States)

    Jindan, C.; Junsong, Z.; Jiujun, Z.

    2015-08-01

    With the development of computer technology and practical verification, digital virtual technology has matured and is increasingly being widely applied to cultural heritage protection and research. With this advancement in technology, there is pressing need to simplify heritage-related puzzles. Thus the main question that has increasingly become the most central and fundamental problem in heritage digitalization work is how to choose the "proper technology" that provides support directly, accurately and rapidly for the research, protection and exchange of cultural heritage. Based on the principles of "authenticity" and "completeness" found in the Venice Charter in regards to dealing with cultural heritage; this paper proposes the concept of the component library which facilitates the improvement and efficiency of virtual reconstruction, provides a visual discussion platform for cultural heritage protection, virtual scene construction, accuracy assessment, and multi-space-time exhibition; thereby implementing the spirit of tolerance and respect found in the Nara Document on Authenticity. The paper further aims to illustrate the significance of the Qing dynasty imperial garden architectural component library for cultural heritage study and protection, the principles for virtual library construction, use and maintenance of the library, and classification approaches, and also provide some suggestions about making high quality 3D models and effective means for database integration.

  5. Shielding Benchmark Computational Analysis

    International Nuclear Information System (INIS)

    Hunter, H.T.; Slater, C.O.; Holland, L.B.; Tracz, G.; Marshall, W.J.; Parsons, J.L.

    2000-01-01

    Over the past several decades, nuclear science has relied on experimental research to verify and validate information about shielding nuclear radiation for a variety of applications. These benchmarks are compared with results from computer code models and are useful for the development of more accurate cross-section libraries, computer code development of radiation transport modeling, and building accurate tests for miniature shielding mockups of new nuclear facilities. When documenting measurements, one must describe many parts of the experimental results to allow a complete computational analysis. Both old and new benchmark experiments, by any definition, must provide a sound basis for modeling more complex geometries required for quality assurance and cost savings in nuclear project development. Benchmarks may involve one or many materials and thicknesses, types of sources, and measurement techniques. In this paper the benchmark experiments of varying complexity are chosen to study the transport properties of some popular materials and thicknesses. These were analyzed using three-dimensional (3-D) models and continuous energy libraries of MCNP4B2, a Monte Carlo code developed at Los Alamos National Laboratory, New Mexico. A shielding benchmark library provided the experimental data and allowed a wide range of choices for source, geometry, and measurement data. The experimental data had often been used in previous analyses by reputable groups such as the Cross Section Evaluation Working Group (CSEWG) and the Organization for Economic Cooperation and Development/Nuclear Energy Agency Nuclear Science Committee (OECD/NEANSC)

  6. Performance Analysis of Ivshmem for High-Performance Computing in Virtual Machines

    Science.gov (United States)

    Ivanovic, Pavle; Richter, Harald

    2018-01-01

    High-Performance computing (HPC) is rarely accomplished via virtual machines (VMs). In this paper, we present a remake of ivshmem which can change this. Ivshmem was a shared memory (SHM) between virtual machines on the same server, with SHM-access synchronization included, until about 5 years ago when newer versions of Linux and its virtualization library libvirt evolved. We restored that SHM-access synchronization feature because it is indispensable for HPC and made ivshmem runnable with contemporary versions of Linux, libvirt, KVM, QEMU and especially MPICH, which is an implementation of MPI - the standard HPC communication library. Additionally, MPICH was transparently modified by us to get ivshmem included, resulting in a three to ten times performance improvement compared to TCP/IP. Furthermore, we have transparently replaced MPI_PUT, a single-side MPICH communication mechanism, by an own MPI_PUT wrapper. As a result, our ivshmem even surpasses non-virtualized SHM data transfers for block lengths greater than 512 KBytes, showing the benefits of virtualization. All improvements were possible without using SR-IOV.

  7. Automated Library of the Future: Estrella Mountain Community College Center.

    Science.gov (United States)

    Community & Junior College Libraries, 1991

    1991-01-01

    Describes plans for the Integrated High Technology Library (IHTL) at the Maricopa County Community College District's new Estrella Mountain campus, covering collaborative planning, the IHTL's design, and guidelines for the new center and campus (e.g., establishing computing/information-access across the curriculum; developing lifelong learners;…

  8. Study on the application of mobile internet cloud computing platform

    Science.gov (United States)

    Gong, Songchun; Fu, Songyin; Chen, Zheng

    2012-04-01

    The innovative development of computer technology promotes the application of the cloud computing platform, which actually is the substitution and exchange of a sort of resource service models and meets the needs of users on the utilization of different resources after changes and adjustments of multiple aspects. "Cloud computing" owns advantages in many aspects which not merely reduce the difficulties to apply the operating system and also make it easy for users to search, acquire and process the resources. In accordance with this point, the author takes the management of digital libraries as the research focus in this paper, and analyzes the key technologies of the mobile internet cloud computing platform in the operation process. The popularization and promotion of computer technology drive people to create the digital library models, and its core idea is to strengthen the optimal management of the library resource information through computers and construct an inquiry and search platform with high performance, allowing the users to access to the necessary information resources at any time. However, the cloud computing is able to promote the computations within the computers to distribute in a large number of distributed computers, and hence implement the connection service of multiple computers. The digital libraries, as a typical representative of the applications of the cloud computing, can be used to carry out an analysis on the key technologies of the cloud computing.

  9. Higher-Order Factor Analysis as a Score Validity Evaluation Tool: An Example with a Measure of Perceptions of Library Service Quality.

    Science.gov (United States)

    Cook, Colleen; Thompson, Bruce

    The SERVQUAL measure was developed by A. Parasuraman, L. Berry, and V. Zeithaml (1988) to measure perceptions of service quality, originally in the retailing sector. However, libraries and other educational institutions are also service providers. Librarians in particular have increasingly become interested in measuring quality of service as the…

  10. Using SERVQUAL in health libraries across Somerset, Devon and Cornwall.

    Science.gov (United States)

    Martin, Susan

    2003-03-01

    This study provides the results of a survey conducted in the autumn of 2001 by ten NHS library services across Somerset, Devon and Cornwall. The aim of the project was to measure the service quality of each individual library and to provide an overall picture of the quality of library services within the south-west peninsula. The survey was based on SERVQUAL, a diagnostic tool developed in the 1980s, which measures service quality in terms of customer expectations and perceptions of service. The survey results have provided the librarians with a wealth of information about service quality. The service as a whole is perceived to be not only meeting but also exceeding expectations in terms of reliability, responsiveness, empathy and assurance. For the first time, the ten health library services can measure their own service quality as well as benchmark themselves against others.

  11. Status of the JEFF nuclear data library

    International Nuclear Information System (INIS)

    Koning, A.J.; Bauge, E.; Dean, C.J.; Dupont, E.; Nordborg, C.; Rugama, Y.; Fischer, U.; Forrest, R.A.; Kellett, M.A.; Jacqmin, R.; Leeb, H.; Mills, R.W.; Pescarini, M.; Rullhusen, P.

    2011-01-01

    The status of the Joint Evaluated Fission and Fusion file (JEFF) is described. Recently, the JEFF-3.1.1 nuclear data library was released and shortly after adopted by the French nuclear power industry for inclusion in their production and analysis codes. Recent updates include actinide evaluations, materials evaluations that have emerged from various European nuclear data projects, the activation library, the decay data and fission yield sub-libraries, and fusion-related data files from the European F4E project. The revisions were motivated by the availability of new measurements, modelling capabilities and trends from integral experiments. Validations have been performed, mainly for criticality, reactivity temperature coefficients, fuel inventory, decay heat and shielding of thermal and fast systems. The next release of the library, JEFF-3.2, will be discussed. This will contain among others a significant increase of covariance data evaluations, modern evaluations for various structural materials, a larger emphasis on minor actinides and addition of high-quality gamma production data for many fission products. (authors)

  12. Libraries and key performance indicators a framework for practitioners

    CERN Document Server

    Appleton, Leo

    2017-01-01

    Libraries and Key Performance Indicators: A Framework for Practitioners explores ways by which libraries across all sectors can demonstrate their value and impact to stakeholders through quality assurance and performance measurement platforms, including library assessment, evaluation methodologies, surveys, and annual reporting. Whilst several different performance measurement tools are considered, the book's main focus is on one tool in particular: Key Performance Indicators (KPIs). KPIs are increasingly being used to measure the performance of library and information services, however, linking KPIs to quality outcomes, such as impact and value can prove very difficult. This book discusses, in detail, the concept of KPIs in the broader context of library assessment and performance measurement. Through reviewing some of the applied theory around using KPIs, along with harvesting examples of current best practices in KPI usage from a variety of different libraries, the book demystifies library KPIs, providing ...

  13. Cross sections in 25 groups obtained from ENDF/B-IV and ENDL/78 libraries, processed with GALAXY and NJOY computer codes

    International Nuclear Information System (INIS)

    Chalhoub, E.S.; Corcuera, R.P.

    1982-01-01

    The discrepancies existing between ENDF/B-IV and ENDL/78 libraries, in diferent energy regions are identified, and the order of the differences in multigroup sections are determined, when GALAXY or NJOY computer codes are used. (E.G.) [pt

  14. Measuring Effectiveness in a Virtual Library

    Science.gov (United States)

    Finch, Jannette L.

    2010-01-01

    Measuring quality of service in academic libraries traditionally includes quantifiable data such as collection size, staff counts, circulation numbers, reference service statistics, qualitative analyses of customer satisfaction, shelving accuracy, and building comfort. In the libraries of the third millennium, virtual worlds, Web content and…

  15. Managing Academic Libraries with Fewer Resources.

    Science.gov (United States)

    Riggs, Donald E.

    1992-01-01

    A discussion of academic library management during retrenchment looks at a variety of issues, including staffing needs in the labor-intensive library environment, acquisitions budgeting, interlibrary cooperation (ownership vs. access to resources), entrepreneurship and strategic planning for problem solving, and use of total quality management…

  16. Helping the Hacker? Library Information, Security, and Social Engineering

    Directory of Open Access Journals (Sweden)

    Samuel T. C. Thompson

    2006-12-01

    Full Text Available Social engineering is the use of nontechnical means to gain unauthorized access to information or computer systems. While this method is recognized as a major security threat in the computer industry, little has been done to address it in the library field. This is of particular concern because libraries increasingly have access to databases of both proprietary and personal information. This tutorial is designed to increase the awareness of library staff in regard to the issue of social engineering.

  17. Building Earth's Largest Library: Driving into the Future.

    Science.gov (United States)

    Coffman, Steve

    1999-01-01

    Examines the Amazon.com online bookstore as a blueprint for designing the world's largest library. Topics include selection; accessibility and convenience; quality of Web sites and search tools; personalized service; library collection development, including interlibrary loan; library catalogs and catalog records; a circulation system; costs;…

  18. Keynote Address: Science Libraries in the Information Age

    Science.gov (United States)

    Huchra, J. P.

    2007-10-01

    The invention of the printing press transformed the world, especially science and education. Today we are witnessing a similar transformation driven by electronic detection, computers and the internet. Where the Gutenberg transformation greatly enabled libraries, the silicon transformation requires libraries to change. The scientific community has new needs and libraries must have new roles in meeting those needs.

  19. Technical Considerations for Reduced Representation Bisulfite Sequencing with Multiplexed Libraries

    Science.gov (United States)

    Chatterjee, Aniruddha; Rodger, Euan J.; Stockwell, Peter A.; Weeks, Robert J.; Morison, Ian M.

    2012-01-01

    Reduced representation bisulfite sequencing (RRBS), which couples bisulfite conversion and next generation sequencing, is an innovative method that specifically enriches genomic regions with a high density of potential methylation sites and enables investigation of DNA methylation at single-nucleotide resolution. Recent advances in the Illumina DNA sample preparation protocol and sequencing technology have vastly improved sequencing throughput capacity. Although the new Illumina technology is now widely used, the unique challenges associated with multiplexed RRBS libraries on this platform have not been previously described. We have made modifications to the RRBS library preparation protocol to sequence multiplexed libraries on a single flow cell lane of the Illumina HiSeq 2000. Furthermore, our analysis incorporates a bioinformatics pipeline specifically designed to process bisulfite-converted sequencing reads and evaluate the output and quality of the sequencing data generated from the multiplexed libraries. We obtained an average of 42 million paired-end reads per sample for each flow-cell lane, with a high unique mapping efficiency to the reference human genome. Here we provide a roadmap of modifications, strategies, and trouble shooting approaches we implemented to optimize sequencing of multiplexed libraries on an a RRBS background. PMID:23193365

  20. MYTHS vesus reality in computed radiography image quality

    International Nuclear Information System (INIS)

    Mango, Steve; Castro, Luiz

    2009-01-01

    As NDE operation - particularly radiographic testing - ransition form analog to digital technologies such as computed radiography (CR), users are learning that there's more to digital image quality than meets the eye. In fact, there are ultiple factors that determine the final perceived image quality of a computed radiograph. Many of these factors are misunderstood, and some are touted as the ''key parameter'' or ''magic bullet'' in producing optiumum image quality, In reality, such claims are oversimplified, and are more marketing hype than reality. The truth?. Perceived image quality results form the cascaded effects of many factor - such as sharpness, system noise, spot size and pixel size, subject contrast, bit depth, radiographic technique, and so on. Many of these factors are within the control of rdiographers or designers of equipment and media. This paper will explain some of these key factors, dispel some of the myths surrounding them, and will show that qualities such as bigger, smaller, more, or less are not always better when it comes to CR image quality. (authors)

  1. Pulse shape analysis and position determination in segmented HPGe detectors: The AGATA detector library

    Energy Technology Data Exchange (ETDEWEB)

    Bruyneel, B. [Universitaet zu Koeln, Institut fuer Kernphysik, Koeln (Germany); Service de Physique Nucleaire, CEA Saclay, Gif-sur-Yvette (France); Birkenbach, B.; Reiter, P. [Universitaet zu Koeln, Institut fuer Kernphysik, Koeln (Germany)

    2016-03-15

    The AGATA Detector Library (ADL) was developed for the calculation of signals from highly segmented large volume high-purity germanium (HPGe) detectors. ADL basis sets comprise a huge amount of calculated position-dependent detector pulse shapes. A basis set is needed for Pulse Shape Analysis (PSA). By means of PSA the interaction position of a γ -ray inside the active detector volume is determined. Theoretical concepts of the calculations are introduced and cover the relevant aspects of signal formation in HPGe. The approximations and the realization of the computer code with its input parameters are explained in detail. ADL is a versatile and modular computer code; new detectors can be implemented in this library. Measured position resolutions of the AGATA detectors based on ADL are discussed. (orig.)

  2. Students lead the library the importance of student contributions to the academic library

    CERN Document Server

    Arnold-Garza, Sara

    2017-01-01

    In six parts-Students as Employees, Students as Curators, Students as Ambassadors, the Library as Client, Student Groups as Library Leaders, and Students as Library Designers-Students Lead the Library provides case studies of programs and initiatives that seek student input, assistance, and leadership in the academic library. Through the library, students can develop leadership skills, cultivate high levels of engagement, and offer peer learning opportunities. Through the students, libraries can create participatory design processes, enhancement and transformation of the library's core functions, and expressed library value for stakeholders.

  3. Rad World -- computer-animated video radiation and hazardous waste-management science curriculum

    International Nuclear Information System (INIS)

    Powell, B.

    1996-01-01

    The Rad World computer-animated video and curriculum materials were developed through a grant from the Waste-management Education and Research Consortium. The package, which includes a computer-animated video, hands-on activities, and multidisciplinary lessons concerning radiation and hazardous-waste management, was created to approach these subjects in an informative, yet entertaining, manner. The lessons and video, designed to supplement studies of energy and physical science at the middle school and high school level, also implement quality and consistent science education as outlined by the New Mexico Science Standards and Benchmarks (1995). Consistent with the curriculum standards and benchmarks, the curriculum includes library research, collaborative learning, hands-on-science, and discovery learning. Pre- and post-tests are included

  4. The Adoption of Innovations in Library Organizations

    Directory of Open Access Journals (Sweden)

    Su-may Sheih Chen

    2001-12-01

    Full Text Available To enhance the quality of library and information service, libraries have borrowed from business theory and practice in planning and evaluating library services. However, an area of business theory that has received relatively little attention in library science is organizational innovation. Reports in the business literature provide evidence that innovation is essential for both development and survival. In increasingly unstable times of rapidly changing information technology, libraries must use the literature of innovation as well as that of planning and evaluation to shape better libraries. This paper explores key issues related to innovation in libraries including the nature, the barriers, and the ways to adopt and implement innovative strategies successfully. [Article content in Chinese

  5. Information Literacy, Learning, and the Public Library: A Study of Danish High School Students

    Science.gov (United States)

    Nielsen, Bo Gerner; Borlund, Pia

    2011-01-01

    The paper reports on a study of 12 Danish high school students' perceptions of public libraries' role in learning, user education, information literacy, and librarians' information competencies. The study is undertaken by use of literature review and interviews with a purposive select sample of public library users in Denmark. The study…

  6. The Vendors' Corner: Biblio-Techniques' Library and Information System (BLIS).

    Science.gov (United States)

    Library Software Review, 1984

    1984-01-01

    Describes online catalog and integrated library computer system designed to enhance Washington Library Network's software. Highlights include system components; implementation options; system features (integrated library functions, database design, system management facilities); support services (installation and training, software maintenance and…

  7. Multimedia technologies in university libraries: opportunities and tasks

    Directory of Open Access Journals (Sweden)

    Tavalbekh Serkhan Ali

    2010-03-01

    Full Text Available Multimedia technologies (MMT are considered as a factor of qualitative changes of informational environment of education. Leading role of libraries in the process of integration of MMT environment of university is determined. Influence of libraries experience in usage of informational technologies both traditional and computer-aided, Internet in particular is observed. Determined that introduction of MMT noticeable improving the importance of library in MMT environment organization of universities, improvement of library informational service. Tasks of university libraries with integration into united media space are defining.

  8. On library information resources construction under network environment

    International Nuclear Information System (INIS)

    Guo Huifang; Wang Jingjing

    2014-01-01

    Information resources construction is the primary task and critical measures for libraries. In the 2lst century, the knowledge economy era, with the continuous development of computer network technology, information resources have become an important part of libraries which have been a significant indicator of its capacity construction. The development of socialized Information, digitalization and internalization has put forward new requirements for library information resources construction. This paper describes the impact of network environment on construction of library information resources and proposes the measures of library information resources. (authors)

  9. Computer based workstation for development of software for high energy physics experiments

    International Nuclear Information System (INIS)

    Ivanchenko, I.M.; Sedykh, Yu.V.

    1987-01-01

    Methodical principles and results of a successful attempt to create on the base of IBM-PC/AT personal computer of effective means for development of programs for high energy physics experiments are analysed. The obtained results permit to combine the best properties and a positive materialized experience accumulated on the existing time sharing collective systems with a high quality of data representation, reliability and convenience of personal computer applications

  10. USGS Spectral Library Version 7

    Science.gov (United States)

    Kokaly, Raymond F.; Clark, Roger N.; Swayze, Gregg A.; Livo, K. Eric; Hoefen, Todd M.; Pearson, Neil C.; Wise, Richard A.; Benzel, William M.; Lowers, Heather A.; Driscoll, Rhonda L.; Klein, Anna J.

    2017-04-10

    We have assembled a library of spectra measured with laboratory, field, and airborne spectrometers. The instruments used cover wavelengths from the ultraviolet to the far infrared (0.2 to 200 microns [μm]). Laboratory samples of specific minerals, plants, chemical compounds, and manmade materials were measured. In many cases, samples were purified, so that unique spectral features of a material can be related to its chemical structure. These spectro-chemical links are important for interpreting remotely sensed data collected in the field or from an aircraft or spacecraft. This library also contains physically constructed as well as mathematically computed mixtures. Four different spectrometer types were used to measure spectra in the library: (1) Beckman™ 5270 covering the spectral range 0.2 to 3 µm, (2) standard, high resolution (hi-res), and high-resolution Next Generation (hi-resNG) models of Analytical Spectral Devices (ASD) field portable spectrometers covering the range from 0.35 to 2.5 µm, (3) Nicolet™ Fourier Transform Infra-Red (FTIR) interferometer spectrometers covering the range from about 1.12 to 216 µm, and (4) the NASA Airborne Visible/Infra-Red Imaging Spectrometer AVIRIS, covering the range 0.37 to 2.5 µm. Measurements of rocks, soils, and natural mixtures of minerals were made in laboratory and field settings. Spectra of plant components and vegetation plots, comprising many plant types and species with varying backgrounds, are also in this library. Measurements by airborne spectrometers are included for forested vegetation plots, in which the trees are too tall for measurement by a field spectrometer. This report describes the instruments used, the organization of materials into chapters, metadata descriptions of spectra and samples, and possible artifacts in the spectral measurements. To facilitate greater application of the spectra, the library has also been convolved to selected spectrometer and imaging spectrometers sampling and

  11. Human Resource Development in Hybrid Libraries

    OpenAIRE

    Prakasan, E. R.; Swarna, T.; Vijai Kumar, *

    2000-01-01

    This paper explores the human resources and development implications in hybrid libraries. Due to technological changes in libraries, which is a result of the proliferation of electronic resources, there has been a shift in workloads and workflow, requiring staff with different skills and educational backgrounds. Training of staff at all levels in information technology is the key to manage change, alleviate anxiety in the workplace and assure quality service in the libraries. Staff developmen...

  12. Library Benchmarking

    Directory of Open Access Journals (Sweden)

    Wiji Suwarno

    2017-02-01

    Full Text Available The term benchmarking has been encountered in the implementation of total quality (TQM or in Indonesian termed holistic quality management because benchmarking is a tool to look for ideas or learn from the library. Benchmarking is a processof measuring and comparing for continuous business process of systematic and continuous measurement, the process of measuring and comparing for continuous business process of an organization to get information that can help these organization improve their performance efforts.

  13. Quality assurance of nuclear medicine computer software

    International Nuclear Information System (INIS)

    Cradduck, T.D.

    1986-01-01

    Although quality assurance activities have become well established for the hardware found in nuclear medicine little attention has been paid to computer software. This paper outlines some of the problems that exist and indicates some of the solutions presently under development. The major thrust has been towards establishment of programming standards and comprehensive documentation. Some manufacturers have developed installation verification procedures which programmers are urged to use as models for their own programs. Items that tend to cause erroneous results are discussed with the emphasis for error detection and correction being placed on proper education and training of the computer operator. The concept of interchangeable data files or 'software phantoms' for purposes of quality assurance is discussed. (Author)

  14. What is exciting at the libraries? – GIS resources at Libraries

    OpenAIRE

    Kong, Nicole

    2017-01-01

    This presentation introduce the new GIS resources offered from Purdue University Libraries. These new support areas include the D-VELoP GIS computing lab, new Esri Development Center (EDC), GIS data portal, GIS data server, Insights for ArcGIS, and GIS VR technology.

  15. Image quality of low-dose CCTA in obese patients: impact of high-definition computed tomography and adaptive statistical iterative reconstruction.

    Science.gov (United States)

    Gebhard, Cathérine; Fuchs, Tobias A; Fiechter, Michael; Stehli, Julia; Stähli, Barbara E; Gaemperli, Oliver; Kaufmann, Philipp A

    2013-10-01

    The accuracy of coronary computed tomography angiography (CCTA) in obese persons is compromised by increased image noise. We investigated CCTA image quality acquired on a high-definition 64-slice CT scanner using modern adaptive statistical iterative reconstruction (ASIR). Seventy overweight and obese patients (24 males; mean age 57 years, mean body mass index 33 kg/m(2)) were studied with clinically-indicated contrast enhanced CCTA. Thirty-five patients underwent a standard definition protocol with filtered backprojection reconstruction (SD-FBP) while 35 patients matched for gender, age, body mass index and coronary artery calcifications underwent a novel high definition protocol with ASIR (HD-ASIR). Segment by segment image quality was assessed using a four-point scale (1 = excellent, 2 = good, 3 = moderate, 4 = non-diagnostic) and revealed better scores for HD-ASIR compared to SD-FBP (1.5 ± 0.43 vs. 1.8 ± 0.48; p ASIR as compared to 1.4 ± 0.4 mm for SD-FBP (p ASIR (388.3 ± 109.6 versus 350.6 ± 90.3 Hounsfield Units, HU; p ASIR vs. SD-ASIR respectively). Compared to a standard definition backprojection protocol (SD-FBP), a newer high definition scan protocol in combination with ASIR (HD-ASIR) incrementally improved image quality and visualization of distal coronary artery segments in overweight and obese individuals, without increasing image noise and radiation dose.

  16. The Human Element in the Virtual Library.

    Science.gov (United States)

    Saunders, Laverna M.

    1999-01-01

    Introduces the concept of the virtual library and explores how the increasing reliance on computers and digital information has affected library users and staff. Discusses users' expectations, democratization of access, human issues, organizational change, technostress, ergonomics, assessment, and strategies for success and survival. Contains 35…

  17. Educational Resources in the ASCC Library

    Science.gov (United States)

    Lin, Steven

    2006-01-01

    After two years of construction, American Samoa Community College opened its new library on September 2, 2003. The library is located on the east side of campus and is equipped with ten computer workstations, four online public access catalogs, three copying machines, and an elevator that is in compliance with the Americans with Disabilities Act.…

  18. USE OF LIBRARIES IN OPEN AND DISTANCE LEARNING SYSTEM: Barriers to the Use of AIOU Libraries by Tutors and Students

    Directory of Open Access Journals (Sweden)

    Abdul Jabbar BHATTI,

    2012-04-01

    Full Text Available This study explores; the library needs of students and tutors of Allama Iqbal Open University (AIOU, utilization level of the library facilities and resources, the problems in the use of library, and suggestions for improvement of library facilities for students and tutors. Data collected from 4080 students and 526 tutors belonging to 15 different regional offices showed that students and tutors needed library for various educational purposes, the regional libraries were not being used much, and both tutors and students were facing various problems such as unsuitable library timing, long distance between library and their residence, non availability of latest journals, non availability of required material, lack of temperature control in the library, insufficient study area, lack of latest books, and inadequate staff. For improving library facility at regional level, the students and tutors suggested to; provide more books and journals, expand library timings, arrange library facility at workshop venues, make arrangements to advertise the resources and services at the library to the students, provide computers and internet service, provide trained staff, and arrange partnership with other academic libraries.

  19. The National Library of Medicine Programs and Services, Fiscal Year 1974.

    Science.gov (United States)

    National Library of Medicine (DHEW), Bethesda, MD.

    The activities and projects of the National Library of Medicine are described. New and continuing programs in library services and operations, on-line computer retrieval services, grants for library assistance, audiovisual programs, and health communications research are included. International activities of the Library are outlined. Summary…

  20. 2008 The State of ICT in University Libraries in South Eastern ...

    African Journals Online (AJOL)

    Gbaje E.S

    the library can boast of standalone computers. Library software, network operating systems and LAN are conspicuously absent in most of the libraries. Though a few of the libraries subscribe to online electronic resources such as EBSCOhost, Access to Global Online Research in Agriculture (AGORA) , Health. Internetwork ...

  1. SCEAPI: A unified Restful Web API for High-Performance Computing

    Science.gov (United States)

    Rongqiang, Cao; Haili, Xiao; Shasha, Lu; Yining, Zhao; Xiaoning, Wang; Xuebin, Chi

    2017-10-01

    The development of scientific computing is increasingly moving to collaborative web and mobile applications. All these applications need high-quality programming interface for accessing heterogeneous computing resources consisting of clusters, grid computing or cloud computing. In this paper, we introduce our high-performance computing environment that integrates computing resources from 16 HPC centers across China. Then we present a bundle of web services called SCEAPI and describe how it can be used to access HPC resources with HTTP or HTTPs protocols. We discuss SCEAPI from several aspects including architecture, implementation and security, and address specific challenges in designing compatible interfaces and protecting sensitive data. We describe the functions of SCEAPI including authentication, file transfer and job management for creating, submitting and monitoring, and how to use SCEAPI in an easy-to-use way. Finally, we discuss how to exploit more HPC resources quickly for the ATLAS experiment by implementing the custom ARC compute element based on SCEAPI, and our work shows that SCEAPI is an easy-to-use and effective solution to extend opportunistic HPC resources.

  2. Students' Perception on Library Service Quality: A Qualitative Study of IIUM Library

    Science.gov (United States)

    Ahmed, Selim; Islam, Rafikul

    2012-01-01

    Academic libraries are currently facing their greatest challenge since the explosion in tertiary education and academic publishing, which began after World War II. The global digital revolution is affecting both the traditional forms of the creation, organisation, and dissemination of knowledge, and the world of tertiary education itself. In the…

  3. The impact of complaint management and service quality on organizational image: A case study at the Malaysian public university library

    Science.gov (United States)

    Tan, Pei Kian; Mohd Suradi, Nur Riza; Saludin, Mohamad Nasir

    2013-04-01

    Service failure frequently occurs. This affects customer expectations which lead to complaint. However, not all dissatisfied customers actually complain. Without customer feedback, it would be impossible for a company to know whether they needed a change for improvement. Thus, complaint management brings a learning experience to organization in order to provide better service. Therefore, it is important to identify customer dissatisfaction through a systematic complaint handling or management. The study proposes a model of systematic complaint management which applied to academic library as a tool of service recovery. As such, the main purpose of this study is to investigate the critical success factors of complaint management towards service quality, customer satisfaction, customer loyalty and the impact to organizational image at academic library. Three academic libraries have been identified and selected for this project, the Library of Tun Sri Lanang, UKM, UTeM and UNIMAS. Using the justice theory, this study investigates the perception of customers on complaint management in terms of outcomes they receive, procedures used by organization and interpersonal treatment. In this study, there are five factors of complaint management identified, which includes speed of recovery, management system, empowerment, culture and psychology and tangible compensation. A questionnaire was designed and used as the data gathering instrument. A total of 600 respondents participated in this study. Ten hypotheses were used to test the relationships between complaint management, service quality, customer satisfaction, customer loyalty and organizational image. To measure the construct relationships, Structural Equation Model (SEM) approach was used. The results show that management system (b = 0.210; p 0.05) do not influence service quality. The second part of this study uses confirmatory factor analysis (CFA) to analyze and confirm the conceptual model proposed in this research

  4. Data Sets, Ensemble Cloud Computing, and the University Library (Invited)

    Science.gov (United States)

    Plale, B. A.

    2013-12-01

    The environmental researcher at the public university has new resources at their disposal to aid in research and publishing. Cloud computing provides compute cycles on demand for analysis and modeling scenarios. Cloud computing is attractive for e-Science because of the ease with which cores can be accessed on demand, and because the virtual machine implementation that underlies cloud computing reduces the cost of porting a numeric or analysis code to a new platform. At the university, many libraries at larger universities are developing the e-Science skills to serve as repositories of record for publishable data sets. But these are confusing times for the publication of data sets from environmental research. The large publishers of scientific literature are advocating a process whereby data sets are tightly tied to a publication. In other words, a paper published in the scientific literature that gives results based on data, must have an associated data set accessible that backs up the results. This approach supports reproducibility of results in that publishers maintain a repository for the papers they publish, and the data sets that the papers used. Does such a solution that maps one data set (or subset) to one paper fit the needs of the environmental researcher who among other things uses complex models, mines longitudinal data bases, and generates observational results? The second school of thought has emerged out of NSF, NOAA, and NASA funded efforts over time: data sets exist coherent at a location, such as occurs at National Snow and Ice Data Center (NSIDC). But when a collection is coherent, reproducibility of individual results is more challenging. We argue for a third complementary option: the university repository as a location for data sets produced as a result of university-based research. This location for a repository relies on the expertise developing in the university libraries across the country, and leverages tools, such as are being developed

  5. Automatic Generation of Agents using Reusable Soft Computing Code Libraries to develop Multi Agent System for Healthcare

    OpenAIRE

    Priti Srinivas Sajja

    2015-01-01

    This paper illustrates architecture for a multi agent system in healthcare domain. The architecture is generic and designed in form of multiple layers. One of the layers of the architecture contains many proactive, co-operative and intelligent agents such as resource management agent, query agent, pattern detection agent and patient management agent. Another layer of the architecture is a collection of libraries to auto-generate code for agents using soft computing techni...

  6. Outsourcing in American Libraries--An Overview.

    Science.gov (United States)

    Bordeianu, Sever; Benaud, Claire-Lise

    1997-01-01

    Discusses the state of outsourcing in American libraries. Highlights include objectives (to reduce cost, increase the quality of service, and achieve a better price/performance objective); operations that can be outsourced; pros and cons; changes in the way library personnel view their work; outsourcing in special, public, academic, and federal…

  7. Influencing Database Use in Public Libraries.

    Science.gov (United States)

    Tenopir, Carol

    1999-01-01

    Discusses results of a survey of factors influencing database use in public libraries. Highlights the importance of content; ease of use; and importance of instruction. Tabulates importance indications for number and location of workstations, library hours, availability of remote login, usefulness and quality of content, lack of other databases,…

  8. Differences between cross-section libraries for neutron dosimetry

    International Nuclear Information System (INIS)

    Tardelli, T.C.; Stecher, L.C.; Coelho, T.S.; Castro, V.A. De; Cavalieri, T.A.; Menzel, F.; Giarola, R.S.; Domingos, D.B.; Yoriyaz, H.

    2013-01-01

    Absorbed dose calculations depend on a consistent set of nuclear data used in simulations in computer codes. Nuclear data are stored in libraries, however, the information available about the differences in dose caused by different libraries are rare. The libraries are processed by a computer system to be able to be used by a radiation transport code. One of the systems capable of processing nuclear data is the NJOY system. The objective of this study is to evaluate the nuclear data libraries for neutrons available in the literature, and to quantify the differences in absorbed dose obtained using the libraries JENDL 4.0, JEFF 3.3.1 and ENDF/B.VII. The absorbed dose calculation was performed on a simple geometric model, as spheres, and in anthropomorphic model of the human body based on the ICRP-110 for neutron transport simulation using the MCNP5 code. The results were compared with literature data. The results obtained with cross sections from the libraries JEFF and ENDF/B.VII have shown to be identical in most cases, except for one case where the difference has exceeded 10%. The results obtained with JENDL library has shown to be considerably different in most cases comparing to other two libraries. Some differences were over 200%. The dose calculations showed differences between the libraries, which is justified by differences in the cross sections. It has been observed that the cross sections values of certain nuclides assume quite different values in different libraries. These differences in turn cause considerable differences in dose calculations. (author)

  9. Research into Learning Resulting from Quality School Library Media Service.

    Science.gov (United States)

    Marchant, Maurice P.; And Others

    1984-01-01

    This annotated bibliography of 20 research reports identifies what has been determined about the effects of library media services on learning and suggests methodologies available for similar studies. Organization is according to area of learning affected--academic achievement; language, reading, and library skills; mathematics; science; social…

  10. The relationship between computer games and quality of life in adolescents.

    Science.gov (United States)

    Dolatabadi, Nayereh Kasiri; Eslami, Ahmad Ali; Mostafavi, Firooze; Hassanzade, Akbar; Moradi, Azam

    2013-01-01

    Term of doing computer games among teenagers is growing rapidly. This popular phenomenon can cause physical and psychosocial issues in them. Therefore, this study examined the relationship between computer games and quality of life domains in adolescents aging 12-15 years. In a cross-sectional study using the 2-stage stratified cluster sampling method, 444 male and female students in Borkhar were selected. The data collection tool consisted of 1) World Health Organization Quality Of Life - BREF questionnaire and 2) personal information questionnaire. The data were analyzed by Pearson correlation, Spearman correlation, chi-square, independent t-tests and analysis of covariance. The total mean score of quality of life in students was 67.11±13.34. The results showed a significant relationship between the age of starting to play games and the overall quality of life score and its fourdomains (range r=-0.13 to -0.18). The mean of overall quality of life score in computer game users was 68.27±13.03 while it was 64.81±13.69 among those who did not play computer games and the difference was significant (P=0.01). There were significant differences in environmental and mental health domains between the two groups (Pcomputer games. Playing computer games for a short time under parental supervision can have positive effects on quality of life in adolescents. However, spending long hours for playing computer games may have negative long-term effects.

  11. Terminals for the interactive input and editing of bibliographic records on the PDP-11 computer in the CERN Library

    International Nuclear Information System (INIS)

    Van Praag, A.

    1975-01-01

    As part of the Library Mechanization Project, special computer terminals have been designed and built for the input and display of a wide range of characters, including Greek letters and mathematical symbols. Each terminal consists of a novel 7-bank keyboard together with two Tektronix 611 storage-tube displays, and there are four terminals connected to a PDP-11/20 computer. After an outline of the hardware for the terminal system, this report gives fuller accounts of the computer interface, display driver, matrix character generator, keyboard, programmable clock, and address comparator. More detailed descriptions of the circuits, the signal formats, and a complete set of schematic diagrams are given as Appendices. (Author)

  12. Phage display peptide libraries: deviations from randomness and correctives

    Science.gov (United States)

    Ryvkin, Arie; Ashkenazy, Haim; Weiss-Ottolenghi, Yael; Piller, Chen; Pupko, Tal; Gershoni, Jonathan M

    2018-01-01

    Abstract Peptide-expressing phage display libraries are widely used for the interrogation of antibodies. Affinity selected peptides are then analyzed to discover epitope mimetics, or are subjected to computational algorithms for epitope prediction. A critical assumption for these applications is the random representation of amino acids in the initial naïve peptide library. In a previous study, we implemented next generation sequencing to evaluate a naïve library and discovered severe deviations from randomness in UAG codon over-representation as well as in high G phosphoramidite abundance causing amino acid distribution biases. In this study, we demonstrate that the UAG over-representation can be attributed to the burden imposed on the phage upon the assembly of the recombinant Protein 8 subunits. This was corrected by constructing the libraries using supE44-containing bacteria which suppress the UAG driven abortive termination. We also demonstrate that the overabundance of G stems from variant synthesis-efficiency and can be corrected using compensating oligonucleotide-mixtures calibrated by mass spectroscopy. Construction of libraries implementing these correctives results in markedly improved libraries that display random distribution of amino acids, thus ensuring that enriched peptides obtained in biopanning represent a genuine selection event, a fundamental assumption for phage display applications. PMID:29420788

  13. Reflection on the talent structure of knowledge-service oriented nuclear technology library

    International Nuclear Information System (INIS)

    Zhang Xue; Zhang Ruiping

    2010-01-01

    Nuclear technology library is the only authoritative organization in collection of nuclear technology literatures.It has exceptional advantage and a large number of customers with great requirement. With promotion of network and digitization of information resource, new situation is posed before nuclear technology library-transforming from traditional library to knowledge-service oriented library. In order to carry on knowledge service effectively and conveniently, a variety of talents are essential. So establishing a talent team with high quality and complete specialities is the fundamental guarantee. Based on a great deal research and discussion, requirements for establishment of a talent team are put forward in the paper and suggestion are present: 5 basic specialized talents are required in nuclear technology library, including organization and management talent, basic operation talent, search service talent, technology application talent, information development talent. (authors)

  14. Weaving libraries into the web OCLC 1998-2008

    CERN Document Server

    Jordan, Jay

    2013-01-01

    The year 1997 found the members of the OCLC (Online Computer Library Center) cooperative in an expansive mood. More than 1,000 library leaders attended the OCLC President's Luncheon in San Francisco, where they celebrated OCLC's 30th anniversary. There were more than 25,000 libraries participating in the cooperative, including nearly 3,000 libraries in 62 countries outside the U.S., and the WorldCat database contained more than 37 million bibliographic records. Over the next ten years, the global digital library would indeed emerge, but in a form that few could have predicted. Ag

  15. A library treasure hunt – An alternative way to introduce new university students to the library

    Directory of Open Access Journals (Sweden)

    Kristina Holmin Verdozzi

    2013-12-01

    Full Text Available Introducing new students to the library and its resources during the first hectic weeks at university can pose a considerable challenge. We have to compete with many other kinds of introductory activities, making new friends, becoming acquainted with the subject, not to mention all the parties and social functions arranged for new students. The traditional lecture, or showing large groups of tired students around the library just wasn't working. So, a few years ago, we started to think about radical new ways of introducing students to the library. We hit upon the idea of arranging a treasure hunt. Students working in small groups have to carry out various tasks at different stations in different libraries. When each task has been successfully completed, a verbal report is made to the librarian at that library, and students can ask questions, before going on to the next task. By carrying out actual tasks, the students become familiar with the important aspects of library resources; learning by doing and having fun at the same time. The "treasure" at the end of the hunt was an apple, a sweet, a pencil, a small LED flashlight and information brochures about the library and the master programme in physics. One of the decisive factors in the success of this activity was the cooperation of lecturers, who integrated it into their introductory programme. It also had high status, as it was a compulsory activity, thanks to the commitment and understanding of lecturers. The treasure hunt is important in the students' later studies as they know which library resources are available, and they recognise the librarians. This lowers the threshold for the further development of the students' information gathering skills. The success of this activity in one of the subjects taught at the Department of Physics has led to the decision to make the library treasure hunt compulsory for all students at the Department, in both the Engineering Faculty and the Science

  16. Computing in high energy physics

    Energy Technology Data Exchange (ETDEWEB)

    Watase, Yoshiyuki

    1991-09-15

    The increasingly important role played by computing and computers in high energy physics is displayed in the 'Computing in High Energy Physics' series of conferences, bringing together experts in different aspects of computing - physicists, computer scientists, and vendors.

  17. What's New in the Library Automation Arena?

    Science.gov (United States)

    Breeding, Marshall

    1998-01-01

    Reviews trends in library automation based on vendors at the 1998 American Library Association Annual Conference. Discusses the major industry trend, a move from host-based computer systems to the new generation of client/server, object-oriented, open systems-based automation. Includes a summary of developments for 26 vendors. (LRW)

  18. IT Strategy for the Library of Congress.

    Science.gov (United States)

    Inouye, Alan

    2000-01-01

    Presents an abstract for a planned technical session to discuss the report of the Committee on the Information Technology Strategy of the Library of Congress, developed by the Computer Science and Telecommunications Board of the National Academies. Highlights include digital information, Web links, preservation, and the management of libraries.…

  19. Generation of a Broad-Group HTGR Library for Use with SCALE

    International Nuclear Information System (INIS)

    Ellis, Ronald James; Lee, Deokjung; Wiarda, Dorothea; Williams, Mark L.; Mertyurek, Ugur

    2012-01-01

    With current and ongoing interest in high temperature gas reactors (HTGRs), the U.S. Nuclear Regulatory Commission (NRC) anticipates the need for nuclear data libraries appropriate for use in applications for modeling, assessing, and analyzing HTGR reactor physics and operating behavior. The objective of this work was to develop a broad-group library suitable for production analyses with SCALE for HTGR applications. Several interim libraries were generated from SCALE fine-group 238- and 999-group libraries, and the final broad-group library was created from Evaluated Nuclear Data File/B Version ENDF/B-VII Release 0 cross-section evaluations using new ORNL methodologies with AMPX, SCALE, and other codes. Furthermore, intermediate resonance (IR) methods were applied to the HTGR broadgroup library, and lambda factors and f-factors were incorporated into the library s nuclear data files. A new version of the SCALE BONAMI module named BONAMI-IR was developed to process the IR data in the new library and, thus, eliminate the need for the CENTRM/PMC modules for resonance selfshielding. This report documents the development of the HTGR broad-group nuclear data library and the results of test and benchmark calculations using the new library with SCALE. The 81-group library is shown to model HTGR cases with similar accuracy to the SCALE 238-group library but with significantly faster computational times due to the reduced number of energy groups and the use of BONAMI-IR instead of BONAMI/CENTRM/PMC for resonance self-shielding calculations.

  20. Assessing high affinity binding to HLA-DQ2.5 by a novel peptide library based approach

    DEFF Research Database (Denmark)

    Jüse, Ulrike; Arntzen, Magnus; Højrup, Peter

    2011-01-01

    Here we report on a novel peptide library based method for HLA class II binding motif identification. The approach is based on water soluble HLA class II molecules and soluble dedicated peptide libraries. A high number of different synthetic peptides are competing to interact with a limited amount...... library. The eluted sequences fit very well with the previously described HLA-DQ2.5 peptide binding motif. This novel method, limited by library complexity and sensitivity of mass spectrometry, allows the analysis of several thousand synthetic sequences concomitantly in a simple water soluble format....

  1. Computing in high energy physics

    International Nuclear Information System (INIS)

    Watase, Yoshiyuki

    1991-01-01

    The increasingly important role played by computing and computers in high energy physics is displayed in the 'Computing in High Energy Physics' series of conferences, bringing together experts in different aspects of computing - physicists, computer scientists, and vendors

  2. Dimensioning storage and computing clusters for efficient High Throughput Computing

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    Scientific experiments are producing huge amounts of data, and they continue increasing the size of their datasets and the total volume of data. These data are then processed by researchers belonging to large scientific collaborations, with the Large Hadron Collider being a good example. The focal point of Scientific Data Centres has shifted from coping efficiently with PetaByte scale storage to deliver quality data processing throughput. The dimensioning of the internal components in High Throughput Computing (HTC) data centers is of crucial importance to cope with all the activities demanded by the experiments, both the online (data acceptance) and the offline (data processing, simulation and user analysis). This requires a precise setup involving disk and tape storage services, a computing cluster and the internal networking to prevent bottlenecks, overloads and undesired slowness that lead to losses cpu cycles and batch jobs failures. In this paper we point out relevant features for running a successful s...

  3. Dimensioning storage and computing clusters for efficient high throughput computing

    International Nuclear Information System (INIS)

    Accion, E; Bria, A; Bernabeu, G; Caubet, M; Delfino, M; Espinal, X; Merino, G; Lopez, F; Martinez, F; Planas, E

    2012-01-01

    Scientific experiments are producing huge amounts of data, and the size of their datasets and total volume of data continues increasing. These data are then processed by researchers belonging to large scientific collaborations, with the Large Hadron Collider being a good example. The focal point of scientific data centers has shifted from efficiently coping with PetaByte scale storage to deliver quality data processing throughput. The dimensioning of the internal components in High Throughput Computing (HTC) data centers is of crucial importance to cope with all the activities demanded by the experiments, both the online (data acceptance) and the offline (data processing, simulation and user analysis). This requires a precise setup involving disk and tape storage services, a computing cluster and the internal networking to prevent bottlenecks, overloads and undesired slowness that lead to losses cpu cycles and batch jobs failures. In this paper we point out relevant features for running a successful data storage and processing service in an intensive HTC environment.

  4. Description of the digital formats of electronic documents used in digital library

    International Nuclear Information System (INIS)

    Galabova, Sevdalina; Trencheva, Tereza; Trenchev, Ivan

    2009-01-01

    A digital library is a library where collections are stored in electronic formats and are easy to be accessed by computers. The digital content may be stored by local machine, or accessed remotely via computer networks. A digital library is a type of information retrieval system. In this paper we will present a few digital format as pdf, djvu and etc. It will be denote the specific of the formats, their history and licenses policy. We will discuss different similarities between these formats.. Keywords: digital library, electronic formats, pdf, djvu

  5. LLNL nuclear data libraries used for fusion calculations

    International Nuclear Information System (INIS)

    Howerton, R.J.

    1984-01-01

    The Physical Data Group of the Computational Physics Division of the Lawrence Livermore National Laboratory has as its principal responsibility the development and maintenance of those data that are related to nuclear reaction processes and are needed for Laboratory programs. Among these are the Magnetic Fusion Energy and the Inertial Confinement Fusion programs. To this end, we have developed and maintain a collection of data files or libraries. These include: files of experimental data of neutron induced reactions; an annotated bibliography of literature related to charged particle induced reactions with light nuclei; and four main libraries of evaluated data. We also maintain files of calculational constants developed from the evaluated libraries for use by Laboratory computer codes. The data used for fusion calculations are usually these calculational constants, but since they are derived by prescribed manipulation of evaluated data this discussion will describe the evaluated libraries

  6. Evaluation of academic library collection using a check-list method

    Directory of Open Access Journals (Sweden)

    Kornelija Petr Balog

    2015-04-01

    Full Text Available The purpose of this paper is to evaluate the quality of the ILS library collection of the Faculty of Humanities and Social Sciences (FHSS in Osijek, Croatia and its congruence with the curriculum. The quality of the collection is measured using the check-list method. The required and optional reading lists of the Department of Information Sciences at the FHSS (academic year 2011/2012 are used as standard lists that the library holdings are compared to. The results found that the library does not have 30.8 per cent of the titles on the reading lists. The remaining 33.9 per cent of the titles are accessible in the library, 28.5 per cent are free electronic resources, and 6.8 per cent of titles are accessible for students through the Department’s Moodle, Learning Management System. The study provides data about the titles available and not available in the FHSS library. However, it does not differentiate between the titles on the required and optional reading lists. This study provides the FHSS librarians with the list of titles that should be obtained in the near future. In Croatia, very few papers on collection assessment have been published so far, and this is the first study about the quality of a library collection at the University of Osijek. The paper attempts to fill that gap and contribute to a deeper understanding of the quality of library collections in the Croatian academic setting.

  7. Resource Provisions of a High School Library Collection

    Science.gov (United States)

    Collins, Karla B.; Doll, Carol A.

    2012-01-01

    The mission of the school library "is to ensure students and staff are effective users of ideas and information" (AASL 2009, 8). The school library collection should, therefore, support instruction throughout the school. However, teachers do not always understand the potential value of the resources available. This research explored…

  8. Extracting DNA from 'jaws': High yield and quality from archived tiger shark (Galeocerdo cuvier) skeletal material

    DEFF Research Database (Denmark)

    Eg Nielsen, Einar; Morgan, J. A T; Maher, S. L.

    2017-01-01

    of tiger sharks (Galeocerdo cuvier). Protocols were compared for DNA yield and quality using a qPCR approach. For jaw swarf, all methods provided relatively high DNA yield and quality, while large differences in yield between protocols were observed for vertebrae. Similar results were obtained from samples...... observed, likely reflecting different preparation and storage methods for the trophies. Trial sequencing of DNA capture genomic libraries using 20 000 baits revealed that a significant proportion of captured sequences were derived from tiger sharks. This study demonstrates that archived shark jaws...

  9. The Profiles in Science Digital Library: Behind the Scenes.

    Science.gov (United States)

    Gallagher, Marie E; Moffatt, Christie

    2012-01-01

    This demonstration shows the Profiles in Science ® digital library. Profiles in Science contains digitized selections from the personal manuscript collections of prominent biomedical researchers, medical practitioners, and those fostering science and health. The Profiles in Science Web site is the delivery mechanism for content derived from the digital library system. The system is designed according to our basic principles for digital library development [1]. The digital library includes the rules and software used for digitizing items, creating and editing database records and performing quality control as well as serving the digital content to the public. Among the types of data managed by the digital library are detailed item-level, collection-level and cross-collection metadata, digitized photographs, papers, audio clips, movies, born-digital electronic files, optical character recognized (OCR) text, and annotations (see Figure 1). The digital library also tracks the status of each item, including digitization quality, sensitivity of content, and copyright. Only items satisfying all required criteria are released to the public through the World Wide Web. External factors have influenced all aspects of the digital library's infrastructure.

  10. Joint-Use Libraries

    Science.gov (United States)

    Casstevens, Susan

    2017-01-01

    The joint-use library is a place where people of all ages, interests, and income levels can find items of interest at no personal cost. The mission of A. H. Meadows Public and High School Library in Midlothian, Texas, is to offer what other public libraries provide: educational and entertainment resources to a community. Yet, the staff also wants…

  11. Curatr: a web application for creating, curating and sharing a mass spectral library.

    Science.gov (United States)

    Palmer, Andrew; Phapale, Prasad; Fay, Dominik; Alexandrov, Theodore

    2018-04-15

    We have developed a web application curatr for the rapid generation of high quality mass spectral fragmentation libraries from liquid-chromatography mass spectrometry datasets. Curatr handles datasets from single or multiplexed standards and extracts chromatographic profiles and potential fragmentation spectra for multiple adducts. An intuitive interface helps users to select high quality spectra that are stored along with searchable molecular information, the providence of each standard and experimental metadata. Curatr supports exports to several standard formats for use with third party software or submission to repositories. We demonstrate the use of curatr to generate the EMBL Metabolomics Core Facility spectral library http://curatr.mcf.embl.de. Source code and example data are at http://github.com/alexandrovteam/curatr/. palmer@embl.de. Supplementary data are available at Bioinformatics online.

  12. The relationship between computer games and quality of life in adolescents

    Science.gov (United States)

    Dolatabadi, Nayereh Kasiri; Eslami, Ahmad Ali; Mostafavi, Firooze; Hassanzade, Akbar; Moradi, Azam

    2013-01-01

    Background: Term of doing computer games among teenagers is growing rapidly. This popular phenomenon can cause physical and psychosocial issues in them. Therefore, this study examined the relationship between computer games and quality of life domains in adolescents aging 12-15 years. Materials and Methods: In a cross-sectional study using the 2-stage stratified cluster sampling method, 444 male and female students in Borkhar were selected. The data collection tool consisted of 1) World Health Organization Quality Of Life – BREF questionnaire and 2) personal information questionnaire. The data were analyzed by Pearson correlation, Spearman correlation, chi-square, independent t-tests and analysis of covariance. Findings: The total mean score of quality of life in students was 67.11±13.34. The results showed a significant relationship between the age of starting to play games and the overall quality of life score and its fourdomains (range r=–0.13 to –0.18). The mean of overall quality of life score in computer game users was 68.27±13.03 while it was 64.81±13.69 among those who did not play computer games and the difference was significant (P=0.01). There were significant differences in environmental and mental health domains between the two groups (Pcomputer games. Conclusion: Playing computer games for a short time under parental supervision can have positive effects on quality of life in adolescents. However, spending long hours for playing computer games may have negative long-term effects. PMID:24083270

  13. Computer and Network Security in Small Libraries: A Guide for Planning.

    Science.gov (United States)

    Williams, Robert L.

    This manual is intended to provide a free resource on essential network security concepts for non-technical managers of small libraries. Managers of other small nonprofit or community organizations will also benefit from it. An introduction defines network security; outlines three goals of network security; discusses why a library should be…

  14. Daily quality control in computed radiography mammography using the manufacturer phantom

    International Nuclear Information System (INIS)

    Jakubiak, Rosangela R.; Messias, Pricila C.; Oliveira, Carlla M.

    2011-01-01

    The quality control (QC) in mammography system involves a large amount of test tools, which implies a large space for storage and a high number of exposure. This work describes a QC system using a phantom, Fuji Computed Radiography (FCR) One Shot Phantom M Plus, that evaluates several parameters with just one exposure. The software offers tests with annual, semi-annual, quarterly, weekly and daily periodicity, and analyzes the conformities of the mammography equipment, image plate and cassettes. Because of the high number of tests, it was evaluated the daily test only for seven months in two mammography equipment. The test, through the software and its image, allows the analysis of ten parameters in QC. The evaluation of these parameters was realized by the average of the values provided by the software. Only one of the evaluated items showed not conformity, but this was observed and the necessary corrections were realized. The monitoring of use of FCR Mammography QC software with the FCR One Shot Phantom M Plus was realized and through this we could investigate that the quality program provided by the system is appropriate for the radiology services that has the Fuji Computed Radiography system. (author)

  15. Security and Crime Challenges in Academic Libraries in Nigeria

    African Journals Online (AJOL)

    Security and prevention of crimes in academic libraries is essential library duties .... From the earliest time to the present, librarians are bothered on how to ensure the protection ..... OPAC (Online Public Access Catalogue) computer system etc.

  16. Library fingerprints: a novel approach to the screening of virtual libraries.

    Science.gov (United States)

    Klon, Anthony E; Diller, David J

    2007-01-01

    We propose a novel method to prioritize libraries for combinatorial synthesis and high-throughput screening that assesses the viability of a particular library on the basis of the aggregate physical-chemical properties of the compounds using a naïve Bayesian classifier. This approach prioritizes collections of related compounds according to the aggregate values of their physical-chemical parameters in contrast to single-compound screening. The method is also shown to be useful in screening existing noncombinatorial libraries when the compounds in these libraries have been previously clustered according to their molecular graphs. We show that the method used here is comparable or superior to the single-compound virtual screening of combinatorial libraries and noncombinatorial libraries and is superior to the pairwise Tanimoto similarity searching of a collection of combinatorial libraries.

  17. 7th International Workshop on Parallel Tools for High Performance Computing

    CERN Document Server

    Gracia, José; Nagel, Wolfgang; Resch, Michael

    2014-01-01

    Current advances in High Performance Computing (HPC) increasingly impact efficient software development workflows. Programmers for HPC applications need to consider trends such as increased core counts, multiple levels of parallelism, reduced memory per core, and I/O system challenges in order to derive well performing and highly scalable codes. At the same time, the increasing complexity adds further sources of program defects. While novel programming paradigms and advanced system libraries provide solutions for some of these challenges, appropriate supporting tools are indispensable. Such tools aid application developers in debugging, performance analysis, or code optimization and therefore make a major contribution to the development of robust and efficient parallel software. This book introduces a selection of the tools presented and discussed at the 7th International Parallel Tools Workshop, held in Dresden, Germany, September 3-4, 2013.  

  18. An in silico MS/MS library for automatic annotation of novel FAHFA lipids.

    Science.gov (United States)

    Ma, Yan; Kind, Tobias; Vaniya, Arpana; Gennity, Ingrid; Fahrmann, Johannes F; Fiehn, Oliver

    2015-01-01

    A new lipid class named 'fatty acid esters of hydroxyl fatty acids' (FAHFA) was recently discovered in mammalian adipose tissue and in blood plasma and some FAHFAs were found to be associated with type 2 diabetes. To facilitate the automatic annotation of FAHFAs in biological specimens, a tandem mass spectra (MS/MS) library is needed. Due to the limitation of the commercial available standard compounds, we proposed building an in silico MS/MS library to extend the coverage of molecules. We developed a computer-generated library with 3267 tandem mass spectra (MS/MS) for 1089 FAHFA species. FAHFA spectra were generated based on authentic standards with negative mode electrospray ionization and 10, 20, and 40 V collision induced dissociation at 4 spectra/s as used in in ultra-high performance liquid chromatography-QTOF mass spectrometry studies. However, positional information of the hydroxyl group is only obtained either at lower QTOF spectra acquisition rates of 1 spectrum/s or at the MS(3) level in ion trap instruments. Therefore, an additional set of 4290 fragment-rich MS/MS spectra was created to enable distinguishing positional FAHFA isomers. The library was generated based on ion fragmentations and ion intensities of FAHFA external reference standards, developing a heuristic model for fragmentation rules and extending these rules to large swaths of computer-generated structures of FAHFAs with varying chain lengths, degrees of unsaturation and hydroxyl group positions. Subsequently, we validated the new in silico library by discovering several new FAHFA species in egg yolk, showing that this library enables high-throughput screening of FAHFA lipids in various biological matrices. The developed library and templates are freely available for commercial or noncommercial use at http://fiehnlab.ucdavis.edu/staff/yanma/fahfa-lipid-library. This in silico MS/MS library allows users to annotate FAHFAs from accurate mass tandem mass spectra in an easy and fast manner

  19. Optimization and quality control of genome-wide Hi-C library preparation.

    Science.gov (United States)

    Zhang, Xiang-Yuan; He, Chao; Ye, Bing-Yu; Xie, De-Jian; Shi, Ming-Lei; Zhang, Yan; Shen, Wen-Long; Li, Ping; Zhao, Zhi-Hu

    2017-09-20

    Highest-throughput chromosome conformation capture (Hi-C) is one of the key assays for genome- wide chromatin interaction studies. It is a time-consuming process that involves many steps and many different kinds of reagents, consumables, and equipments. At present, the reproducibility is unsatisfactory. By optimizing the key steps of the Hi-C experiment, such as crosslinking, pretreatment of digestion, inactivation of restriction enzyme, and in situ ligation etc., we established a robust Hi-C procedure and prepared two biological replicates of Hi-C libraries from the GM12878 cells. After preliminary quality control by Sanger sequencing, the two replicates were high-throughput sequenced. The bioinformatics analysis of the raw sequencing data revealed the mapping-ability and pair-mate rate of the raw data were around 90% and 72%, respectively. Additionally, after removal of self-circular ligations and dangling-end products, more than 96% of the valid pairs were reached. Genome-wide interactome profiling shows clear topological associated domains (TADs), which is consistent with previous reports. Further correlation analysis showed that the two biological replicates strongly correlate with each other in terms of both bin coverage and all bin pairs. All these results indicated that the optimized Hi-C procedure is robust and stable, which will be very helpful for the wide applications of the Hi-C assay.

  20. Flight Software Math Library

    Science.gov (United States)

    McComas, David

    2013-01-01

    The flight software (FSW) math library is a collection of reusable math components that provides typical math utilities required by spacecraft flight software. These utilities are intended to increase flight software quality reusability and maintainability by providing a set of consistent, well-documented, and tested math utilities. This library only has dependencies on ANSI C, so it is easily ported. Prior to this library, each mission typically created its own math utilities using ideas/code from previous missions. Part of the reason for this is that math libraries can be written with different strategies in areas like error handling, parameters orders, naming conventions, etc. Changing the utilities for each mission introduces risks and costs. The obvious risks and costs are that the utilities must be coded and revalidated. The hidden risks and costs arise in miscommunication between engineers. These utilities must be understood by both the flight software engineers and other subsystem engineers (primarily guidance navigation and control). The FSW math library is part of a larger goal to produce a library of reusable Guidance Navigation and Control (GN&C) FSW components. A GN&C FSW library cannot be created unless a standardized math basis is created. This library solves the standardization problem by defining a common feature set and establishing policies for the library s design. This allows the libraries to be maintained with the same strategy used in its initial development, which supports a library of reusable GN&C FSW components. The FSW math library is written for an embedded software environment in C. This places restrictions on the language features that can be used by the library. Another advantage of the FSW math library is that it can be used in the FSW as well as other environments like the GN&C analyst s simulators. This helps communication between the teams because they can use the same utilities with the same feature set and syntax.

  1. High-efficiency photorealistic computer-generated holograms based on the backward ray-tracing technique

    Science.gov (United States)

    Wang, Yuan; Chen, Zhidong; Sang, Xinzhu; Li, Hui; Zhao, Linmin

    2018-03-01

    Holographic displays can provide the complete optical wave field of a three-dimensional (3D) scene, including the depth perception. However, it often takes a long computation time to produce traditional computer-generated holograms (CGHs) without more complex and photorealistic rendering. The backward ray-tracing technique is able to render photorealistic high-quality images, which noticeably reduce the computation time achieved from the high-degree parallelism. Here, a high-efficiency photorealistic computer-generated hologram method is presented based on the ray-tracing technique. Rays are parallelly launched and traced under different illuminations and circumstances. Experimental results demonstrate the effectiveness of the proposed method. Compared with the traditional point cloud CGH, the computation time is decreased to 24 s to reconstruct a 3D object of 100 ×100 rays with continuous depth change.

  2. Student project of optical system analysis API-library development

    Science.gov (United States)

    Ivanova, Tatiana; Zhukova, Tatiana; Dantcaranov, Ruslan; Romanova, Maria; Zhadin, Alexander; Ivanov, Vyacheslav; Kalinkina, Olga

    2017-08-01

    In the paper API-library software developed by students of Applied and Computer Optics Department (ITMO University) for optical system design is presented. The library performs paraxial and real ray tracing, calculates 3d order (Seidel) aberration and real ray aberration of axis and non-axis beams (wave, lateral, longitudinal, coma, distortion etc.) and finally, approximate wave aberration by Zernike polynomials. Real aperture can be calculated by considering of real rays tracing failure on each surface. So far we assume optical system is centered, with spherical or 2d order aspherical surfaces. Optical glasses can be set directly by refraction index or by dispersion coefficients. The library can be used for education or research purposes in optical system design area. It provides ready to use software functions for optical system simulation and analysis that developer can simply plug into their software development for different purposes, for example for some specific synthesis tasks or investigation of new optimization modes. In the paper we present an example of using the library for development of cemented doublet synthesis software based on Slusarev's methodology. The library is used in optical system optimization recipes course for deep studying of optimization model and its application for optical system design. Development of such software is an excellent experience for students and help to understanding optical image modeling and quality analysis. This development is organized as student group joint project. We try to organize it as a group in real research and development project, so each student has his own role in the project and then use whole library functionality in his own master or bachelor thesis. Working in such group gives students useful experience and opportunity to work as research and development engineer of scientific software in the future.

  3. Experience of Google's latest deep learning library, TensorFlow, in a large-scale WLCG cluster

    Energy Technology Data Exchange (ETDEWEB)

    Kawamura, Gen; Smith, Joshua Wyatt; Quadt, Arnulf [II. Physikalisches Institut, Georg-August-Universitaet Goettingen (Germany)

    2016-07-01

    The researchers at the Google Brain team released their second generation's Deep Learning library, TensorFlow, as an open-source package under the Apache 2.0 license in November, 2015. Google has already deployed the first generation's library using DistBlief in various systems such as Google Search, advertising systems, speech recognition systems, Google Images, Google Maps, Street View, Google Translate and many other latest products. In addition, many researchers in high energy physics have recently started to understand and use Deep Learning algorithms in their own research and analysis. We conceive a first use-case scenario of TensorFlow to create the Deep Learning models from high-dimensional inputs like physics analysis data in a large-scale WLCG computing cluster. TensorFlow carries out computations using a dataflow model and graph structure onto a wide variety of different hardware platforms and systems, such as many CPU architectures, GPUs and smartphone platforms. Having a single library that can distribute the computations to create a model to the various platforms and systems would significantly simplify the use of Deep Learning algorithms in high energy physics. We deploy TensorFlow with the Docker container environments and present the first use in our grid system.

  4. Library construction and evaluation for site saturation mutagenesis.

    Science.gov (United States)

    Sullivan, Bradford; Walton, Adam Z; Stewart, Jon D

    2013-06-10

    We developed a method for creating and evaluating site-saturation libraries that consistently yields an average of 27.4±3.0 codons of the 32 possible within a pool of 95 transformants. This was verified by sequencing 95 members from 11 independent libraries within the gene encoding alkene reductase OYE 2.6 from Pichia stipitis. Correct PCR primer design as well as a variety of factors that increase transformation efficiency were critical contributors to the method's overall success. We also developed a quantitative analysis of library quality (Q-values) that defines library degeneracy. Q-values can be calculated from standard fluorescence sequencing data (capillary electropherograms) and the degeneracy predicted from an early stage of library construction (pooled plasmids from the initial transformation) closely matched that observed after ca. 1000 library members were sequenced. Based on this experience, we suggest that this analysis can be a useful guide when applying our optimized protocol to new systems, allowing one to focus only on good-quality libraries and reject substandard libraries at an early stage. This advantage is particularly important when lower-throughput screening techniques such as chiral-phase GC must be employed to identify protein variants with desirable properties, e.g., altered stereoselectivities or when multiple codons are targeted for simultaneous randomization. Copyright © 2013 Elsevier Inc. All rights reserved.

  5. Israeli Special Libraries

    Science.gov (United States)

    Foster, Barbara

    1974-01-01

    Israel is sprinkled with a noteworthy representation of special libraries which run the gamut from modest kibbutz efforts to highly technical scientific and humanities libraries. A few examples are discussed here. (Author/CH)

  6. The JPL Library information retrieval system

    Science.gov (United States)

    Walsh, J.

    1975-01-01

    The development, capabilities, and products of the computer-based retrieval system of the Jet Propulsion Laboratory Library are described. The system handles books and documents, produces a book catalog, and provides a machine search capability. Programs and documentation are available to the public through NASA's computer software dissemination program.

  7. Integrated evolutionary computation neural network quality controller for automated systems

    Energy Technology Data Exchange (ETDEWEB)

    Patro, S.; Kolarik, W.J. [Texas Tech Univ., Lubbock, TX (United States). Dept. of Industrial Engineering

    1999-06-01

    With increasing competition in the global market, more and more stringent quality standards and specifications are being demands at lower costs. Manufacturing applications of computing power are becoming more common. The application of neural networks to identification and control of dynamic processes has been discussed. The limitations of using neural networks for control purposes has been pointed out and a different technique, evolutionary computation, has been discussed. The results of identifying and controlling an unstable, dynamic process using evolutionary computation methods has been presented. A framework for an integrated system, using both neural networks and evolutionary computation, has been proposed to identify the process and then control the product quality, in a dynamic, multivariable system, in real-time.

  8. Development of Calcomp compatible interface library 'piflib' on X Window System

    International Nuclear Information System (INIS)

    Tanabe, Hidenobu; Yokokawa, Mitsuo; Onuma, Yoshio.

    1993-05-01

    Graphics processing at JAERI mainly has been executed on mainframe computers with Calcomp compatible graphics libraries. With spread of engineering workstations (EWS), it is important that those enormous graphics softwares be able to be carried out on EWS. The Calcomp compatible interface library 'piflib' has been developed on the X Window System, which is the most popular window environments on EWS. In this report, specifications of the library 'piflib' and its usages are presented. The cooperative processing with mainframe computers is also described. (author)

  9. Computational analysis of neutronic parameters for TRIGA Mark-II research reactor using evaluated nuclear data libraries

    International Nuclear Information System (INIS)

    Uddin, M.N.; Sarker, M.M.; Khan, M.J.H.; Islam, S.M.A.

    2010-01-01

    The aim of this study is to analyze the neutronic parameters of TRIGA Mark-II research reactor using the chain of NJOY-WIMS-CITATION computer codes based on evaluated nuclear data libraries CENDL-2.2 and JEFF-3.1.1. The nuclear data processing code NJOY99.0 has been employed to generate the 69 group WIMS library for the isotopes of TRIGA core. The cell code WIMSD-5B was used to generate the cross sections in CITATION format and then 3-dimensional diffusion code CITTATION was used to calculate the neutronic parameters of the TRIGA Mark-II research reactor. All the analyses were performed using the 7-group macroscopic cross section library. The CITATION test-runs using different cross section sets based on different models applied in WIMS calculations have shown a strong influence of those models on the final integral parameters. Some of the cells were specially treated with PRIZE options available in WIMSD-5B to take into account the fine structure of the flux gradient in the fuel-reflector interface region. It was observed that two basic parameters, the effective multiplication factor, k eff and the thermal neutron flux, were in good agreement among the calculated results with each other as well as the measured values. The maximum power densities at the hot spot were 1.0446E02 W/cc and 1.0426E02 W/cc for the libraries CENDL-2.2 and JEFF-3.1.1 respectively. The calculated total peaking factors 5.793 and 5.745 were compared to the original SAR value of 5.6325 as well as MCNP result. Consequently, this analysis will be helpful to enhance the neutronic calculations and also be used for the further thermal-hydraulics study of the TRIGA core.

  10. Understanding Patterns of Library Use Among Undergraduate Students from Different Disciplines

    Directory of Open Access Journals (Sweden)

    Ellen Collins

    2014-09-01

    Full Text Available Objective – To test whether routinely-generated library usage data could be linked with information about students to understand patterns of library use among students from different disciplines at the University of Huddersfield. This information is important for librarians seeking to demonstrate the value of the library, and to ensure that they are providing services which meet user needs. The study seeks to join two strands of library user research which until now have been kept rather separate – an interest in disciplinary differences in usage, and a methodology which involves large-scale routinely-generated data. Methods – The study uses anonymized data about individual students derived from two sources: routinely-generated data on various dimensions of physical and electronic library resource usage, and information from the student registry on the course studied by each student. Courses were aggregated at a subject and then disciplinary level. Kruskal-Wallis and Mann Whitney tests were used to identify statistically significant differences between the high-level disciplinary groups, and within each disciplinary group at the subject level. Results – The study identifies a number of statistically significant differences on various dimensions of usage between both high-level disciplinary groupings and lower subject-level groupings. In some cases, differences are not the same as those observed in earlier studies, reflecting distinctive usage patterns and differences in the way that disciplines or subjects are defined and organised. While music students at Huddersfield are heavy library users within the arts subject-level grouping arts students use library resources less than those in social science disciplines, contradicting findings from studies at other institutions, Computing and engineering students were relatively similar, although computing students were more likely to download PDFs, and engineering students were more likely to

  11. Scientific Library Offers New Training Options | Poster

    Science.gov (United States)

    The Scientific Library is expanding its current training opportunities by offering webinars, allowing employees to take advantage of trainings from the comfort of their own offices. Due to the nature of their work, some employees find it inconvenient to attend in-person training classes; others simply prefer to use their own computers. The Scientific Library has been

  12. Which educational role can Libraries play in a University learning environment?

    Directory of Open Access Journals (Sweden)

    Alexandra Angeletaki

    2010-07-01

    • • Classroom instruction and observation of skills and technology application proficiencies • Face to face conversation with the students and the faculty members involved in the program. • Web-organised library survey. Project coordinator: Alexandra Angeletaki, University library of Trondheim Email: alexandra.angeletaki@ub.ntnu.no Description: The traditional way of assessing library service quality is to measure the numbers of users and resource materials purchased each year by the library users (Quantitative. But can this type of information help the Library to establish itself as an important educational component, meeting its role in the digital information world with a high academic standard that can influence the research outcome of the faculty it serves. What will the future Library environment be, if one takes in consideration the technological change of the library in place to the library in “Space”? The aim should be to maximise not only the services in numbers as they are easy numeric figures to measure, but in quality that meets the academic requirements of a research Library with educational programs exerting influence on the learning experience of its users. It is consequent then that such a measurement will have to be empowered in order to increase academic literacy and research competence. The University Library of Trondheim has been working the last 2 years in collecting data about the learning process of archaeology students trained in Information literacy workshops in collaboration with the Institute of Archaeology from the University of Trondheim. In 2010 our department introduced the use of reading devices for first year students of two different curriculums Archaeology and Chemistry. Three reading devices were filled up with the texts of the subjects taught and the students that were chosen to participate in the program will be giving at the end of the Spring semester 2010 an account of the use of the reading devices. The overall

  13. Research on Current Trends and Developments in Library Automation Systems.

    Science.gov (United States)

    Clyde, Laurel A.

    A research project was undertaken in 1992 in order to identify, describe, and analyze current trends and developments in library automation systems and library software. The starting point is work conducted in 1988/1989 that formed the foundation of the book "Computer Software for School Libraries: A Directory." A 1992 survey of software…

  14. DendroPy: a Python library for phylogenetic computing.

    Science.gov (United States)

    Sukumaran, Jeet; Holder, Mark T

    2010-06-15

    DendroPy is a cross-platform library for the Python programming language that provides for object-oriented reading, writing, simulation and manipulation of phylogenetic data, with an emphasis on phylogenetic tree operations. DendroPy uses a splits-hash mapping to perform rapid calculations of tree distances, similarities and shape under various metrics. It contains rich simulation routines to generate trees under a number of different phylogenetic and coalescent models. DendroPy's data simulation and manipulation facilities, in conjunction with its support of a broad range of phylogenetic data formats (NEXUS, Newick, PHYLIP, FASTA, NeXML, etc.), allow it to serve a useful role in various phyloinformatics and phylogeographic pipelines. The stable release of the library is available for download and automated installation through the Python Package Index site (http://pypi.python.org/pypi/DendroPy), while the active development source code repository is available to the public from GitHub (http://github.com/jeetsukumaran/DendroPy).

  15. PLANNING APPLICATION OF WEB 2.0 FOR ORGANIZATIONAL LEARNING IN UNIVERSITAS PENDIDIKAN INDONESIA LIBRARY

    Directory of Open Access Journals (Sweden)

    Santi Santika

    2017-07-01

    Full Text Available Library of Universitas Pendidikan Indonesia (UPI has a quality policy commitment to continuous improvement in every area and process. It can be achieved by continuously optimizing organizational learning. Web 2.0 is a media application that can help the organizational learning process because it has the characteristics of read and write, as well as having the flexibility of time use, but the application must be in accordance with the culture and character of the organization. Therefore, this study aimed to find out the Web 2.0 application that can be applied to the organizational learning in the Library of UPI. The method used is a mixed method qualitative and quantitative approach. Research stage refers to the stage of planning and support phases of Web 2.0 Tools Implementation Model. The results showed that the application of Web 2.0 can be applied to the organizational learning in the Library UPI. It refers to the tendency of organizational culture Library of UPI that is good and tendency of HR Library UPI attitude against the use of the Internet and computers are very good. Web 2.0 applications that can be used by UPI library are blogs, online forums, and wiki as a primary tools. Facebook, Youtube, chat application, twitter and Instagram as a supporting tools.

  16. Training for QSC: How McDonald's Makes Library Managers.

    Science.gov (United States)

    Aycock, Anthony

    2001-01-01

    Written by a former McDonald's manager who earned a master's degree in library science, this article describes how "QSC"--quality, services, and cleanliness--can be applied to libraries in the same way it is used by restaurants. Argues that libraries are simply businesses that cater to customers. (Contains 42 references.) (NB)

  17. SLATEC-4.1, Subroutine Library for Solution of Mathematical Problems

    International Nuclear Information System (INIS)

    Boland, W.R.

    1999-01-01

    1 - Description of problem or function: SLATEC4.1 is a large collection of FORTRAN mathematical subprograms brought together in a joint effort by the Air Force Phillips Laboratory, Lawrence Livermore National Laboratory, Los Alamos National Laboratory, Magnetic Fusion Energy Computing Center, National Institute of Standards and Technology, Sandia National Laboratories (Albuquerque and Livermore), and Oak Ridge National Laboratory. SLATEC is characterized by portability, good numerical technology, good documentation, robustness, and quality assurance. The library can be divided into the following subsections following the lines of the GAMS classification system: Error Analysis, Elementary and Special Functions, Elementary Vector Operations, Solutions of Systems of Linear Equations, Eigen analysis, QR Decomposition, Singular Value Decomposition, Overdetermined or Underdetermined Systems, Interpolation, Solution of Nonlinear Equations, Optimization, Quadrature, Ordinary Differential Equations, Partial Differential Equations, Fast Fourier Transforms, Approximation, Pseudo-random Number Generation, Sorting, Machine Constants, and Diagnostics and Error Handling. 2 - Method of solution: This information is provided by comments within the individual library subroutines

  18. Benchmarking and Its Relevance to the Library and Information Sector. Interim Findings of "Best Practice Benchmarking in the Library and Information Sector," a British Library Research and Development Department Project.

    Science.gov (United States)

    Kinnell, Margaret; Garrod, Penny

    This British Library Research and Development Department study assesses current activities and attitudes toward quality management in library and information services (LIS) in the academic sector as well as the commercial/industrial sector. Definitions and types of benchmarking are described, and the relevance of benchmarking to LIS is evaluated.…

  19. Emotional Branding as an Effort to Improve the Quality of Library Service in Order to Increase the Use of Library

    Directory of Open Access Journals (Sweden)

    Neneng Komariah

    2016-12-01

    Full Text Available This paper is a study of implementation of emotional branding in library services.    Emotional branding is a new paradigm in marketing world. It means create emotional nuance and its objective is to create an emotional nearness between a brand and its consumer. The library as an institution which provide information services may implement emotional branding concept in creating emotional nuance which will build emotional nearness between the library and its users. It is hope that the emotional nearness  will create users’ satisfaction and  loyality, so the use of library will  increase. Emotional branding in library can be created through working performance of librarian who always try to create and maintain good relationships with users. Some strategies could be implemented, those are always ask the user first, ask their name and try to remember it, know information which usually needed, tell them when there is new information which relevant for them, create a situation that librarian always ready to help, show them that the librarian like to discuss with them, be a good listener, the librarian should be wear uniform with name tag, and be a mobile staff and ask the user if they need some helps.  The implementation of emotional branding in library services need a crucial change in mindset of library management in order to provide qualified services which suitable with users’ demand.

  20. Emotional Branding as an Effort to Improve the Quality of Library Service in Order to Increase the Use of Library

    Directory of Open Access Journals (Sweden)

    Neneng Komariah

    2018-01-01

    Full Text Available This paper is a study of implementation of emotional branding in library services.    Emotional branding is a new paradigm in marketing world. It means create emotional nuance and its objective is to create an emotional nearness between a brand and its consumer. The library as an institution which provide information services may implement emotional branding concept in creating emotional nuance which will build emotional nearness between the library and its users. It is hope that the emotional nearness  will create users’ satisfaction and  loyality, so the use of library will  increase. Emotional branding in library can be created through working performance of librarian who always try to create and maintain good relationships with users. Some strategies could be implemented, those are always ask the user first, ask their name and try to remember it, know information which usually needed, tell them when there is new information which relevant for them, create a situation that librarian always ready to help, show them that the librarian like to discuss with them, be a good listener, the librarian should be wear uniform with name tag, and be a mobile staff and ask the user if they need some helps.  The implementation of emotional branding in library services need a crucial change in mindset of library management in order to provide qualified services which suitable with users’ demand.

  1. Computing in high energy physics

    Energy Technology Data Exchange (ETDEWEB)

    Smith, Sarah; Devenish, Robin [Nuclear Physics Laboratory, Oxford University (United Kingdom)

    1989-07-15

    Computing in high energy physics has changed over the years from being something one did on a slide-rule, through early computers, then a necessary evil to the position today where computers permeate all aspects of the subject from control of the apparatus to theoretical lattice gauge calculations. The state of the art, as well as new trends and hopes, were reflected in this year's 'Computing In High Energy Physics' conference held in the dreamy setting of Oxford's spires. The conference aimed to give a comprehensive overview, entailing a heavy schedule of 35 plenary talks plus 48 contributed papers in two afternoons of parallel sessions. In addition to high energy physics computing, a number of papers were given by experts in computing science, in line with the conference's aim – 'to bring together high energy physicists and computer scientists'.

  2. Issues for bringing digital libraries into public use

    Science.gov (United States)

    Flater, David W.; Yesha, Yelena

    1993-01-01

    In much the same way that the field of artificial intelligence produced a cult which fervently believed that computers would soon think like human beings, the existence of electronic books has resurrected the paperless society as a utopian vision to some, an apocalyptic horror to others. In this essay we have attempted to provide realistic notions of what digital libraries are likely to become if they are a popular success. E-books are capable of subsuming most of the media we use today and have the potential for added functionality by being interactive. The environmental impact of having millions more computers will be offset to some degree, perhaps even exceeded, by the fact that televisions, stereos, VCR's, CD players, newspapers, magazines, and books will become part of the computer system or be made redundant. On the whole, large-scale use of digital libraries is likely to be a winning proposition. Whether or not this comes to pass depends on the directions taken by today's researchers and software developers. By involving the public, the effort being put into digital libraries can be leveraged into something which is big enough to make a real change for the better. If digital libraries remain the exclusive property of government, universities, and large research firms, then large parts of the world will remain without digital libraries for years to come, just as they have remained without digital phone service for far too long. If software companies try to scuttle the project by patenting crucial algorithms and using proprietary data formats, all of us will suffer. Let us reverse the errors of the past and create a truly open digital library system.

  3. Systems Analysis, Machineable Circulation Data and Library Users and Non-Users.

    Science.gov (United States)

    Lubans, John, Jr.

    A study to be made with computer-based circulation data of the non-use and use of a large academic library is discussed. A search of the literature reveals that computer-based circulation systems can be, but have not been, utilized to provide data bases for systematic analyses of library users and resources. The data gathered in the circulation…

  4. Damsel: A Data Model Storage Library for Exascale Science

    Energy Technology Data Exchange (ETDEWEB)

    Koziol, Quincey [The HDF Group, Champaign, IL (United States)

    2014-11-26

    The goal of this project is to enable exascale computational science applications to interact conveniently and efficiently with storage through abstractions that match their data models. We will accomplish this through three major activities: (1) identifying major data model motifs in computational science applications and developing representative benchmarks; (2) developing a data model storage library, called Damsel, that supports these motifs, provides efficient storage data layouts, incorporates optimizations to enable exascale operation, and is tolerant to failures; and (3) productizing Damsel and working with computational scientists to encourage adoption of this library by the scientific community.

  5. Damsel - A Data Model Storage Library for Exascale Science

    Energy Technology Data Exchange (ETDEWEB)

    Samatova, Nagiza F

    2014-07-18

    The goal of this project is to enable exascale computational science applications to interact conveniently and efficiently with storage through abstractions that match their data models. We will accomplish this through three major activities: (1) identifying major data model motifs in computational science applications and developing representative benchmarks; (2) developing a data model storage library, called Damsel, that supports these motifs, provides efficient storage data layouts, incorporates optimizations to enable exascale operation, and is tolerant to failures; and (3) productizing Damsel and working with computational scientists to encourage adoption of this library by the scientific community.

  6. Study of college library appealing information system: A case of Longyan University

    Science.gov (United States)

    Liao, Jin-Hui

    2014-10-01

    The complaints from the readers at university libraries mainly focus on the aspects of service attitude, quality of service, reading environment, the management system, etc. Librarians should realize that reader complaints can actually promote the role of the library service and communicate with readers who complain in a friendly manner. In addition, the Longyan University library should establish an internal management system, improve library hardware facilities, improve the quality of librarians and optimize the knowledge structure of librarians, so as to improve the quality of the service for readers and reduce complaints. Based on this point, we have designed an appealing information system in cryptography machine basis, to provide readers online, remote and anonymous complaint functions.

  7. Image quality in coronary computed tomography angiography

    DEFF Research Database (Denmark)

    Precht, Helle; Gerke, Oke; Thygesen, Jesper

    2018-01-01

    Background Computed tomography (CT) technology is rapidly evolving and software solution developed to optimize image quality and/or lower radiation dose. Purpose To investigate the influence of adaptive statistical iterative reconstruction (ASIR) at different radiation doses in coronary CT...

  8. Droplet Digital™ PCR Next-Generation Sequencing Library QC Assay.

    Science.gov (United States)

    Heredia, Nicholas J

    2018-01-01

    Digital PCR is a valuable tool to quantify next-generation sequencing (NGS) libraries precisely and accurately. Accurately quantifying NGS libraries enable accurate loading of the libraries on to the sequencer and thus improve sequencing performance by reducing under and overloading error. Accurate quantification also benefits users by enabling uniform loading of indexed/barcoded libraries which in turn greatly improves sequencing uniformity of the indexed/barcoded samples. The advantages gained by employing the Droplet Digital PCR (ddPCR™) library QC assay includes the precise and accurate quantification in addition to size quality assessment, enabling users to QC their sequencing libraries with confidence.

  9. Analyses of criticality and reactivity for TRACY experiments based on JENDL-3.3 data library

    International Nuclear Information System (INIS)

    Sono, Hiroki; Miyoshi, Yoshinori; Nakajima, Ken

    2003-01-01

    The parameters on criticality and reactivity employed for computational simulations of the TRACY supercritical experiments were analyzed using a recently revised nuclear data library, JENDL-3.3. The parameters based on the JENDL-3.3 library were compared to those based on two former-used libraries, JENDL-3.2 and ENDF/B-VI. In the analyses computational codes, MVP, MCNP version 4C and TWOTRAN, were used. The following conclusions were obtained from the analyses: (1) The computational biases of the effective neutron multiplication factor attributable to the nuclear data libraries and to the computational codes do not depend the TRACY experimental conditions such as fuel conditions. (2) The fractional discrepancies in the kinetic parameters and coefficients of reactivity are within ∼5% between the three libraries. By comparison between calculations and measurements of the parameters, the JENDL-3.3 library is expected to give closer values to the measurements than the JENDL-3.2 and ENDF/B-VI libraries. (3) While the reactivity worth of transient rods expressed in the $ unit shows ∼5% discrepancy between the three libraries according to their respective β eff values, there is little discrepancy in that expressed in the Δk/k unit. (author)

  10. High-level language computer architecture

    CERN Document Server

    Chu, Yaohan

    1975-01-01

    High-Level Language Computer Architecture offers a tutorial on high-level language computer architecture, including von Neumann architecture and syntax-oriented architecture as well as direct and indirect execution architecture. Design concepts of Japanese-language data processing systems are discussed, along with the architecture of stack machines and the SYMBOL computer system. The conceptual design of a direct high-level language processor is also described.Comprised of seven chapters, this book first presents a classification of high-level language computer architecture according to the pr

  11. Design of focused and restrained subsets from extremely large virtual libraries.

    Science.gov (United States)

    Jamois, Eric A; Lin, Chien T; Waldman, Marvin

    2003-11-01

    With the current and ever-growing offering of reagents along with the vast palette of organic reactions, virtual libraries accessible to combinatorial chemists can reach sizes of billions of compounds or more. Extracting practical size subsets for experimentation has remained an essential step in the design of combinatorial libraries. A typical approach to computational library design involves enumeration of structures and properties for the entire virtual library, which may be unpractical for such large libraries. This study describes a new approach termed as on the fly optimization (OTFO) where descriptors are computed as needed within the subset optimization cycle and without intermediate enumeration of structures. Results reported herein highlight the advantages of coupling an ultra-fast descriptor calculation engine to subset optimization capabilities. We also show that enumeration of properties for the entire virtual library may not only be unpractical but also wasteful. Successful design of focused and restrained subsets can be achieved while sampling only a small fraction of the virtual library. We also investigate the stability of the method and compare results obtained from simulated annealing (SA) and genetic algorithms (GA).

  12. Job life cycle management libraries for CMS workflow management projects

    International Nuclear Information System (INIS)

    Lingen, Frank van; Wilkinson, Rick; Evans, Dave; Foulkes, Stephen; Afaq, Anzar; Vaandering, Eric; Ryu, Seangchan

    2010-01-01

    Scientific analysis and simulation requires the processing and generation of millions of data samples. These tasks are often comprised of multiple smaller tasks divided over multiple (computing) sites. This paper discusses the Compact Muon Solenoid (CMS) workflow infrastructure, and specifically the Python based workflow library which is used for so called task lifecycle management. The CMS workflow infrastructure consists of three layers: high level specification of the various tasks based on input/output data sets, life cycle management of task instances derived from the high level specification and execution management. The workflow library is the result of a convergence of three CMS sub projects that respectively deal with scientific analysis, simulation and real time data aggregation from the experiment. This will reduce duplication and hence development and maintenance costs.

  13. Iterative optimization of performance libraries by hierarchical division of codes

    International Nuclear Information System (INIS)

    Donadio, S.

    2007-09-01

    The increasing complexity of hardware features incorporated in modern processors makes high performance code generation very challenging. Library generators such as ATLAS, FFTW and SPIRAL overcome this issue by empirically searching in the space of possible program versions for the one that performs the best. This thesis explores fully automatic solution to adapt a compute-intensive application to the target architecture. By mimicking complex sequences of transformations useful to optimize real codes, we show that generative programming is a practical tool to implement a new hierarchical compilation approach for the generation of high performance code relying on the use of state-of-the-art compilers. As opposed to ATLAS, this approach is not application-dependant but can be applied to fairly generic loop structures. Our approach relies on the decomposition of the original loop nest into simpler kernels. These kernels are much simpler to optimize and furthermore, using such codes makes the performance trade off problem much simpler to express and to solve. Finally, we propose a new approach for the generation of performance libraries based on this decomposition method. We show that our method generates high-performance libraries, in particular for BLAS. (author)

  14. High-performance computing using FPGAs

    CERN Document Server

    Benkrid, Khaled

    2013-01-01

    This book is concerned with the emerging field of High Performance Reconfigurable Computing (HPRC), which aims to harness the high performance and relative low power of reconfigurable hardware–in the form Field Programmable Gate Arrays (FPGAs)–in High Performance Computing (HPC) applications. It presents the latest developments in this field from applications, architecture, and tools and methodologies points of view. We hope that this work will form a reference for existing researchers in the field, and entice new researchers and developers to join the HPRC community.  The book includes:  Thirteen application chapters which present the most important application areas tackled by high performance reconfigurable computers, namely: financial computing, bioinformatics and computational biology, data search and processing, stencil computation e.g. computational fluid dynamics and seismic modeling, cryptanalysis, astronomical N-body simulation, and circuit simulation.     Seven architecture chapters which...

  15. Computing in high energy physics

    International Nuclear Information System (INIS)

    Smith, Sarah; Devenish, Robin

    1989-01-01

    Computing in high energy physics has changed over the years from being something one did on a slide-rule, through early computers, then a necessary evil to the position today where computers permeate all aspects of the subject from control of the apparatus to theoretical lattice gauge calculations. The state of the art, as well as new trends and hopes, were reflected in this year's 'Computing In High Energy Physics' conference held in the dreamy setting of Oxford's spires. The conference aimed to give a comprehensive overview, entailing a heavy schedule of 35 plenary talks plus 48 contributed papers in two afternoons of parallel sessions. In addition to high energy physics computing, a number of papers were given by experts in computing science, in line with the conference's aim – 'to bring together high energy physicists and computer scientists'

  16. Technological Advances Impacting Libraries and Libraries of the Future Kütüphaneleri Etkileyen Teknolojik Yenilikler ve Geleceğin Kütüphaneleri

    Directory of Open Access Journals (Sweden)

    Nazan Özenç Uçak

    1995-03-01

    Full Text Available Developments in libraries which began with the use of computers accelerate with the emergence of new technologies every day. Especially, the widespread use of microcomputers, developments in communication and optic disk technologies, and the use ofartifxcal intelligence and expert systems in libraries have impact upon the library profession. The assortment, quality and speed of library services have increased due to these developments. Along with the technological advances, the organizational structure of , and the management approcah to, libraries changes, too. These ongoing changes bring forth the question of what would happen to the libraries and librarianship of the future. Kütüphanelerde bilgisayarların kullanımı ile başlayan gelişmeler, her geçen gün yeni teknolojilerin ortaya çıkmasıyla giderek büyümektedir, özellikle mikrobilgisayarların geniş kullanım alanı bulması, iletişim teknolojisinin hızlı gelişimi, optik disk teknolojisindeki yenilikler, yapay zeka ve uzman sistemlerden kütüphanecilik alanında yararlanılabilmesi kütüphanecilik mesleğini etkilemektedir. Bu gelişmelere bağlı olarak verilen hizmetlerin çeşidi, niteliği ve hızı artmaktadır. Kütüphanelerin Örgüt yapılarında, yönetim anlayışlarında da değişimler söz konusudur. Bugün yaşanan bu değişimler geleceğin kütüphanelerinin ve kütüphaneciliğinin neolacağı sorularını gündeme getirmektedir.

  17. Fee-based services in sci-tech libraries

    CERN Document Server

    Mount, Ellis

    2013-01-01

    This timely and important book explores how fee-based services have developed in various types of sci-tech libraries. The authoritative contributors focus on the current changing financial aspects of the sci-tech library operation and clarify for the reader how these changes have brought about conditions in which traditional methods of funding are no longer adequate. What new options are open and how they are best being applied in today's sci-tech libraries is fully and clearly explained and illustrated. Topics explored include cost allocation and cost recovery, fees for computer searching, an

  18. Bringing Computational Thinking into the High School Science and Math Classroom

    Science.gov (United States)

    Trouille, Laura; Beheshti, E.; Horn, M.; Jona, K.; Kalogera, V.; Weintrop, D.; Wilensky, U.; University CT-STEM Project, Northwestern; University CenterTalent Development, Northwestern

    2013-01-01

    Computational thinking (for example, the thought processes involved in developing algorithmic solutions to problems that can then be automated for computation) has revolutionized the way we do science. The Next Generation Science Standards require that teachers support their students’ development of computational thinking and computational modeling skills. As a result, there is a very high demand among teachers for quality materials. Astronomy provides an abundance of opportunities to support student development of computational thinking skills. Our group has taken advantage of this to create a series of astronomy-based computational thinking lesson plans for use in typical physics, astronomy, and math high school classrooms. This project is funded by the NSF Computing Education for the 21st Century grant and is jointly led by Northwestern University’s Center for Interdisciplinary Exploration and Research in Astrophysics (CIERA), the Computer Science department, the Learning Sciences department, and the Office of STEM Education Partnerships (OSEP). I will also briefly present the online ‘Astro Adventures’ courses for middle and high school students I have developed through NU’s Center for Talent Development. The online courses take advantage of many of the amazing online astronomy enrichment materials available to the public, including a range of hands-on activities and the ability to take images with the Global Telescope Network. The course culminates with an independent computational research project.

  19. News from the Library

    CERN Multimedia

    CERN Library

    2010-01-01

    Even more books available electronically!   For several years now, the Library has been offering a large collection of electronic books in a wide range of disciplines. The books can be accessed by all CERN users with a Nice account and, like printed books, can be borrowed for a given period. In a few clicks of the mouse, you can leaf through and read books and even print parts of them from your computer. The Library catalogue now comprises a total of more than 10,000 different e-books. The long-awaited electronic versions of O'Reilly book titles are now available: 70 titles have recently been added to the Library's collection and many others will follow in the coming weeks. This collection of books, mainly on IT subjects, is widely used in the development field. Their availability on line is thus a clear bonus. But there's no need for fans of paper versions to worry: the Library will continue to expand its collection of printed books. The two collections exist side by side and even complement ea...

  20. Public library consumer health information pilot project: results of a National Library of Medicine evaluation.

    Science.gov (United States)

    Wood, F B; Lyon, B; Schell, M B; Kitendaugh, P; Cid, V H; Siegel, E R

    2000-10-01

    In October 1998, the National Library of Medicine (NLM) launched a pilot project to learn about the role of public libraries in providing health information to the public and to generate information that would assist NLM and the National Network of Libraries of Medicine (NN/LM) in learning how best to work with public libraries in the future. Three regional medical libraries (RMLs), eight resource libraries, and forty-one public libraries or library systems from nine states and the District of Columbia were selected for participation. The pilot project included an evaluation component that was carried out in parallel with project implementation. The evaluation ran through September 1999. The results of the evaluation indicated that participating public librarians were enthusiastic about the training and information materials provided as part of the project and that many public libraries used the materials and conducted their own outreach to local communities and groups. Most libraries applied the modest funds to purchase additional Internet-accessible computers and/or upgrade their health-reference materials. However, few of the participating public libraries had health information centers (although health information was perceived as a top-ten or top-five topic of interest to patrons). Also, the project generated only minimal usage of NLM's consumer health database, known as MEDLINEplus, from the premises of the monitored libraries (patron usage from home or office locations was not tracked). The evaluation results suggested a balanced follow-up by NLM and the NN/LM, with a few carefully selected national activities, complemented by a package of targeted activities that, as of January 2000, are being planned, developed, or implemented. The results also highlighted the importance of building an evaluation component into projects like this one from the outset, to assure that objectives were met and that evaluative information was available on a timely basis, as was

  1. Quality assurance of computed tomography (CT) scanners

    International Nuclear Information System (INIS)

    Sankaran, A.; Sanu, K.K. . Email : a_sankaran@vsnl.com

    2004-01-01

    This article reviews the present status of research work and development of various test objects, phantoms and detector/instrumentation systems for quality assurance (QA) of computed tomography (CT) scanners, carried out in advanced countries, with emphasis on similar work done in this research centre. CT scanner is a complex equipment and routine quality control procedures are essential to the maintenance of image quality with optimum patient dose. Image quality can be ensured only through correlation between prospective monitoring of system components and tests of overall performance with standard phantoms. CT examinations contribute a large share to the population dose in advanced countries. The unique dosimetry problems in CT necessitate special techniques. This article describes a comprehensive kit developed indigenously for the following QA and type approval tests as well as for research studies on image quality/dosimetry on CT scanners

  2. Cloud-based services for your library a LITA guide

    CERN Document Server

    Mitchell, Erik T

    2013-01-01

    By exploring specific examples of cloud computing and virtualization, this book allows libraries considering cloud computing to start their exploration of these systems with a more informed perspective.

  3. The State of ICT in University Libraries in South Eastern Nigeria ...

    African Journals Online (AJOL)

    Many of the Libraries especially the state funded have no Internet, or LAN workstation but each of the library can boast of standalone computers. ... Though a few of the libraries subscribe to online electronic resources such as EBSCOhost, Access to Global Online Research in Agriculture (AGORA) , Health Internetwork ...

  4. Nuclear data, cross section libraries and their application in nuclear technology

    International Nuclear Information System (INIS)

    1985-01-01

    These proceedings contain the articles presented at the named seminar. The articles deal with evaluated nuclear data libraries, computer codes for neutron transport and reactor calculations using nuclear data libraries, and the application of nuclear data libraries for the calculation of the interaction of neutron beams with materials. (HSI)

  5. A computerized energy systems code and information library at Soreq

    Energy Technology Data Exchange (ETDEWEB)

    Silverman, I; Shapira, M; Caner, D; Sapier, D [Israel Atomic Energy Commission, Yavne (Israel). Soreq Nuclear Research Center

    1996-12-01

    In the framework of the contractual agreement between the Ministry of Energy and Infrastructure and the Division of Nuclear Engineering of the Israel Atomic Energy Commission, both Soreq-NRC and Ben-Gurion University have agreed to establish, in 1991, a code center. This code center contains a library of computer codes and relevant data, with particular emphasis on nuclear power plant research and development support. The code center maintains existing computer codes and adapts them to the ever changing computing environment, keeps track of new code developments in the field of nuclear engineering, and acquires the most recent revisions of computer codes of interest. An attempt is made to collect relevant codes developed in Israel and to assure that proper documentation and application instructions are available. En addition to computer programs, the code center collects sample problems and international benchmarks to verify the codes and their applications to various areas of interest to nuclear power plant engineering and safety evaluation. Recently, the reactor simulation group at Soreq acquired, using funds provided by the Ministry of Energy and Infrastructure, a PC work station operating under a Linux operating system to give users of the library an easy on-line way to access resources available at the library. These resources include the computer codes and their documentation, reports published by the reactor simulation group, and other information databases available at Soreq. Registered users set a communication line, through a modem, between their computer and the new workstation at Soreq and use it to download codes and/or information or to solve their problems, using codes from the library, on the computer at Soreq (authors).

  6. A computerized energy systems code and information library at Soreq

    International Nuclear Information System (INIS)

    Silverman, I.; Shapira, M.; Caner, D.; Sapier, D.

    1996-01-01

    In the framework of the contractual agreement between the Ministry of Energy and Infrastructure and the Division of Nuclear Engineering of the Israel Atomic Energy Commission, both Soreq-NRC and Ben-Gurion University have agreed to establish, in 1991, a code center. This code center contains a library of computer codes and relevant data, with particular emphasis on nuclear power plant research and development support. The code center maintains existing computer codes and adapts them to the ever changing computing environment, keeps track of new code developments in the field of nuclear engineering, and acquires the most recent revisions of computer codes of interest. An attempt is made to collect relevant codes developed in Israel and to assure that proper documentation and application instructions are available. En addition to computer programs, the code center collects sample problems and international benchmarks to verify the codes and their applications to various areas of interest to nuclear power plant engineering and safety evaluation. Recently, the reactor simulation group at Soreq acquired, using funds provided by the Ministry of Energy and Infrastructure, a PC work station operating under a Linux operating system to give users of the library an easy on-line way to access resources available at the library. These resources include the computer codes and their documentation, reports published by the reactor simulation group, and other information databases available at Soreq. Registered users set a communication line, through a modem, between their computer and the new workstation at Soreq and use it to download codes and/or information or to solve their problems, using codes from the library, on the computer at Soreq (authors)

  7. Towards a cyber-physical era: soft computing framework based multi-sensor array for water quality monitoring

    Science.gov (United States)

    Bhardwaj, Jyotirmoy; Gupta, Karunesh K.; Gupta, Rajiv

    2018-02-01

    New concepts and techniques are replacing traditional methods of water quality parameter measurement systems. This paper introduces a cyber-physical system (CPS) approach for water quality assessment in a distribution network. Cyber-physical systems with embedded sensors, processors and actuators can be designed to sense and interact with the water environment. The proposed CPS is comprised of sensing framework integrated with five different water quality parameter sensor nodes and soft computing framework for computational modelling. Soft computing framework utilizes the applications of Python for user interface and fuzzy sciences for decision making. Introduction of multiple sensors in a water distribution network generates a huge number of data matrices, which are sometimes highly complex, difficult to understand and convoluted for effective decision making. Therefore, the proposed system framework also intends to simplify the complexity of obtained sensor data matrices and to support decision making for water engineers through a soft computing framework. The target of this proposed research is to provide a simple and efficient method to identify and detect presence of contamination in a water distribution network using applications of CPS.

  8. Emotional Branding as an Effort to Improve the Quality of Library Service in Order to Increase the Use of Library

    OpenAIRE

    Neneng Komariah; Saleha Rodiah; Encang Saepudin

    2016-01-01

    This paper is a study of implementation of emotional branding in library services.    Emotional branding is a new paradigm in marketing world. It means create emotional nuance and its objective is to create an emotional nearness between a brand and its consumer. The library as an institution which provide information services may implement emotional branding concept in creating emotional nuance which will build emotional nearness between the library and its users. It is hope that the emotiona...

  9. RFID for libraries a practical guide

    CERN Document Server

    Pandian, M Paul

    2010-01-01

    Many modern technologies give the impression that they somehow work by magic, particularly when they operate automatically and their mechanisms are invisible. A technology called RFID (Radio Frequency Identification), which is relatively new, has exactly this characteristic. Librarians everywhere are closely watching RFID technology. Advances over using bar codes on library materials, RFID tags are being touted as a way to radically redesign how library materials are handled. But it is expensive. The tags are vulnerable to wear and tear and the technology is not fully developed. It's unclear what sort of return on investment (ROI) it provides. While it is common for libraries to now have information technology expertise within their organizations, RFID may appear unfamiliar. RFID for Libraries: A practical guide examines what RFID technology is and how it works. The book reviews the development of computer technology and its effects on library operations over the last couple of decades. Further examination fe...

  10. National Software Reference Library (NSRL)

    Science.gov (United States)

    National Software Reference Library (NSRL) (PC database for purchase)   A collaboration of the National Institute of Standards and Technology (NIST), the National Institute of Justice (NIJ), the Federal Bureau of Investigation (FBI), the Defense Computer Forensics Laboratory (DCFL),the U.S. Customs Service, software vendors, and state and local law enforement organizations, the NSRL is a tool to assist in fighting crime involving computers.

  11. Canadian Public Library Users are Unaware of Their Information Literacy Deficiencies as Related to Internet Use and Public Libraries are Challenged to Address These Needs. A Review of: Julien, Heidi and Cameron Hoffman. “Information Literacy Training in Canada’s Public Libraries.” Library Quarterly 78.1 (2008: 19‐41.

    Directory of Open Access Journals (Sweden)

    Martha Ingrid Preddie

    2009-12-01

    searching and for e‐mail.It was also observed that in general the physical environment was not conducive to computer use due to uncomfortable or absent seating and a lack of privacy. Additionally, only two sites had areas specifically designated for IL instruction.Of the 25 respondents, 19 reported at least five years experience with the Internet, 9 of whom cited experience of 10 years or more. Self‐reported confidence with the Internet was high: 16 individuals claimed to be very confident, 7 somewhat confident, and only 2 lacking in confidence. There was a weak positive correlation between years of use and individuals’ reported levels of confidence.Customers reported interest in improving computer literacy (e.g., keyboarding ability and IL skills (ability to use more sources of information. Some expressed a desire “to improve certain personal attitudes” (30, such as patience when conducting Internet searches. When presented with the Association of College and Research Libraries’ definition of IL, 13 (52% of those interviewed claimed to be information literate, 8 were ambivalent, and 4 admitted to being information illiterate. Those who professed to be information literate had no particular feeling about this state of being, however 10 interviewees admitted feeling positive about being able to use the Internet to retrieve information. Most of those interviewed (15 disagreed that a paucity of IL skills is a deterrent to “accessing online information efficiently and effectively” (30. Eleven reported development of information skills through self teaching, while 8 cited secondary schools or tertiary educational institutions. However, such training was more in terms of computer technology education than IL. Eleven of the participants expressed a desire for additional IL training, 5 of whom indicated a preference for the public library to supply such training. Customers identified face‐to‐face, rather than online, as the ideal training format. Four

  12. A Model for Designing Library Instruction for Distance Learning

    Science.gov (United States)

    Rand, Angela Doucet

    2013-01-01

    Providing library instruction in distance learning environments presents a unique set of challenges for instructional librarians. Innovations in computer-mediated communication and advances in cognitive science research provide the opportunity for designing library instruction that meets a variety of student information seeking needs. Using a…

  13. Comprehensive evaluation and optimization of amplicon library preparation methods for high-throughput antibody sequencing.

    Science.gov (United States)

    Menzel, Ulrike; Greiff, Victor; Khan, Tarik A; Haessler, Ulrike; Hellmann, Ina; Friedensohn, Simon; Cook, Skylar C; Pogson, Mark; Reddy, Sai T

    2014-01-01

    High-throughput sequencing (HTS) of antibody repertoire libraries has become a powerful tool in the field of systems immunology. However, numerous sources of bias in HTS workflows may affect the obtained antibody repertoire data. A crucial step in antibody library preparation is the addition of short platform-specific nucleotide adapter sequences. As of yet, the impact of the method of adapter addition on experimental library preparation and the resulting antibody repertoire HTS datasets has not been thoroughly investigated. Therefore, we compared three standard library preparation methods by performing Illumina HTS on antibody variable heavy genes from murine antibody-secreting cells. Clonal overlap and rank statistics demonstrated that the investigated methods produced equivalent HTS datasets. PCR-based methods were experimentally superior to ligation with respect to speed, efficiency, and practicality. Finally, using a two-step PCR based method we established a protocol for antibody repertoire library generation, beginning from inputs as low as 1 ng of total RNA. In summary, this study represents a major advance towards a standardized experimental framework for antibody HTS, thus opening up the potential for systems-based, cross-experiment meta-analyses of antibody repertoires.

  14. Development of electronic libraries in Nigerian universities | Edem ...

    African Journals Online (AJOL)

    Development of electronic libraries in Nigerian universities. ... electronic libraries to enhance the quality of teaching, learning and research. ... as the instrument for data collection and academic librarians as respondents . ... Furthermore, frequent power outage and poor attitude o f university management s respectively are ...

  15. Streamlined library programming how to improve services and cut costs

    CERN Document Server

    Porter-Reynolds, Daisy

    2014-01-01

    In their roles as community centers, public libraries offer many innovative and appealing programs; but under current budget cuts, library resources are stretched thin. With slashed budgets and limited staff hours, what can libraries do to best serve their publics? This how-to guide provides strategies for streamlining library programming in public libraries while simultaneously maintaining-or even improving-quality delivery. The wide variety of principles and techniques described can be applied on a selective basis to libraries of all sizes. Based upon the author's own extensive experience as

  16. Efficient Management of High-Throughput Screening Libraries with SAVANAH

    DEFF Research Database (Denmark)

    List, Markus; Elnegaard, Marlene Pedersen; Schmidt, Steffen

    2017-01-01

    ) sample information from the library to experimental results from the assay plates. All results can be exported to the R statistical environment or piped into HiTSeekR (http://hitseekr.compbio.sdu.dk) for comprehensive follow-up analyses. In summary, SAVANAH supports the HTS community in managing...... for such screens are molecular libraries, that is, microtiter plates with solubilized reagents such as siRNAs, shRNAs, miRNA inhibitors or mimics, and sgRNAs, or small compounds, that is, drugs. These reagents are typically condensed to provide enough material for covering several screens. Library plates thus need...... to be serially diluted before they can be used as assay plates. This process, however, leads to an explosion in the number of plates and samples to be tracked. Here, we present SAVANAH, the first tool to effectively manage molecular screening libraries across dilution series. It conveniently links (connects...

  17. A simple, high throughput method to locate single copy sequences from Bacterial Artificial Chromosome (BAC libraries using High Resolution Melt analysis

    Directory of Open Access Journals (Sweden)

    Caligari Peter DS

    2010-05-01

    Full Text Available Abstract Background The high-throughput anchoring of genetic markers into contigs is required for many ongoing physical mapping projects. Multidimentional BAC pooling strategies for PCR-based screening of large insert libraries is a widely used alternative to high density filter hybridisation of bacterial colonies. To date, concerns over reliability have led most if not all groups engaged in high throughput physical mapping projects to favour BAC DNA isolation prior to amplification by conventional PCR. Results Here, we report the first combined use of Multiplex Tandem PCR (MT-PCR and High Resolution Melt (HRM analysis on bacterial stocks of BAC library superpools as a means of rapidly anchoring markers to BAC colonies and thereby to integrate genetic and physical maps. We exemplify the approach using a BAC library of the model plant Arabidopsis thaliana. Super pools of twenty five 384-well plates and two-dimension matrix pools of the BAC library were prepared for marker screening. The entire procedure only requires around 3 h to anchor one marker. Conclusions A pre-amplification step during MT-PCR allows high multiplexing and increases the sensitivity and reliability of subsequent HRM discrimination. This simple gel-free protocol is more reliable, faster and far less costly than conventional PCR screening. The option to screen in parallel 3 genetic markers in one MT-PCR-HRM reaction using templates from directly pooled bacterial stocks of BAC-containing bacteria further reduces time for anchoring markers in physical maps of species with large genomes.

  18. High energy physics and grid computing

    International Nuclear Information System (INIS)

    Yu Chuansong

    2004-01-01

    The status of the new generation computing environment of the high energy physics experiments is introduced briefly in this paper. The development of the high energy physics experiments and the new computing requirements by the experiments are presented. The blueprint of the new generation computing environment of the LHC experiments, the history of the Grid computing, the R and D status of the high energy physics grid computing technology, the network bandwidth needed by the high energy physics grid and its development are described. The grid computing research in Chinese high energy physics community is introduced at last. (authors)

  19. Changing the Concept and Measure of Service Quality in Academic Libraries.

    Science.gov (United States)

    Nitecki, Danuta A.

    1996-01-01

    The diagnostic instrument SERVQUAL has been developed to measure customer expectations and perceptions of service. This article describes a study that investigated how applicable SERVQUAL is to academic libraries and how influential the study's findings might be in changing concepts of the management of academic library services. Two SERVQUAL…

  20. Towards Portable Large-Scale Image Processing with High-Performance Computing.

    Science.gov (United States)

    Huo, Yuankai; Blaber, Justin; Damon, Stephen M; Boyd, Brian D; Bao, Shunxing; Parvathaneni, Prasanna; Noguera, Camilo Bermudez; Chaganti, Shikha; Nath, Vishwesh; Greer, Jasmine M; Lyu, Ilwoo; French, William R; Newton, Allen T; Rogers, Baxter P; Landman, Bennett A

    2018-05-03

    High-throughput, large-scale medical image computing demands tight integration of high-performance computing (HPC) infrastructure for data storage, job distribution, and image processing. The Vanderbilt University Institute for Imaging Science (VUIIS) Center for Computational Imaging (CCI) has constructed a large-scale image storage and processing infrastructure that is composed of (1) a large-scale image database using the eXtensible Neuroimaging Archive Toolkit (XNAT), (2) a content-aware job scheduling platform using the Distributed Automation for XNAT pipeline automation tool (DAX), and (3) a wide variety of encapsulated image processing pipelines called "spiders." The VUIIS CCI medical image data storage and processing infrastructure have housed and processed nearly half-million medical image volumes with Vanderbilt Advanced Computing Center for Research and Education (ACCRE), which is the HPC facility at the Vanderbilt University. The initial deployment was natively deployed (i.e., direct installations on a bare-metal server) within the ACCRE hardware and software environments, which lead to issues of portability and sustainability. First, it could be laborious to deploy the entire VUIIS CCI medical image data storage and processing infrastructure to another HPC center with varying hardware infrastructure, library availability, and software permission policies. Second, the spiders were not developed in an isolated manner, which has led to software dependency issues during system upgrades or remote software installation. To address such issues, herein, we describe recent innovations using containerization techniques with XNAT/DAX which are used to isolate the VUIIS CCI medical image data storage and processing infrastructure from the underlying hardware and software environments. The newly presented XNAT/DAX solution has the following new features: (1) multi-level portability from system level to the application level, (2) flexible and dynamic software