WorldWideScience

Sample records for computing software structural

  1. A new entropy based method for computing software structural complexity

    CERN Document Server

    Roca, J L

    2002-01-01

    In this paper a new methodology for the evaluation of software structural complexity is described. It is based on the entropy evaluation of the random uniform response function associated with the so called software characteristic function SCF. The behavior of the SCF with the different software structures and their relationship with the number of inherent errors is investigated. It is also investigated how the entropy concept can be used to evaluate the complexity of a software structure considering the SCF as a canonical representation of the graph associated with the control flow diagram. The functions, parameters and algorithms that allow to carry out this evaluation are also introduced. After this analytic phase follows the experimental phase, verifying the consistency of the proposed metric and their boundary conditions. The conclusion is that the degree of software structural complexity can be measured as the entropy of the random uniform response function of the SCF. That entropy is in direct relation...

  2. New ATLAS Software & Computing Organization

    CERN Multimedia

    Barberis, D

    Following the election by the ATLAS Collaboration Board of Dario Barberis (Genoa University/INFN) as Computing Coordinator and David Quarrie (LBNL) as Software Project Leader, it was considered necessary to modify the organization of the ATLAS Software & Computing ("S&C") project. The new organization is based upon the following principles: separation of the responsibilities for computing management from those of software development, with the appointment of a Computing Coordinator and a Software Project Leader who are both members of the Executive Board; hierarchical structure of responsibilities and reporting lines; coordination at all levels between TDAQ, S&C and Physics working groups; integration of the subdetector software development groups with the central S&C organization. A schematic diagram of the new organization can be seen in Fig.1. Figure 1: new ATLAS Software & Computing organization. Two Management Boards will help the Computing Coordinator and the Software Project...

  3. Computing and software

    Directory of Open Access Journals (Sweden)

    White, G. C.

    2004-06-01

    Full Text Available The reality is that the statistical methods used for analysis of data depend upon the availability of software. Analysis of marked animal data is no different than the rest of the statistical field. The methods used for analysis are those that are available in reliable software packages. Thus, the critical importance of having reliable, up–to–date software available to biologists is obvious. Statisticians have continued to develop more robust models, ever expanding the suite of potential analysis methods available. But without software to implement these newer methods, they will languish in the abstract, and not be applied to the problems deserving them. In the Computers and Software Session, two new software packages are described, a comparison of implementation of methods for the estimation of nest survival is provided, and a more speculative paper about how the next generation of software might be structured is presented. Rotella et al. (2004 compare nest survival estimation with different software packages: SAS logistic regression, SAS non–linear mixed models, and Program MARK. Nests are assumed to be visited at various, possibly infrequent, intervals. All of the approaches described compute nest survival with the same likelihood, and require that the age of the nest is known to account for nests that eventually hatch. However, each approach offers advantages and disadvantages, explored by Rotella et al. (2004. Efford et al. (2004 present a new software package called DENSITY. The package computes population abundance and density from trapping arrays and other detection methods with a new and unique approach. DENSITY represents the first major addition to the analysis of trapping arrays in 20 years. Barker & White (2004 discuss how existing software such as Program MARK require that each new model’s likelihood must be programmed specifically for that model. They wishfully think that future software might allow the user to combine

  4. Computing and software

    Science.gov (United States)

    White, Gary C.; Hines, J.E.

    2004-01-01

    The reality is that the statistical methods used for analysis of data depend upon the availability of software. Analysis of marked animal data is no different than the rest of the statistical field. The methods used for analysis are those that are available in reliable software packages. Thus, the critical importance of having reliable, up–to–date software available to biologists is obvious. Statisticians have continued to develop more robust models, ever expanding the suite of potential analysis methodsavailable. But without software to implement these newer methods, they will languish in the abstract, and not be applied to the problems deserving them.In the Computers and Software Session, two new software packages are described, a comparison of implementation of methods for the estimation of nest survival is provided, and a more speculative paper about how the next generation of software might be structured is presented.Rotella et al. (2004) compare nest survival estimation with different software packages: SAS logistic regression, SAS non–linear mixed models, and Program MARK. Nests are assumed to be visited at various, possibly infrequent, intervals. All of the approaches described compute nest survival with the same likelihood, and require that the age of the nest is known to account for nests that eventually hatch. However, each approach offers advantages and disadvantages, explored by Rotella et al. (2004).Efford et al. (2004) present a new software package called DENSITY. The package computes population abundance and density from trapping arrays and other detection methods with a new and unique approach. DENSITY represents the first major addition to the analysis of trapping arrays in 20 years.Barker & White (2004) discuss how existing software such as Program MARK require that each new model’s likelihood must be programmed specifically for that model. They wishfully think that future software might allow the user to combine pieces of likelihood

  5. Copyright and Computer Software.

    Science.gov (United States)

    Haugness, C. A.

    1985-01-01

    Explores some of the difficult areas of copyright protection for computer software including legal and illegal copying; values imparted to students if illegal software copies are used; and teacher and administrator responsibility. A suggested school district policy on software copyright is presented. (MBR)

  6. Who Owns Computer Software?

    Science.gov (United States)

    Branscomb, Anne Wells

    1995-01-01

    Discusses the protection of intellectual property as it applies to computer software and its impact on private enterprise and the public good. Highlights include the role of patents, copyrights, and trade secrets; some court cases; and recommendations for alternatives to the existing legal framework for protecting computer software. (KRN)

  7. OpenStructure: an integrated software framework for computational structural biology.

    Science.gov (United States)

    Biasini, M; Schmidt, T; Bienert, S; Mariani, V; Studer, G; Haas, J; Johner, N; Schenk, A D; Philippsen, A; Schwede, T

    2013-05-01

    Research projects in structural biology increasingly rely on combinations of heterogeneous sources of information, e.g. evolutionary information from multiple sequence alignments, experimental evidence in the form of density maps and proximity constraints from proteomics experiments. The OpenStructure software framework, which allows the seamless integration of information of different origin, has previously been introduced. The software consists of C++ libraries which are fully accessible from the Python programming language. Additionally, the framework provides a sophisticated graphics module that interactively displays molecular structures and density maps in three dimensions. In this work, the latest developments in the OpenStructure framework are outlined. The extensive capabilities of the framework will be illustrated using short code examples that show how information from molecular-structure coordinates can be combined with sequence data and/or density maps. The framework has been released under the LGPL version 3 license and is available for download from http://www.openstructure.org.

  8. Computer software profiles

    Energy Technology Data Exchange (ETDEWEB)

    Anon.

    2009-04-15

    A review of various computer software programs designed for use in the petroleum industry was presented with reference to each programs capabilities, efficiencies, and operational parameters. This article highlighted 3 software packages developed by Epic Consulting Services Ltd. for reservoir surveillance, forecasting and waterflooding optimization. Two oil and gas software solutions developed by Energy Navigator were presented, notably AFE Navigator for capital tracking and Value Navigator for full suite engineering applications. Six reservoir simulation packages developed by Fekete Associates Inc. were also presented along with software programs developed by 3esi for resource planning, reservoir forecasting, scheduling tracking, archiving, variance reporting and project performance monitoring. Streamsim Technologies Inc. has also developed software packages known as studioSL and 3DSL for geologists and reservoir engineers. Three multiphase flow modelling software packages developed by Neotec were also highlighted. These included Wellflo, Pipeflo and Forgas to model pipelines, flow conditions and wellbore conditions. The consulting software packages developed by Schlumberger Information Solutions for use in geology and geophysics were also highlighted along with software solutions for reservoir engineering; production optimization; reserves risk management; and information management. tabs., figs.

  9. A state-of-the-art report on software operation structure of the digital control computer system

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Bong Kee; Lee, Kyung Hoh; Joo, Jae Yoon; Jang, Yung Woo; Shin, Hyun Kook [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1994-06-01

    CANDU Nuclear Power Plants including Wolsong 1 and 2/3/4 are controlled by a real-time plant control computer system. This report was written to provide an overview on the station control computer software which belongs to one of the most advanced real-time computing application area, along with the Fuel Handling Machine design concepts. The combination of well designed control computer and Fuel Handling Machine allow changing fuel bundles while the plant is in operation. Design methodologies and software structure are discussed along with the interface between the two systems. 29 figs., 2 tabs., 20 refs. (Author).

  10. CMS software and computing

    CERN Document Server

    Charlot, C

    2003-01-01

    CMS is one of the two general-purpose HEP experiments currently under construction for the Large Hadron Collider at CERN. The handling of multi-petabyte data samples in a worldwide context requires computing and software systems with unprecedented scale and complexity. We describe how CMS is meeting the many data analysis challenges in the LHC area. We cover in particular our system of globally distributed regional centers, the status of our object-oriented software, and our strategies for Grid-enriched data analysis.

  11. Software abstractions and computational issues in parallel structure adaptive mesh methods for electronic structure calculations

    Energy Technology Data Exchange (ETDEWEB)

    Kohn, S.; Weare, J.; Ong, E.; Baden, S.

    1997-05-01

    We have applied structured adaptive mesh refinement techniques to the solution of the LDA equations for electronic structure calculations. Local spatial refinement concentrates memory resources and numerical effort where it is most needed, near the atomic centers and in regions of rapidly varying charge density. The structured grid representation enables us to employ efficient iterative solver techniques such as conjugate gradient with FAC multigrid preconditioning. We have parallelized our solver using an object- oriented adaptive mesh refinement framework.

  12. 48 CFR 227.7202 - Commercial computer software and commercial computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7202 Commercial computer software and commercial computer software documentation. ... software and commercial computer software documentation. 227.7202 Section 227.7202 Federal Acquisition...

  13. 48 CFR 227.7203 - Noncommercial computer software and noncommercial computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203 Noncommercial computer software and noncommercial computer software documentation. ... software and noncommercial computer software documentation. 227.7203 Section 227.7203 Federal Acquisition...

  14. Computer software profiles

    Energy Technology Data Exchange (ETDEWEB)

    Anon

    2005-03-01

    A variety of currently available computer software packages are described. The following packages are included: By Advanced Geotechnology Inc: (1) STABView(TM), a state-of-the-art well planning package for wellbore stability, lost circulation, fracturing and sand production risk analysis, and (2) ROCKSBank(TM) which is a new database and analysis toolkit for a wide range of exploration, drilling, completion and production applications. The Computer Modelling Group Inc offers: (1) IMEX-3D, a package designed to do full-field conventional black-oil, condensate, pseudo miscible, and polymer simulation, and (2) GEM, an Equation of State Simulator. The latest enhancements include geochemistry for acid gases and greenhouse gas injection and storage, and enhanced modelling of gas solubility in brine, among others. (3) STARS is another member of the Computer Modelling Group family; it is a 3-D Advanced Thermal Stimulator, recently augmented to do relative permeability hysteresis, creep model in geomechanics, and multiphase co injection for ES-SAGD. Epic Consulting Services Ltd offers ResWorks(TM) and ResAssist(TM), the former to view pool production history on a macro scale and perform surveillance evaluation of patterns and pods, the latter to perform rapid water flood optimization simulations. Fekete Associates Inc has a stable of nine software packages of interest to the reservoir engineer ranging from F.A.S.T. CBM for coalbed methane development or reservoir analysis to F.A.S.T. Welltest to analyse pressure data, permeability, wellbore skin, drainage area, hydrocarbon-in-place, and to do simulation exercises. Neotechnology Consultants Ltd provide PIPEFLO and PIPEFLOW AS to conduct single and multi-phase flow analysis and PVT behaviour modelling for pipeline transmission and gathering systems. Also available are the FORGAS family of software packages for forecasting multiphase production from the reservoir to the plant. Streamsim Technologies Inc has 3DSL, a fully 3-D

  15. Software and Computing News

    CERN Multimedia

    Barberis, D

    The last several months have been very busy ones for the ATLAS software developers. They've been trying to cope with the competing demands of multiple software stress tests and testbeds. These include Data Challenge Two (DC2), the Combined Testbeam (CTB), preparations for the Physics Workshop to be held in Rome in June 2005, and other testbeds, primarily one for the High-Level Trigger. Data Challenge 2 (DC2) The primary goal of this was to validate the computing model and to provide a test of simulating a day's worth of ATLAS data (10 million events) and of fully processing it and making it available to the physicists within 10 days (i.e. a 10% scale test). DC2 consists of three parts - the generation, simulation, and mixing of a representative sample of physics events with background events; the reconstruction of the mixed samples with initial classification into the different physics signatures; and the distribution of the data to multiple remote sites (Tier-1 centers) for analysis by physicists. Figu...

  16. Computer games and software engineering

    CERN Document Server

    Cooper, Kendra M L

    2015-01-01

    Computer games represent a significant software application domain for innovative research in software engineering techniques and technologies. Game developers, whether focusing on entertainment-market opportunities or game-based applications in non-entertainment domains, thus share a common interest with software engineers and developers on how to best engineer game software.Featuring contributions from leading experts in software engineering, the book provides a comprehensive introduction to computer game software development that includes its history as well as emerging research on the inte

  17. Computing and software

    OpenAIRE

    White, G C; Hines, J.E.

    2004-01-01

    The reality is that the statistical methods used for analysis of data depend upon the availability of software. Analysis of marked animal data is no different than the rest of the statistical field. The methods used for analysis are those that are available in reliable software packages. Thus, the critical importance of having reliable, up–to–date software available to biologists is obvious. Statisticians have continued to develop more robust models, ever expanding the suite of potential ...

  18. Computing and software

    OpenAIRE

    White, G C; Hines, J.E.

    2004-01-01

    The reality is that the statistical methods used for analysis of data depend upon the availability of software. Analysis of marked animal data is no different than the rest of the statistical field. The methods used for analysis are those that are available in reliable software packages. Thus, the critical importance of having reliable, up–to–date software available to biologists is obvious. Statisticians have continued to develop more robust models, ever expanding the suite of potential anal...

  19. 48 CFR 12.212 - Computer software.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Computer software. 12.212... software. (a) Commercial computer software or commercial computer software documentation shall be acquired... required to— (1) Furnish technical information related to commercial computer software or commercial...

  20. Software system for computing material and structural properties of bone and muscle in the lower extremity from pQCT

    Science.gov (United States)

    Makrogiannis, Sokratis; Ferrucci, Luigi

    2014-06-01

    Peripheral Quantitative Computed Tomography (pQCT) is a non-invasive imaging technology that is well-suited for quantification of bone structural and material properties. Because of its increasing use and applicability, the development of automated quantification methods for pQCT images is an appealing field of research. In this paper we introduce a software system for hard and soft tissue quantification in the lower leg using pQCT imaging data. The main stages of our approach are the segmentation and identification of bone, muscle and fat, and the computation of densitometric and geometric variables of each regional tissue type. Our system was validated against reference area and densitometric measurements over a set of test images and produced encouraging results.

  1. 48 CFR 252.227-7014 - Rights in noncommercial computer software and noncommercial computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... documentation. (d) Third party copyrighted computer software or computer software documentation. The Contractor... Government under any patent. (j) Limitation on charges for rights in computer software or computer software... computer software and noncommercial computer software documentation. 252.227-7014 Section 252.227-7014...

  2. 48 CFR 227.7203-15 - Subcontractor rights in computer software or computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203-15 Subcontractor rights in computer software or computer software documentation. (a... computer software or computer software documentation. 227.7203-15 Section 227.7203-15 Federal Acquisition...

  3. 48 CFR 227.7202-3 - Rights in commercial computer software or commercial computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7202-3 Rights in commercial computer software or commercial computer software documentation... computer software or commercial computer software documentation. 227.7202-3 Section 227.7202-3 Federal...

  4. Computing handbook computer science and software engineering

    CERN Document Server

    Gonzalez, Teofilo; Tucker, Allen

    2014-01-01

    Overview of Computer Science Structure and Organization of Computing Peter J. DenningComputational Thinking Valerie BarrAlgorithms and Complexity Data Structures Mark WeissBasic Techniques for Design and Analysis of Algorithms Edward ReingoldGraph and Network Algorithms Samir Khuller and Balaji RaghavachariComputational Geometry Marc van KreveldComplexity Theory Eric Allender, Michael Loui, and Kenneth ReganFormal Models and Computability Tao Jiang, Ming Li, and Bala

  5. Computer systems and software engineering

    Science.gov (United States)

    Mckay, Charles W.

    1988-01-01

    The High Technologies Laboratory (HTL) was established in the fall of 1982 at the University of Houston Clear Lake. Research conducted at the High Tech Lab is focused upon computer systems and software engineering. There is a strong emphasis on the interrelationship of these areas of technology and the United States' space program. In Jan. of 1987, NASA Headquarters announced the formation of its first research center dedicated to software engineering. Operated by the High Tech Lab, the Software Engineering Research Center (SERC) was formed at the University of Houston Clear Lake. The High Tech Lab/Software Engineering Research Center promotes cooperative research among government, industry, and academia to advance the edge-of-knowledge and the state-of-the-practice in key topics of computer systems and software engineering which are critical to NASA. The center also recommends appropriate actions, guidelines, standards, and policies to NASA in matters pertinent to the center's research. Results of the research conducted at the High Tech Lab/Software Engineering Research Center have given direction to many decisions made by NASA concerning the Space Station Program.

  6. DFI Computer Modeling Software (CMS)

    Energy Technology Data Exchange (ETDEWEB)

    Cazalet, E.G.; Deziel, L.B. Jr.; Haas, S.M.; Martin, T.W.; Nesbitt, D.M.; Phillips, R.L.

    1979-10-01

    The data base management system used to create, edit and store models data and solutions for the LEAP system is described. The software is entirely in FORTRAN-G for the IBM 370 series of computers and provides interface with a commercial data base system SYSTEM-2000.

  7. 48 CFR 227.7203-2 - Acquisition of noncommercial computer software and computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... practicable, solicitations and contracts— (1) Identify the types of computer software and the quantity of... documentation will be delivered; (2) Establish each type of computer software or computer software documentation... computer software or computer software documentation under a fixed-price type contract; (4) Include...

  8. Adoption of High Performance Computational (HPC) Modeling Software for Widespread Use in the Manufacture of Welded Structures

    Energy Technology Data Exchange (ETDEWEB)

    Brust, Frederick W. [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Punch, Edward F. [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Twombly, Elizabeth Kurth [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Kalyanam, Suresh [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Kennedy, James [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Hattery, Garty R. [Engineering Mechanics Corporation of Columbus, Columbus, OH (United States); Dodds, Robert H. [Professional Consulting Services, Inc., Lisle, IL (United States); Mach, Justin C [Caterpillar, Peoria, IL (United States); Chalker, Alan [Ohio Supercomputer Center (OSC), Columbus, OH (United States); Nicklas, Jeremy [Ohio Supercomputer Center (OSC), Columbus, OH (United States); Gohar, Basil M [Ohio Supercomputer Center (OSC), Columbus, OH (United States); Hudak, David [Ohio Supercomputer Center (OSC), Columbus, OH (United States)

    2016-12-30

    This report summarizes the final product developed for the US DOE Small Business Innovation Research (SBIR) Phase II grant made to Engineering Mechanics Corporation of Columbus (Emc2) between April 16, 2014 and August 31, 2016 titled ‘Adoption of High Performance Computational (HPC) Modeling Software for Widespread Use in the Manufacture of Welded Structures’. Many US companies have moved fabrication and production facilities off shore because of cheaper labor costs. A key aspect in bringing these jobs back to the US is the use of technology to render US-made fabrications more cost-efficient overall with higher quality. One significant advantage that has emerged in the US over the last two decades is the use of virtual design for fabrication of small and large structures in weld fabrication industries. Industries that use virtual design and analysis tools have reduced material part size, developed environmentally-friendly fabrication processes, improved product quality and performance, and reduced manufacturing costs. Indeed, Caterpillar Inc. (CAT), one of the partners in this effort, continues to have a large fabrication presence in the US because of the use of weld fabrication modeling to optimize fabrications by controlling weld residual stresses and distortions and improving fatigue, corrosion, and fracture performance. This report describes Emc2’s DOE SBIR Phase II final results to extend an existing, state-of-the-art software code, Virtual Fabrication Technology (VFT®), currently used to design and model large welded structures prior to fabrication - to a broader range of products with widespread applications for small and medium-sized enterprises (SMEs). VFT® helps control distortion, can minimize and/or control residual stresses, control welding microstructure, and pre-determine welding parameters such as weld-sequencing, pre-bending, thermal-tensioning, etc. VFT® uses material properties, consumable properties, etc. as inputs

  9. 48 CFR 212.212 - Computer software.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 3 2010-10-01 2010-10-01 false Computer software. 212.212... Acquisition of Commercial Items 212.212 Computer software. (1) Departments and agencies shall identify and... technology development), opportunities for the use of commercial computer software and other non...

  10. Designing Scientific Software for Heterogeneous Computing

    DEFF Research Database (Denmark)

    Glimberg, Stefan Lemvig

    , algorithms and data structures must be designed to utilize the underlying parallel architecture. The architectural changes in hardware design within the last decade, from single to multi and many-core architectures, require software developers to identify and properly implement methods that both exploit...... makes parallel software design applicable, but also a challenge for scientific software developers at all levels. We have developed a generic C++ library for fast prototyping of large-scale PDEs solvers based on flexible-order finite difference approximations on structured regular grids. The library...... is designed with a high abstraction interface to improve developer productivity. The library is based on modern template-based design concepts as described in Glimberg, Engsig-Karup, Nielsen & Dammann (2013). The library utilizes heterogeneous CPU/GPU environments in order to maximize computational throughput...

  11. 48 CFR 227.7203-16 - Providing computer software or computer software documentation to foreign governments, foreign...

    Science.gov (United States)

    2010-10-01

    ... software or computer software documentation to foreign governments, foreign contractors, or international... REGULATIONS SYSTEM, DEPARTMENT OF DEFENSE GENERAL CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203-16 Providing computer software...

  12. Modelling Software for Structure Metrics

    NARCIS (Netherlands)

    van den Broek, P.M.; van den Berg, Klaas

    In the approach to structural software metrics, software is modelled by means of flowgraphs. A tacit assumption in this approach is that the structure of a program is reflected by the structure of the flowgraph. When only the flow of control between commands is considered this assumption is valid;

  13. Computer software summaries. Numbers 325 through 423

    Energy Technology Data Exchange (ETDEWEB)

    1978-08-01

    The National Energy Software Center (NESC) serves as the software exchange and information center for the U.S. Department of Energy and the Nuclear Regulatory Commission. These summaries describe agency-sponsored software which is at the specification stage, under development, being checked out, in use, or available at agency offices, laboratories, and contractor installations. They describe software which is not included in the NESC library due to its preliminary status or because it is believed to be of limited interest. Codes dealing with the following subjects are included: cross section and resonance integral calculations; spectrum calculations, generation of group constants, lattice and cell problems; static design studies; depletion, fuel management, cost analysis, and power plant economics; space-independent kinetics; space--time kinetics, coupled neutronics--hydrodynamics--thermodynamics; and excursion simulations; radiological safety, hazard and accident analysis; heat transfer and fluid flow; deformation and stress distribution computations, structural analysis, and engineering design studies; gamma heating and shield design; reactor systems analysis; data preparation; data management; subsidiary calculations; experimental data processing; general mathematical and computing system routines; materials; environmental and earth sciences; space sciences; electronics and engineering equipment; chemistry; particle accelerators and high-voltage machines; physics; magnetic fusion research; biology and medicine; and data. (RWR)

  14. The ANS mathematics and computation software standards

    Energy Technology Data Exchange (ETDEWEB)

    Smetana, A. O. [Savannah River National Laboratory, Washington Savannah River Company, Aiken, SC 29808 (United States)

    2006-07-01

    The Mathematics and Computations Div. of the American Nuclear Society sponsors the ANS-10 Standards Subcommittee. This subcommittee, which is part of the ANS Standards Committee, currently maintains three ANSI/ANS software standards. These standards are: Portability of Scientific and Engineering Software, ANS-10.2; Guidelines for the Verification and Validation of Scientific and Engineering Computer Programs for the Nuclear Industry, ANS-10.4; and Accommodating User Needs in Scientific and Engineering Computer Software Development, ANS-10.5. A fourth Standard, Documentation of Computer Software, ANS-10.3, is available as a historical Standard. (authors)

  15. Ethics in computer software design and development

    Science.gov (United States)

    Alan J. Thomson; Daniel L. Schmoldt

    2001-01-01

    Over the past 20 years, computer software has become integral and commonplace for operational and management tasks throughout agricultural and natural resource disciplines. During this software infusion, however, little thought has been afforded human impacts, both good and bad. This paper examines current ethical issues of software system design and development in...

  16. 48 CFR 227.7203-3 - Early identification of computer software or computer software documentation to be furnished to...

    Science.gov (United States)

    2010-10-01

    ..., DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203-3 Early identification of computer software or computer software documentation to be furnished to the Government with... computer software or computer software documentation to be furnished to the Government with restrictions on...

  17. Computer Software & Intellectual Property. Background Paper.

    Science.gov (United States)

    Congress of the U.S., Washington, DC. Office of Technology Assessment.

    This background paper reviews copyright, patent, and trade secret protections as these issues are related to computer software. Topics discussed include current issues regarding legal protection for computer software including the necessity for defining intellectual property, determining what should or should not be protected, commerical piracy,…

  18. Socorro Electronic Structure Software

    Energy Technology Data Exchange (ETDEWEB)

    2017-08-03

    Socorro can be used to compute the ground-state electron density for a periodically repeated simulation cell in which the external potential is obtained from norm-conserving pseudopotentials or projector-augmented-wave functions. The ground-state electron density is computed by summation over one-electron orbitals which are obtained using the Kohn-Sham formulation of density-functional theory and which are expanded in a plane-wave basis. Various quantities can be computed from the ground-state solution, including atomic forces which can be used to optimize atom positions in the simulation cell and to perform molecular dynamics simulations.

  19. Copyright law, computer software, and government acquisition

    OpenAIRE

    Dauphinais, Paul Raymond

    1984-01-01

    Approved for public release; distribution is unlimited This thesis examines copyright law as it relates to computer software and how this law affects the Government acquisition of computer software. Following a differentiation of copyright law, patent law, and trade secrets, a brief history of the evolution of copyright law is presented. Current Government software acquisition practices are examined with respect to copyright statutes. The 1984 Betamax case is examined and related to sof...

  20. Collection Of Software For Computer Graphics

    Science.gov (United States)

    Hibbard, Eric A.; Makatura, George

    1990-01-01

    Ames Research Graphics System (ARCGRAPH) collection of software libraries and software utilities assisting researchers in generating, manipulating, and visualizing graphical data. Defines metafile format containing device-independent graphical data. File format used with various computer-graphics-manipulation and -animation software packages at Ames, including SURF (COSMIC Program ARC-12381) and GAS (COSMIC Program ARC-12379). Consists of two-stage "pipeline" used to put out graphical primitives. ARCGRAPH libraries developed on VAX computer running VMS.

  1. Computer software project management: an introduction

    OpenAIRE

    Liberto, Samuel Matthew

    1991-01-01

    Approved for public release; distribution is unlimited This thesis addresses the general principles of computer software project management. The main objective is to aid perspective software project managers in dealing with the development and management of software projects. The definition of the classical software development life cycle is given. The components include system engineering, analysis, design, coding, testing, and maintenance. The thesis contains a description of the reasons...

  2. Software Accelerates Computing Time for Complex Math

    Science.gov (United States)

    2014-01-01

    Ames Research Center awarded Newark, Delaware-based EM Photonics Inc. SBIR funding to utilize graphic processing unit (GPU) technology- traditionally used for computer video games-to develop high-computing software called CULA. The software gives users the ability to run complex algorithms on personal computers with greater speed. As a result of the NASA collaboration, the number of employees at the company has increased 10 percent.

  3. Innovations in Computing Sciences and Software Engineering

    CERN Document Server

    Sobh, Tarek

    2010-01-01

    "Innovations in Computing Sciences and Software Engineering" includes a set of rigorously reviewed world-class manuscripts addressing and detailing state-of-the-art research projects in the areas of Computer Science, Software Engineering, Computer Engineering, and Systems Engineering and Sciences. The topics covered include: Image and Pattern Recognition: Compression, Image processing, Signal Processing Architectures, Signal Processing for Communication, Signal Processing Implementation, Speech Compression, and Video Coding Architectures; Languages and Systems: Algorithms, Databases,

  4. Music Learning Based on Computer Software

    Directory of Open Access Journals (Sweden)

    Baihui Yan

    2017-12-01

    Full Text Available In order to better develop and improve students’ music learning, the authors proposed the method of music learning based on computer software. It is still a new field to use computer music software to assist teaching. Hereby, we conducted an in-depth analysis on the computer-enabled music learning and the music learning status in secondary schools, obtaining the specific analytical data. Survey data shows that students have many cognitive problems in the current music classroom, and yet teachers have not found a reasonable countermeasure to them. Against this background, the introduction of computer music software to music learning is a new trial that can not only cultivate the students’ initiatives of music learning, but also enhance their abilities to learn music. Therefore, it is concluded that the computer software based music learning is of great significance to improving the current music learning modes and means.

  5. 48 CFR 227.7203-8 - Deferred delivery and deferred ordering of computer software and computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... GENERAL CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203-8 Deferred delivery and deferred ordering of computer software and computer... deferred ordering of computer software and computer software documentation. 227.7203-8 Section 227.7203-8...

  6. Analysing qualitative research data using computer software.

    Science.gov (United States)

    McLafferty, Ella; Farley, Alistair H

    An increasing number of clinical nurses are choosing to undertake qualitative research. A number of computer software packages are available designed for the management and analysis of qualitative data. However, while it is claimed that the use of these programs is also increasing, this claim is not supported by a search of recent publications. This paper discusses the advantages and disadvantages of using computer software packages to manage and analyse qualitative data.

  7. Computer Software for Working with Language.

    Science.gov (United States)

    Winograd, Terry

    1984-01-01

    Discusses four types of computer programs that can manipulate linguistic symbols with great facility. These programs deal with: (1) machine translation; (2) word processing; (3) question answering; and (4) adjuncts to electronic mail known as coordination systems. Also considers the future of computer software dealing with language. (JN)

  8. JPL Robotics Laboratory computer vision software library

    Science.gov (United States)

    Cunningham, R.

    1984-01-01

    The past ten years of research on computer vision have matured into a powerful real time system comprised of standardized commercial hardware, computers, and pipeline processing laboratory prototypes, supported by anextensive set of image processing algorithms. The software system was constructed to be transportable via the choice of a popular high level language (PASCAL) and a widely used computer (VAX-11/750), it comprises a whole realm of low level and high level processing software that has proven to be versatile for applications ranging from factory automation to space satellite tracking and grappling.

  9. IUE Data Analysis Software for Personal Computers

    Science.gov (United States)

    Thompson, R.; Caplinger, J.; Taylor, L.; Lawton , P.

    1996-01-01

    This report summarizes the work performed for the program titled, "IUE Data Analysis Software for Personal Computers" awarded under Astrophysics Data Program NRA 92-OSSA-15. The work performed was completed over a 2-year period starting in April 1994. As a result of the project, 450 IDL routines and eight database tables are now available for distribution for Power Macintosh computers and Personal Computers running Windows 3.1.

  10. Technical innovation: Multidimensional computerized software enabled subtraction computed tomographic angiography.

    Science.gov (United States)

    Bhatia, Mona; Rosset, Antoine; Platon, Alexandra; Didier, Dominique; Becker, Christoph D; Poletti, Pierre-Alexandre

    2010-01-01

    Computed tomographic angiography (CTA) is a frequent noninvasive alternative to digital subtraction angiography. We previously reported the development of a new subtraction software to overcome limitations of adjacent bone and calcification in CT angiographic subtraction. Our aim was to further develop and improve this fast and automated computerized software, universally available for free use and compatible with most CT scanners, thus enabling better delineation of vascular structures, artifact reduction, and shorter reading times with potential clinical benefits. This computer-based free software will be available as an open source in the next release of OsiriX at the Web site http://www.osirix-viewer.com.

  11. Computational intelligence and quantitative software engineering

    CERN Document Server

    Succi, Giancarlo; Sillitti, Alberto

    2016-01-01

    In a down-to-the earth manner, the volume lucidly presents how the fundamental concepts, methodology, and algorithms of Computational Intelligence are efficiently exploited in Software Engineering and opens up a novel and promising avenue of a comprehensive analysis and advanced design of software artifacts. It shows how the paradigm and the best practices of Computational Intelligence can be creatively explored to carry out comprehensive software requirement analysis, support design, testing, and maintenance. Software Engineering is an intensive knowledge-based endeavor of inherent human-centric nature, which profoundly relies on acquiring semiformal knowledge and then processing it to produce a running system. The knowledge spans a wide variety of artifacts, from requirements, captured in the interaction with customers, to design practices, testing, and code management strategies, which rely on the knowledge of the running system. This volume consists of contributions written by widely acknowledged experts ...

  12. Computer-assisted qualitative data analysis software.

    Science.gov (United States)

    Cope, Diane G

    2014-05-01

    Advances in technology have provided new approaches for data collection methods and analysis for researchers. Data collection is no longer limited to paper-and-pencil format, and numerous methods are now available through Internet and electronic resources. With these techniques, researchers are not burdened with entering data manually and data analysis is facilitated by software programs. Quantitative research is supported by the use of computer software and provides ease in the management of large data sets and rapid analysis of numeric statistical methods. New technologies are emerging to support qualitative research with the availability of computer-assisted qualitative data analysis software (CAQDAS).CAQDAS will be presented with a discussion of advantages, limitations, controversial issues, and recommendations for this type of software use.

  13. CADOProSys Software. General structure

    Directory of Open Access Journals (Sweden)

    Chivu Catrina

    2017-01-01

    Full Text Available Design of flexible manufacturing systems is a challenge basedon the field, structure, medium and long term management strategies of theproduction system. On the market there are many ERP software designedfor integrated management. Generally, this type of software is not focusedon manufacturing lines, managing only the documents inside themanufacturing systems, being not appropriate for design. Anyway, all theapplications from the market assumed that the user knows how to designPLM software. This assumption was the motivation for the authors of thepresent paper to design a computer aided application that will help thebeginners, the graduates, to design a manufacturing line. The softwareincludes many modules. Thus, there are modules for designing: space,number of machine-tools and / or equipment, human resource, layout,production planning, logistics (supply-chain and material handling,automation. The software was designed especially for automotive industry,industry that dominates the Brasov area. Present paper is focused on thegeneral structure of the CADOProSys (Computer Aided Design andOptimisation of Production Systems Software, including the logicdiagram and some opportunities provided by the application.

  14. Fun and software exploring pleasure, paradox and pain in computing

    CERN Document Server

    Goriunova, Olga

    2014-01-01

    Fun and Software offers the untold story of fun as constitutive of the culture and aesthetics of computing. Fun in computing is a mode of thinking, making and experiencing. It invokes and convolutes the question of rationalism and logical reason, addresses the sensibilities and experience of computation and attests to its creative drives. By exploring topics as diverse as the pleasure and pain of the programmer, geek wit, affects of play and coding as a bodily pursuit of the unique in recursive structures, Fun and Software helps construct a different point of entry to the understanding of soft

  15. The HEP Software and Computing Knowledge Base

    Science.gov (United States)

    Wenaus, T.

    2017-10-01

    HEP software today is a rich and diverse domain in itself and exists within the mushrooming world of open source software. As HEP software developers and users we can be more productive and effective if our work and our choices are informed by a good knowledge of what others in our community have created or found useful. The HEP Software and Computing Knowledge Base, hepsoftware.org, was created to facilitate this by serving as a collection point and information exchange on software projects and products, services, training, computing facilities, and relating them to the projects, experiments, organizations and science domains that offer them or use them. It was created as a contribution to the HEP Software Foundation, for which a HEP S&C knowledge base was a much requested early deliverable. This contribution will motivate and describe the system, what it offers, its content and contributions both existing and needed, and its implementation (node.js based web service and javascript client app) which has emphasized ease of use for both users and contributors.

  16. 48 CFR 227.7203-10 - Contractor identification and marking of computer software or computer software documentation to...

    Science.gov (United States)

    2010-10-01

    ... and marking of computer software or computer software documentation to be furnished with restrictive... REGULATIONS SYSTEM, DEPARTMENT OF DEFENSE GENERAL CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Computer Software and Computer Software Documentation 227.7203-10 Contractor identification and...

  17. Application Software Structure Enables NIF Operations

    Energy Technology Data Exchange (ETDEWEB)

    Fong, K W; Estes, C M; Fisher, J M; Shelton, R T

    2001-10-17

    The NIF Integrated Computer Control System (ICCS) application software uses a set of service frameworks that assures uniform behavior spanning the front-end processors (FEPs) and supervisor programs. This uniformity is visible both in the way each program employs shared services and in the flexibility it affords for attaching graphical user interfaces (GUIs). Uniformity of structure across applications is desired for the benefit of programmers who will be maintaining the many programs that constitute the ICCS. In this paper, the framework components that have the greatest impact on the application structure are discussed.

  18. Computing platforms for software-defined radio

    CERN Document Server

    Nurmi, Jari; Isoaho, Jouni; Garzia, Fabio

    2017-01-01

    This book addresses Software-Defined Radio (SDR) baseband processing from the computer architecture point of view, providing a detailed exploration of different computing platforms by classifying different approaches, highlighting the common features related to SDR requirements and by showing pros and cons of the proposed solutions. Coverage includes architectures exploiting parallelism by extending single-processor environment (such as VLIW, SIMD, TTA approaches), multi-core platforms distributing the computation to either a homogeneous array or a set of specialized heterogeneous processors, and architectures exploiting fine-grained, coarse-grained, or hybrid reconfigurability. Describes a computer engineering approach to SDR baseband processing hardware; Discusses implementation of numerous compute-intensive signal processing algorithms on single and multicore platforms; Enables deep understanding of optimization techniques related to power and energy consumption of multicore platforms using several basic a...

  19. 78 FR 47015 - Software Requirement Specifications for Digital Computer Software Used in Safety Systems of...

    Science.gov (United States)

    2013-08-02

    ... COMMISSION Software Requirement Specifications for Digital Computer Software Used in Safety Systems of... 1 of RG 1.172, ``Software Requirement Specifications for Digital Computer Software used in Safety... (IEEE) Standard (Std.) 830-1998, ``IEEE Recommended Practice for Software Requirements Specifications...

  20. Computational structural mechanics for engine structures

    Science.gov (United States)

    Chamis, C. C.

    1989-01-01

    The computational structural mechanics (CSM) program at Lewis encompasses: (1) fundamental aspects for formulating and solving structural mechanics problems, and (2) development of integrated software systems to computationally simulate the performance/durability/life of engine structures. It is structured to mainly supplement, complement, and whenever possible replace, costly experimental efforts which are unavoidable during engineering research and development programs. Specific objectives include: investigate unique advantages of parallel and multiprocesses for: reformulating/solving structural mechanics and formulating/solving multidisciplinary mechanics and develop integrated structural system computational simulators for: predicting structural performances, evaluating newly developed methods, and for identifying and prioritizing improved/missing methods needed. Herein the CSM program is summarized with emphasis on the Engine Structures Computational Simulator (ESCS). Typical results obtained using ESCS are described to illustrate its versatility.

  1. Computer software for process hazards analysis.

    Science.gov (United States)

    Hyatt, N

    2000-10-01

    Computerized software tools are assuming major significance in conducting HAZOPs. This is because they have the potential to offer better online presentations and performance to HAZOP teams, as well as better documentation and downstream tracking. The chances of something being "missed" are greatly reduced. We know, only too well, that HAZOP sessions can be like the industrial equivalent of a trip to the dentist. Sessions can (and usually do) become arduous and painstaking. To make the process easier for all those involved, we need all the help computerized software can provide. In this paper I have outlined the challenges addressed in the production of Windows software for performing HAZOP and other forms of PHA. The object is to produce more "intelligent", more user-friendly software for performing HAZOP where technical interaction between team members is of key significance. HAZOP techniques, having already proven themselves, are extending into the field of computer control and human error. This makes further demands on HAZOP software and emphasizes its importance.

  2. Software agents in molecular computational biology.

    Science.gov (United States)

    Keele, John W; Wray, James E

    2005-12-01

    Progress made in applying agent systems to molecular computational biology is reviewed and strategies by which to exploit agent technology to greater advantage are investigated. Communities of software agents could play an important role in helping genome scientists design reagents for future research. The advent of genome sequencing in cattle and swine increases the complexity of data analysis required to conduct research in livestock genomics. Databases are always expanding and semantic differences among data are common. Agent platforms have been developed to deal with generic issues such as agent communication, life cycle management and advertisement of services (white and yellow pages). This frees computational biologists from the drudgery of having to re-invent the wheel on these common chores, giving them more time to focus on biology and bioinformatics. Agent platforms that comply with the Foundation for Intelligent Physical Agents (FIPA) standards are able to interoperate. In other words, agents developed on different platforms can communicate and cooperate with one another if domain-specific higher-level communication protocol details are agreed upon between different agent developers. Many software agent platforms are peer-to-peer, which means that even if some of the agents and data repositories are temporarily unavailable, a subset of the goals of the system can still be met. Past use of software agents in bioinformatics indicates that an agent approach should prove fruitful. Examination of current problems in bioinformatics indicates that existing agent platforms should be adaptable to novel situations.

  3. The December 2006 ATLAS Computing & Software Workshop

    CERN Multimedia

    Fred Luehring

    The 29th ATLAS Computing & Software Workshop was held on December 11-15 at CERN. With the rapidly approaching onset of data taking, the workshop participants had an air of urgency about them. There was considerable discussion on hot topics such as physics validation of the software, data analysis, actual software production on the GRID, and the schedule of work for 2007 including the Final Dress Rehearsal (FDR). However don't be fooled, the workshop was not all work - there were also two social events which were greatly enjoyed by the attendees. The workshop welcomed Wouter Verkerke as the new Physics Validation Coordinator (replacing Davide Costanzo). Most recent validation work has centered on the 12.0.X release series that will be used for the Computing System Commissioning (CSC) exercise. The validation is now a big job because it needs to be done over a variety of conditions (magnetic field on/off, aligned/misaligned geometry) for every candidate release. Luckily there have been a large number of pe...

  4. Software for computing and annotating genomic ranges.

    Directory of Open Access Journals (Sweden)

    Michael Lawrence

    Full Text Available We describe Bioconductor infrastructure for representing and computing on annotated genomic ranges and integrating genomic data with the statistical computing features of R and its extensions. At the core of the infrastructure are three packages: IRanges, GenomicRanges, and GenomicFeatures. These packages provide scalable data structures for representing annotated ranges on the genome, with special support for transcript structures, read alignments and coverage vectors. Computational facilities include efficient algorithms for overlap and nearest neighbor detection, coverage calculation and other range operations. This infrastructure directly supports more than 80 other Bioconductor packages, including those for sequence analysis, differential expression analysis and visualization.

  5. 77 FR 50726 - Software Requirement Specifications for Digital Computer Software and Complex Electronics Used in...

    Science.gov (United States)

    2012-08-22

    ... COMMISSION Software Requirement Specifications for Digital Computer Software and Complex Electronics Used in... issuing for public comment draft regulatory guide (DG), DG-1209, ``Software Requirement Specifications for Digital Computer Software and Complex Electronics used in Safety Systems of Nuclear Power Plants.'' The DG...

  6. Delivering LHC software to HPC compute elements

    CERN Document Server

    Blomer, Jakob; Hardi, Nikola; Popescu, Radu

    2017-01-01

    In recent years, there was a growing interest in improving the utilization of supercomputers by running applications of experiments at the Large Hadron Collider (LHC) at CERN when idle cores cannot be assigned to traditional HPC jobs. At the same time, the upcoming LHC machine and detector upgrades will produce some 60 times higher data rates and challenge LHC experiments to use so far untapped compute resources. LHC experiment applications are tailored to run on high-throughput computing resources and they have a different anatomy than HPC applications. LHC applications comprise a core framework that allows hundreds of researchers to plug in their specific algorithms. The software stacks easily accumulate to many gigabytes for a single release. New releases are often produced on a daily basis. To facilitate the distribution of these software stacks to world-wide distributed computing resources, LHC experiments use a purpose-built, global, POSIX file system, the CernVM File System. CernVM-FS pre-processes dat...

  7. Computer- Aided Design in Power Engineering Application of Software Tools

    CERN Document Server

    Stojkovic, Zlatan

    2012-01-01

    This textbooks demonstrates the application of software tools in solving a series of problems from the field of designing power system structures and systems. It contains four chapters: The first chapter leads the reader through all the phases necessary in the procedures of computer aided modeling and simulation. It guides through the complex problems presenting on the basis of eleven original examples. The second chapter presents  application of software tools in power system calculations of power systems equipment design. Several design example calculations are carried out using engineering standards like MATLAB, EMTP/ATP, Excel & Access, AutoCAD and Simulink. The third chapters focuses on the graphical documentation using a collection of software tools (AutoCAD, EPLAN, SIMARIS SIVACON, SIMARIS DESIGN) which enable the complete automation of the development of graphical documentation of a power systems. In the fourth chapter, the application of software tools in the project management in power systems ...

  8. 48 CFR 212.7003 - Technical data and computer software.

    Science.gov (United States)

    2010-10-01

    ... computer software. 212.7003 Section 212.7003 Federal Acquisition Regulations System DEFENSE ACQUISITION... data and computer software. For purposes of establishing delivery requirements and license rights for technical data under 227.7102 and for computer software under 227.7202, there shall be a rebuttable...

  9. 14 CFR 415.123 - Computing systems and software.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Computing systems and software. 415.123... Launch Vehicle From a Non-Federal Launch Site § 415.123 Computing systems and software. (a) An applicant's safety review document must describe all computing systems and software that perform a safety...

  10. Computer Software and Copyright Law. CRS Report for Congress.

    Science.gov (United States)

    Weimer, Douglas Reid

    The purpose of this report is to provide an overview of copyright law as it relates to computer software. The computer industry terminology, case law, and software protection by patents and trade secret legislation are examined, and it is argued that the protection of ownership rights of computer software poses major challenges to existing U.S.…

  11. Computer imaging software for profile photograph analysis.

    Science.gov (United States)

    Tollefson, Travis T; Sykes, Jonathan M

    2007-01-01

    To describe a novel calibration technique for photographs of different sizes and to test a new method of chin evaluation in relation to established analysis measurements. A photograph analysis and medical record review of 14 patients who underwent combined rhinoplasty and chin correction at an academic center. Patients undergoing concurrent orthognathic surgery, rhytidectomy, or submental liposuction were excluded. Preoperative and postoperative digital photographs were analyzed using computer imaging software with a new method, the soft tissue porion to pogonion distance, and with established measurements, including the cervicomental angle, the mentocervical angle, and the facial convexity angle. The porion to pogonion distance consistently increased after the chin correction procedure (more in the osseous group). All photograph angle measurements changed toward the established normal range postoperatively. Surgery for facial disharmony requires artistic judgment and objective evaluation. Although 3-dimensional video analysis of the face seems promising, its clinical use is limited by cost. For surgeons who use computer imaging software, analysis of profile photographs is the most valuable tool. Even when preoperative and postoperative photographs are of different sizes, relative distance comparisons are possible with a new calibration technique using the constant facial landmarks, the porion and the pupil. The porion-pogonion distance is a simple reproducible measurement that can be used along with established soft tissue measurements as a guide for profile facial analysis.

  12. 48 CFR 227.7203-14 - Conformity, acceptance, and warranty of computer software and computer software documentation.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 3 2010-10-01 2010-10-01 false Conformity, acceptance... Software Documentation 227.7203-14 Conformity, acceptance, and warranty of computer software and computer...) Conformity and acceptance. Solicitations and contracts requiring the delivery of computer software shall...

  13. Development of an Aero-Optics Software Library and Integration into Structured Overset and Unstructured Computational Fluid Dynamics (CFD) Flow Solvers

    Science.gov (United States)

    2011-04-01

    Computational Fluid Dynamics ( CFD ) Flow Solvers William J. Coirier and James Stutts KRATOS/Digital Fusion, Inc. April 2011 Final Report...Integration into Structured Overset and Unstructured Computational Fluid Dynamics ( CFD ) Flow Solvers 5a. CONTRACT NUMBER In-house 5b. GRANT

  14. CMS Computing Software and Analysis Challenge 2006

    Science.gov (United States)

    De Filippis, N.; CMS Collaboration

    2007-10-01

    The CMS (Compact Muon Solenoid) collaboration is making a big effort to test the workflow and the dataflow associated with the data handling model. With this purpose the computing, software and analysis Challenge 2006, namely CSA06, started the 15th of September. It was a 50 million event exercise that included all the steps of the analysis chain, like the prompt reconstruction, the data streaming, calibration and alignment iterative executions, the data distribution to regional sites, up to the end-user analysis. Grid tools provided by the LCG project are also experimented to gain access to the data and the resources by providing a user friendly interface to the physicists submitting the production and the analysis jobs. An overview of the status and results of the CSA06 is presented in this work.

  15. 48 CFR 27.405-3 - Commercial computer software.

    Science.gov (United States)

    2010-10-01

    ... software. 27.405-3 Section 27.405-3 Federal Acquisition Regulations System FEDERAL ACQUISITION REGULATION... Commercial computer software. (a) When contracting other than from GSA's Multiple Award Schedule contracts for the acquisition of commercial computer software, no specific contract clause prescribed in this...

  16. Data structure and software engineering challenges and improvements

    CERN Document Server

    Antonakos, James L

    2011-01-01

    Data structure and software engineering is an integral part of computer science. This volume presents new approaches and methods to knowledge sharing, brain mapping, data integration, and data storage. The author describes how to manage an organization's business process and domain data and presents new software and hardware testing methods. The book introduces a game development framework used as a learning aid in a software engineering at the university level. It also features a review of social software engineering metrics and methods for processing business information. It explains how to

  17. The benefit of introducing audit software into curricula for computer ...

    African Journals Online (AJOL)

    The use of computers in the learning process is a well-researched area. The introduction of computers and related audit software in the auditing field has had a major influence on the auditing process. Very little research has been done on the inclusion of computer audit software in the auditing syllabus. Even less research ...

  18. Computer, Network, Software, and Hardware Engineering with Applications

    CERN Document Server

    Schneidewind, Norman F

    2012-01-01

    There are many books on computers, networks, and software engineering but none that integrate the three with applications. Integration is important because, increasingly, software dominates the performance, reliability, maintainability, and availability of complex computer and systems. Books on software engineering typically portray software as if it exists in a vacuum with no relationship to the wider system. This is wrong because a system is more than software. It is comprised of people, organizations, processes, hardware, and software. All of these components must be considered in an integr

  19. Computers boost structural technology

    Science.gov (United States)

    Noor, Ahmed K.; Venneri, Samuel L.

    1989-01-01

    Derived from matrix methods of structural analysis and finite element methods developed over the last three decades, computational structures technology (CST) blends computer science, numerical analysis, and approximation theory into structural analysis and synthesis. Recent significant advances in CST include stochastic-based modeling, strategies for performing large-scale structural calculations on new computing systems, and the integration of CST with other disciplinary modules for multidisciplinary analysis and design. New methodologies have been developed at NASA for integrated fluid-thermal structural analysis and integrated aerodynamic-structure-control design. The need for multiple views of data for different modules also led to the development of a number of sophisticated data-base management systems. For CST to play a role in the future development of structures technology and in the multidisciplinary design of future flight vehicles, major advances and computational tools are needed in a number of key areas.

  20. Infinite possibilities: Computational structures technology

    Science.gov (United States)

    Beam, Sherilee F.

    1994-01-01

    Computational Fluid Dynamics (or CFD) methods are very familiar to the research community. Even the general public has had some exposure to CFD images, primarily through the news media. However, very little attention has been paid to CST--Computational Structures Technology. Yet, no important design can be completed without it. During the first half of this century, researchers only dreamed of designing and building structures on a computer. Today their dreams have become practical realities as computational methods are used in all phases of design, fabrication and testing of engineering systems. Increasingly complex structures can now be built in even shorter periods of time. Over the past four decades, computer technology has been developing, and early finite element methods have grown from small in-house programs to numerous commercial software programs. When coupled with advanced computing systems, they help engineers make dramatic leaps in designing and testing concepts. The goals of CST include: predicting how a structure will behave under actual operating conditions; designing and complementing other experiments conducted on a structure; investigating microstructural damage or chaotic, unpredictable behavior; helping material developers in improving material systems; and being a useful tool in design systems optimization and sensitivity techniques. Applying CST to a structure problem requires five steps: (1) observe the specific problem; (2) develop a computational model for numerical simulation; (3) develop and assemble software and hardware for running the codes; (4) post-process and interpret the results; and (5) use the model to analyze and design the actual structure. Researchers in both industry and academia continue to make significant contributions to advance this technology with improvements in software, collaborative computing environments and supercomputing systems. As these environments and systems evolve, computational structures technology will

  1. Science Computer Software. A Handbook on Selection and Classroom Use.

    Science.gov (United States)

    Good, Ron

    Designed for use by science teachers, this document provides selected lists of computer software collected by the Science and Mathematics Software (SAMS) lab at Florida State University. The commercially available software is compiled under the subject areas of biology, chemistry, and physics. Each list provides: (1) recommended grade level; (2)…

  2. 14 CFR 417.123 - Computing systems and software.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Computing systems and software. 417.123... systems and software. (a) A launch operator must document a system safety process that identifies the... systems and software. (b) A launch operator must identify all safety-critical functions associated with...

  3. A Novel Coupling Pattern in Computational Science and Engineering Software

    Science.gov (United States)

    Computational science and engineering (CSE) software is written by experts of certain area(s). Due to the specialization,existing CSE software may need to integrate other CSE software systems developed by different groups of experts. Thecoupling problem is one of the challenges f...

  4. Software engineering frameworks for the cloud computing paradigm

    CERN Document Server

    Mahmood, Zaigham

    2013-01-01

    This book presents the latest research on Software Engineering Frameworks for the Cloud Computing Paradigm, drawn from an international selection of researchers and practitioners. The book offers both a discussion of relevant software engineering approaches and practical guidance on enterprise-wide software deployment in the cloud environment, together with real-world case studies. Features: presents the state of the art in software engineering approaches for developing cloud-suitable applications; discusses the impact of the cloud computing paradigm on software engineering; offers guidance an

  5. Software Synthesis for High Productivity Exascale Computing

    Energy Technology Data Exchange (ETDEWEB)

    Bodik, Rastislav [Univ. of Washington, Seattle, WA (United States)

    2010-09-01

    Over the three years of our project, we accomplished three key milestones: We demonstrated how ideas from generative programming and software synthesis can help support the development of bulk-synchronous distributed memory kernels. These ideas are realized in a new language called MSL, a C-like language that combines synthesis features with high level notations for array manipulation and bulk-synchronous parallelism to simplify the semantic analysis required for synthesis. We also demonstrated that these high level notations map easily to low level C code and show that the performance of this generated code matches that of handwritten Fortran. Second, we introduced the idea of solver-aided domain-specific languages (SDSLs), which are an emerging class of computer-aided programming systems. SDSLs ease the construction of programs by automating tasks such as verification, debugging, synthesis, and non-deterministic execution. SDSLs are implemented by translating the DSL program into logical constraints. Next, we developed a symbolic virtual machine called Rosette, which simplifies the construction of such SDSLs and their compilers. We have used Rosette to build SynthCL, a subset of OpenCL that supports synthesis. Third, we developed novel numeric algorithms that move as little data as possible, either between levels of a memory hierarchy or between parallel processors over a network. We achieved progress in three aspects of this problem. First we determined lower bounds on communication. Second, we compared these lower bounds to widely used versions of these algorithms, and noted that these widely used algorithms usually communicate asymptotically more than is necessary. Third, we identified or invented new algorithms for most linear algebra problems that do attain these lower bounds, and demonstrated large speed-ups in theory and practice.

  6. 77 FR 50722 - Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2012-08-22

    ... COMMISSION Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants... regulatory guide (DG), DG-1208, ``Software Unit Testing for Digital Computer Software used in Safety Systems... revision endorses, with clarifications, the enhanced consensus practices for testing of computer software...

  7. Infinite possibilities: Computational structures technology

    Science.gov (United States)

    Beam, Sherilee F.

    1994-12-01

    Computational Fluid Dynamics (or CFD) methods are very familiar to the research community. Even the general public has had some exposure to CFD images, primarily through the news media. However, very little attention has been paid to CST--Computational Structures Technology. Yet, no important design can be completed without it. During the first half of this century, researchers only dreamed of designing and building structures on a computer. Today their dreams have become practical realities as computational methods are used in all phases of design, fabrication and testing of engineering systems. Increasingly complex structures can now be built in even shorter periods of time. Over the past four decades, computer technology has been developing, and early finite element methods have grown from small in-house programs to numerous commercial software programs. When coupled with advanced computing systems, they help engineers make dramatic leaps in designing and testing concepts. The goals of CST include: predicting how a structure will behave under actual operating conditions; designing and complementing other experiments conducted on a structure; investigating microstructural damage or chaotic, unpredictable behavior; helping material developers in improving material systems; and being a useful tool in design systems optimization and sensitivity techniques. Applying CST to a structure problem requires five steps: (1) observe the specific problem; (2) develop a computational model for numerical simulation; (3) develop and assemble software and hardware for running the codes; (4) post-process and interpret the results; and (5) use the model to analyze and design the actual structure. Researchers in both industry and academia continue to make significant contributions to advance this technology with improvements in software, collaborative computing environments and supercomputing systems. As these environments and systems evolve, computational structures technology will

  8. Survey and analyses of computer software usage in Calabar ...

    African Journals Online (AJOL)

    Use of computer in Nigeria has become a house hold phenomenon. Calabar metropolis is one of the areas where computers are used for many activities. This work is to find out the most used software and the type of jobs mostly done. A descriptive analysis using simple percentages revealed that word processing software ...

  9. Assessment of Computer Software Usage for Estimating and Tender ...

    African Journals Online (AJOL)

    The objectives of this research is to evaluate the level of computer software usage for estimating and tender analysis while also assessing the challenges faced by Nigerian Quantity Surveyors in using computer software for estimating and tender analysis. The survey design methodology was adopted for this research.

  10. Value Engineering: An Application to Computer Software

    Science.gov (United States)

    1995-06-01

    software to effectively implement VE in software acquisition. To add to this problem, the DOD Value Engineering program as a whole has had a history ...of VE in sortware acquisitionfdevelopment (ie. educacion , award programs, designate Govt. savings for use in 77 generating additional savings

  11. Evaluation of Software for Introducing Protein Structure: Visualization and Simulation

    Science.gov (United States)

    White, Brian; Kahriman, Azmin; Luberice, Lois; Idleh, Farhia

    2010-01-01

    Communicating an understanding of the forces and factors that determine a protein's structure is an important goal of many biology and biochemistry courses at a variety of levels. Many educators use computer software that allows visualization of these complex molecules for this purpose. Although visualization is in wide use and has been associated…

  12. Computer-Aided Software Engineering - An approach to real-time software development

    Science.gov (United States)

    Walker, Carrie K.; Turkovich, John J.

    1989-01-01

    A new software engineering discipline is Computer-Aided Software Engineering (CASE), a technology aimed at automating the software development process. This paper explores the development of CASE technology, particularly in the area of real-time/scientific/engineering software, and a history of CASE is given. The proposed software development environment for the Advanced Launch System (ALS CASE) is described as an example of an advanced software development system for real-time/scientific/engineering (RT/SE) software. The Automated Programming Subsystem of ALS CASE automatically generates executable code and corresponding documentation from a suitably formatted specification of the software requirements. Software requirements are interactively specified in the form of engineering block diagrams. Several demonstrations of the Automated Programming Subsystem are discussed.

  13. Discrete computational structures

    CERN Document Server

    Korfhage, Robert R

    1974-01-01

    Discrete Computational Structures describes discrete mathematical concepts that are important to computing, covering necessary mathematical fundamentals, computer representation of sets, graph theory, storage minimization, and bandwidth. The book also explains conceptual framework (Gorn trees, searching, subroutines) and directed graphs (flowcharts, critical paths, information network). The text discusses algebra particularly as it applies to concentrates on semigroups, groups, lattices, propositional calculus, including a new tabular method of Boolean function minimization. The text emphasize

  14. Software design for resilient computer systems

    CERN Document Server

    Schagaev, Igor

    2016-01-01

    This book addresses the question of how system software should be designed to account for faults, and which fault tolerance features it should provide for highest reliability. The authors first show how the system software interacts with the hardware to tolerate faults. They analyze and further develop the theory of fault tolerance to understand the different ways to increase the reliability of a system, with special attention on the role of system software in this process. They further develop the general algorithm of fault tolerance (GAFT) with its three main processes: hardware checking, preparation for recovery, and the recovery procedure. For each of the three processes, they analyze the requirements and properties theoretically and give possible implementation scenarios and system software support required. Based on the theoretical results, the authors derive an Oberon-based programming language with direct support of the three processes of GAFT. In the last part of this book, they introduce a simulator...

  15. JSMAA: open source software for SMAA computations

    Science.gov (United States)

    Tervonen, Tommi

    2014-01-01

    Most software for multi-criteria decision analysis (MCDA) implement a small set of compatible methods as a closed monolithic program. With such software tools, the decision models have to be input by hand. In some applications, however, the model can be generated using external information sources, and thus it would be beneficial if the MCDA software could integrate in the comprehensive information infrastructure. This article motivates for the need of model generation in the methodological context of stochastic multicriteria acceptability analysis (SMAA), and describes the JSMAA software that implements SMAA-2, SMAA-O and SMAA-TRI methods. JSMAA is an open source and divided in separate graphical user interface and library components, enabling its use in systems with a model generation subsystem.

  16. Bioconductor: open software development for computational biology and bioinformatics

    DEFF Research Database (Denmark)

    Gentleman, R.C.; Carey, V.J.; Bates, D.M.

    2004-01-01

    The Bioconductor project is an initiative for the collaborative creation of extensible software for computational biology and bioinformatics. The goals of the project include: fostering collaborative development and widespread use of innovative software, reducing barriers to entry into interdisci......The Bioconductor project is an initiative for the collaborative creation of extensible software for computational biology and bioinformatics. The goals of the project include: fostering collaborative development and widespread use of innovative software, reducing barriers to entry...... into interdisciplinary scientific research, and promoting the achievement of remote reproducibility of research results. We describe details of our aims and methods, identify current challenges, compare Bioconductor to other open bioinformatics projects, and provide working examples....

  17. Computer organization and design the hardware/software interface

    CERN Document Server

    Hennessy, John L

    1994-01-01

    Computer Organization and Design: The Hardware/Software Interface presents the interaction between hardware and software at a variety of levels, which offers a framework for understanding the fundamentals of computing. This book focuses on the concepts that are the basis for computers.Organized into nine chapters, this book begins with an overview of the computer revolution. This text then explains the concepts and algorithms used in modern computer arithmetic. Other chapters consider the abstractions and concepts in memory hierarchies by starting with the simplest possible cache. This book di

  18. 78 FR 47011 - Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2013-08-02

    ... COMMISSION Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants..., ``Software Unit Testing for Digital Computer Software Used in Safety Systems of Nuclear Power Plants.'' This... (ANSI/IEEE) Standard (Std.) 1008-1987, ``IEEE Standard for Software Unit Testing'' with the...

  19. 76 Computer Assisted Language Learning (CALL) Software ...

    African Journals Online (AJOL)

    Ike Odimegwu

    It is clearly inadequate to consider the design of the software in isolation from its pedagogical setting. It is noted that some of the most innovative educational activities occur when teachers and students actually subvert the designer's intention, and these interactions are necessarily complex in nature. Furthermore, Anderson ...

  20. Micro-computer simulation software: A review

    Directory of Open Access Journals (Sweden)

    P.S. Kruger

    2003-12-01

    Full Text Available Simulation modelling has proved to be one of the most powerful tools available to the Operations Research Analyst. The development of micro-computer technology has reached a state of maturity where the micro-computer can provide the necessary computing power and consequently various powerful and inexpensive simulation languages for micro-computers have became available. This paper will attempt to provide an introduction to the general philosophy and characteristics of some of the available micro-computer simulation languages. The emphasis will be on the characteristics of the specific micro-computer implementation rather than on a comparison of the modelling features of the various languages. Such comparisons may be found elsewhere.

  1. Test case generation for On Board Computer Software Major Components

    OpenAIRE

    Dinesha H A; V.K Agrawal

    2013-01-01

    PES Institute of Technology along with five other institutions is developing a student imaging satellite. In imaging satellite development, software implementation plays an important role. On Board Computer (OBC) is one which runs the satellite software. Satellite software has its sub components like telemetry, control modes, data processing, actuator, tele command and etc. Design, development and testing of these components are done successfully. Testing plays very important role to assure i...

  2. Caesy: A software tool for computer-aided engineering

    Science.gov (United States)

    Wette, Matt

    1993-01-01

    A new software tool, Caesy, is described. This tool provides a strongly typed programming environment for research in the development of algorithms and software for computer-aided control system design. A description of the user language and its implementation as they currently stand are presented along with a description of work in progress and areas of future work.

  3. Imprinting Community College Computer Science Education with Software Engineering Principles

    Science.gov (United States)

    Hundley, Jacqueline Holliday

    2012-01-01

    Although the two-year curriculum guide includes coverage of all eight software engineering core topics, the computer science courses taught in Alabama community colleges limit student exposure to the programming, or coding, phase of the software development lifecycle and offer little experience in requirements analysis, design, testing, and…

  4. The experiential modification of a computer software package for ...

    African Journals Online (AJOL)

    Erna Kinsey

    The experiential modification of a computer software package for graphing algebraic functions. J.G. Maree, S. Scholtz,* H.J. Botha and S. Van Putten. School of Teacher Training, Faculty of Education, University of Pretoria, Pretoria, 0002 South Africa. * To whom correspondence should be addressed. Graphing software and ...

  5. Mathematics Software, A Computer Lab, and the Hearing Impaired.

    Science.gov (United States)

    Castle, Dan J.

    1982-01-01

    A teacher of deaf high school students describes considerations in developing a computer lab, including equipment selection, class and subject scheduling, and suitable software. Examples of software used in his lab are cited, including Mathware's Math City and Apple's SUPERMATH program. (CL)

  6. Software Safety Risk in Legacy Safety-Critical Computer Systems

    Science.gov (United States)

    Hill, Janice; Baggs, Rhoda

    2007-01-01

    Safety-critical computer systems must be engineered to meet system and software safety requirements. For legacy safety-critical computer systems, software safety requirements may not have been formally specified during development. When process-oriented software safety requirements are levied on a legacy system after the fact, where software development artifacts don't exist or are incomplete, the question becomes 'how can this be done?' The risks associated with only meeting certain software safety requirements in a legacy safety-critical computer system must be addressed should such systems be selected as candidates for reuse. This paper proposes a method for ascertaining formally, a software safety risk assessment, that provides measurements for software safety for legacy systems which may or may not have a suite of software engineering documentation that is now normally required. It relies upon the NASA Software Safety Standard, risk assessment methods based upon the Taxonomy-Based Questionnaire, and the application of reverse engineering CASE tools to produce original design documents for legacy systems.

  7. Onboard Computers, Onboard Software and Satellite Operations An Introduction

    CERN Document Server

    Eickhoff, Jens

    2012-01-01

    This book is intended as a system engineer's compendium, explaining the dependencies and technical interactions between the onboard computer hardware, the onboard software and the spacecraft operations from ground. After a brief introduction on the subsequent development in all three fields over the spacecraft engineering phases each of the main topis is treated in depth in a separate part. The features of today’s onboard computers are explained at hand of their historic evolution over the decades from the early days of spaceflight up to today. Latest system-on-chip processor architectures are treated as well as all onboard computer major components. After the onboard computer hardware the corresponding software is treated in a separate part. Both the software static architecture as well as the dynamic architecture are covered, and development technologies as well as software verification approaches are included. Following these two parts on the onboard architecture, the last part covers the concepts of spa...

  8. Computer-assisted software transcription of qualitative interviews.

    Science.gov (United States)

    Alcock, John; Iphofen, Ron

    2007-01-01

    John Alcock and Ron Iphofen examine a method to assist with interview transcription using free, open source computer software and digital recordings, and consider some underlying practical, ethical and philosophical issues.

  9. CMS Software and Computing Ready for Run 2

    CERN Document Server

    Bloom, Kenneth

    2015-01-01

    In Run 1 of the Large Hadron Collider, software and computing was a strategic strength of the Compact Muon Solenoid experiment. The timely processing of data and simulation samples and the excellent performance of the reconstruction algorithms played an important role in the preparation of the full suite of searches used for the observation of the Higgs boson in 2012. In Run 2, the LHC will run at higher intensities and CMS will record data at a higher trigger rate. These new running conditions will provide new challenges for the software and computing systems. Over the two years of Long Shutdown 1, CMS has built upon the successes of Run 1 to improve the software and computing to meet these challenges. In this presentation we will describe the new features in software and computing that will once again put CMS in a position of physics leadership.

  10. Software Defined Radio Datalink Implementation Using PC-Type Computers

    National Research Council Canada - National Science Library

    Zafeiropoulos, Georgios

    2003-01-01

    The objective of this thesis was to examine the feasibility of implementation and the performance of a Software Defined Radio datalink, using a common PC type host computer and a high level programming language...

  11. AASERT: Software Tools for Experimentation in Computational Geometry

    National Research Council Canada - National Science Library

    Dobkin, David

    2001-01-01

    This research has considered problems in computer graphics and visualization. The work has aimed to bring theoretical tools to practical problems as well as to develop tools with which to aid in the building of geometric software...

  12. Reviews of computing technology: Software overview

    Energy Technology Data Exchange (ETDEWEB)

    Hartshorn, W.R.; Johnson, A.L.

    1994-01-05

    The Savannah River Site Computing Architecture states that the site computing environment will be standards-based, data-driven, and workstation-oriented. Larger server systems deliver needed information to users in a client-server relationship. Goals of the Architecture include utilizing computing resources effectively, maintaining a high level of data integrity, developing a robust infrastructure, and storing data in such a way as to promote accessibility and usability. This document describes the current storage environment at Savannah River Site (SRS) and presents some of the problems that will be faced and strategies that are planned over the next few years.

  13. Super instruction architecture of petascale electronic structure software: the story

    Science.gov (United States)

    Lotrich, V. F.; Ponton, J. M.; Perera, A. S.; Deumens, E.; Bartlett, R. J.; Sanders, B. A.

    2010-11-01

    Theoretical methods in chemistry lead to algorithms for the computation of electronic energies and other properties of electronic wave functions that require large numbers of floating point operations and involve large data sets. Thus, computational chemists are very interested in using massively parallel computer systems and in particular the new petascale systems. In this paper we discuss a new programming paradigm that was developed at the Quantum Theory Project to construct electronic structure software that can scale to large numbers of cores of the order of 100,000 and beyond to solve problems in materials engineering relevant to the problems facing society today.

  14. Software for Computing the Tetrachoric Correlation Coefficient

    OpenAIRE

    Rubén Daniel Ledesma; Guillermo Macbeth; Pedro Valero-Mora

    2011-01-01

    La correlación tetracórica es un caso particular de análisis de correlación entre variables continuas distribuidas normalmente pero que han sido medidas en formato dicotómico. En este artículo se presenta ViSta-Tetrachor, un software gratuito que emplea una aproximación al coeficiente de correlación tetracórica caracterizada por ser más fácil de computar que las ecuaciones originales, así como más eficiente que otras aproximaciones propuestas. ViSta-Tetrachor proporciona estimaciones puntuale...

  15. CMS software and computing for LHC Run 2

    CERN Document Server

    INSPIRE-00067576

    2016-11-09

    The CMS offline software and computing system has successfully met the challenge of LHC Run 2. In this presentation, we will discuss how the entire system was improved in anticipation of increased trigger output rate, increased rate of pileup interactions and the evolution of computing technology. The primary goals behind these changes was to increase the flexibility of computing facilities where ever possible, as to increase our operational efficiency, and to decrease the computing resources needed to accomplish the primary offline computing workflows. These changes have resulted in a new approach to distributed computing in CMS for Run 2 and for the future as the LHC luminosity should continue to increase. We will discuss changes and plans to our data federation, which was one of the key changes towards a more flexible computing model for Run 2. Our software framework and algorithms also underwent significant changes. We will summarize the our experience with a new multi-threaded framework as deployed on ou...

  16. Web Solutions Inspire Cloud Computing Software

    Science.gov (United States)

    2013-01-01

    An effort at Ames Research Center to standardize NASA websites unexpectedly led to a breakthrough in open source cloud computing technology. With the help of Rackspace Inc. of San Antonio, Texas, the resulting product, OpenStack, has spurred the growth of an entire industry that is already employing hundreds of people and generating hundreds of millions in revenue.

  17. Computer Assisted Language Learning (CALL) Software: Evaluation ...

    African Journals Online (AJOL)

    Evaluating the nature and extent of the influence of Computer Assisted Language Learning (CALL) on the quality of language learning is highly problematic. This is owing to the number and complexity of interacting variables involved in setting the items for teaching and learning languages. This paper identified and ...

  18. Developments in Computer Aided Software Maintenance

    Science.gov (United States)

    1974-09-01

    differences in intelligence ( Adorno et al, 1950). The present authors speculate, however, that there may be a correlation between programming ability and...Administrative Side." Computer Decisions, October 1973. Adorno , T.W., Frenkel-Brunswick, E., Levinson, D.J., and Sanford, R.N. The Authoritarian

  19. What Do Computer Science Students Think about Software Piracy?

    Science.gov (United States)

    Konstantakis, Nikos I.; Palaigeorgiou, George E.; Siozos, Panos D.; Tsoukalas, Ioannis A.

    2010-01-01

    Today, software piracy is an issue of global importance. Computer science students are the future information and communication technologies professionals and it is important to study the way they approach this issue. In this article, we attempt to study attitudes, behaviours and the corresponding reasoning of computer science students in Greece…

  20. General-Purpose Software For Computer Graphics

    Science.gov (United States)

    Rogers, Joseph E.

    1992-01-01

    NASA Device Independent Graphics Library (NASADIG) is general-purpose computer-graphics package for computer-based engineering and management applications which gives opportunity to translate data into effective graphical displays for presentation. Features include two- and three-dimensional plotting, spline and polynomial interpolation, control of blanking of areas, multiple log and/or linear axes, control of legends and text, control of thicknesses of curves, and multiple text fonts. Included are subroutines for definition of areas and axes of plots; setup and display of text; blanking of areas; setup of style, interpolation, and plotting of lines; control of patterns and of shading of colors; control of legends, blocks of text, and characters; initialization of devices; and setting of mixed alphabets. Written in FORTRAN 77.

  1. Software for Computing the Tetrachoric Correlation Coefficient

    Directory of Open Access Journals (Sweden)

    Rubén Daniel Ledesma

    2011-01-01

    Full Text Available La correlación tetracórica es un caso particular de análisis de correlación entre variables continuas distribuidas normalmente pero que han sido medidas en formato dicotómico. En este artículo se presenta ViSta-Tetrachor, un software gratuito que emplea una aproximación al coeficiente de correlación tetracórica caracterizada por ser más fácil de computar que las ecuaciones originales, así como más eficiente que otras aproximaciones propuestas. ViSta-Tetrachor proporciona estimaciones puntuales e intervalos de confianza para el estadístico. También permite generar matrices de correlación tetracórica y aplicar un análisis factorial a estas matrices. Se presenta una breve descripción del programa junto con diversos ejemplos de aplicación.

  2. Computer hardware and software for robotic control

    Science.gov (United States)

    Davis, Virgil Leon

    1987-01-01

    The KSC has implemented an integrated system that coordinates state-of-the-art robotic subsystems. It is a sensor based real-time robotic control system performing operations beyond the capability of an off-the-shelf robot. The integrated system provides real-time closed loop adaptive path control of position and orientation of all six axes of a large robot; enables the implementation of a highly configurable, expandable testbed for sensor system development; and makes several smart distributed control subsystems (robot arm controller, process controller, graphics display, and vision tracking) appear as intelligent peripherals to a supervisory computer coordinating the overall systems.

  3. Leveraging software architectures to guide and verify the development of sense/compute/control applications

    DEFF Research Database (Denmark)

    Cassou, Damien; Balland, Emilie; Consel, Charles

    2011-01-01

    A software architecture describes the structure of a computing system by specifying software components and their interactions. Mapping a software architecture to an implementation is a well known challenge. A key element of this mapping is the architecture’s description of the data and control...... interaction specifications. We introduce a notion of interaction contract that expresses allowed interactions between components, describing both data and control-flow constraints. This declaration is part of the architecture description, allows generation of extensive programming support, and enables various...... verifications. We instantiate our approach in an architecture description language for Sense/Compute/Control applications, and describe associated compilation and verification strategies....

  4. Integrated Hardware and Software for No-Loss Computing

    Science.gov (United States)

    James, Mark

    2007-01-01

    When an algorithm is distributed across multiple threads executing on many distinct processors, a loss of one of those threads or processors can potentially result in the total loss of all the incremental results up to that point. When implementation is massively hardware distributed, then the probability of a hardware failure during the course of a long execution is potentially high. Traditionally, this problem has been addressed by establishing checkpoints where the current state of some or part of the execution is saved. Then in the event of a failure, this state information can be used to recompute that point in the execution and resume the computation from that point. A serious problem arises when one distributes a problem across multiple threads and physical processors is that one increases the likelihood of the algorithm failing due to no fault of the scientist but as a result of hardware faults coupled with operating system problems. With good reason, scientists expect their computing tools to serve them and not the other way around. What is novel here is a unique combination of hardware and software that reformulates an application into monolithic structure that can be monitored in real-time and dynamically reconfigured in the event of a failure. This unique reformulation of hardware and software will provide advanced aeronautical technologies to meet the challenges of next-generation systems in aviation, for civilian and scientific purposes, in our atmosphere and in atmospheres of other worlds. In particular, with respect to NASA s manned flight to Mars, this technology addresses the critical requirements for improving safety and increasing reliability of manned spacecraft.

  5. Workshop on Software Development Tools for Petascale Computing

    Energy Technology Data Exchange (ETDEWEB)

    Vetter, Jeffrey [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Georgia Inst. of Technology, Atlanta, GA (United States)

    2007-08-01

    Petascale computing systems will soon be available to the DOE science community. Recent studies in the productivity of HPC platforms point to better software environments as a key enabler to science on these systems. To prepare for the deployment and productive use of these petascale platforms, the DOE science and general HPC community must have the software development tools, such as performance analyzers and debuggers that meet application requirements for scalability, functionality, reliability, and ease of use. In this report, we identify and prioritize the research opportunities in the area of software development tools for high performance computing. To facilitate this effort, DOE hosted a group of 55 leading international experts in this area at the Software Development Tools for PetaScale Computing (SDTPC) Workshop, which was held in Washington, D.C. on August 1 and 2, 2007. Software development tools serve as an important interface between the application teams and the target HPC architectures. Broadly speaking, these roles can be decomposed into three categories: performance tools, correctness tools, and development environments. Accordingly, this SDTPC report has four technical thrusts: performance tools, correctness tools, development environment infrastructures, and scalable tool infrastructures. The last thrust primarily targets tool developers per se, rather than end users. Finally, this report identifies non-technical strategic challenges that impact most tool development. The organizing committee emphasizes that many critical areas are outside the scope of this charter; these important areas include system software, compilers, and I/O.

  6. Service-oriented Software Defined Optical Networks for Cloud Computing

    Science.gov (United States)

    Liu, Yuze; Li, Hui; Ji, Yuefeng

    2017-10-01

    With the development of big data and cloud computing technology, the traditional software-defined network is facing new challenges (e.g., ubiquitous accessibility, higher bandwidth, more flexible management and greater security). This paper proposes a new service-oriented software defined optical network architecture, including a resource layer, a service abstract layer, a control layer and an application layer. We then dwell on the corresponding service providing method. Different service ID is used to identify the service a device can offer. Finally, we experimentally evaluate that proposed service providing method can be applied to transmit different services based on the service ID in the service-oriented software defined optical network.

  7. Copyright protection of computer software and its patentability

    OpenAIRE

    Kalíšek, Jindřich

    2009-01-01

    | Autorskoprávní ochrana počítačových programů a jejich patentovatelnost | |94| Diploma Thesis Abstract Author Jindřich Kalíšek Faculty Faculty of Law Tutor JUDr. Tomáš Dobřichovský, Ph.D. Work Title The Copyright Protection for Software in Accordance to its Patenting Attachments Count 2 (P. A, P. B) Letters Count ≈ 258 830 inc. spaces and footnotes Information Sources Count 90 (thereof 63 available online) Keywords (7) computer program, software, copyright, software copyright, patent, patent...

  8. Computer Software Configuration Item-Specific Flight Software Image Transfer Script Generator

    Science.gov (United States)

    Bolen, Kenny; Greenlaw, Ronald

    2010-01-01

    A K-shell UNIX script enables the International Space Station (ISS) Flight Control Team (FCT) operators in NASA s Mission Control Center (MCC) in Houston to transfer an entire or partial computer software configuration item (CSCI) from a flight software compact disk (CD) to the onboard Portable Computer System (PCS). The tool is designed to read the content stored on a flight software CD and generate individual CSCI transfer scripts that are capable of transferring the flight software content in a given subdirectory on the CD to the scratch directory on the PCS. The flight control team can then transfer the flight software from the PCS scratch directory to the Electronically Erasable Programmable Read Only Memory (EEPROM) of an ISS Multiplexer/ Demultiplexer (MDM) via the Indirect File Transfer capability. The individual CSCI scripts and the CSCI Specific Flight Software Image Transfer Script Generator (CFITSG), when executed a second time, will remove all components from their original execution. The tool will identify errors in the transfer process and create logs of the transferred software for the purposes of configuration management.

  9. A design methodology for portable software on parallel computers

    Science.gov (United States)

    Nicol, David M.; Miller, Keith W.; Chrisman, Dan A.

    1993-01-01

    This final report for research that was supported by grant number NAG-1-995 documents our progress in addressing two difficulties in parallel programming. The first difficulty is developing software that will execute quickly on a parallel computer. The second difficulty is transporting software between dissimilar parallel computers. In general, we expect that more hardware-specific information will be included in software designs for parallel computers than in designs for sequential computers. This inclusion is an instance of portability being sacrificed for high performance. New parallel computers are being introduced frequently. Trying to keep one's software on the current high performance hardware, a software developer almost continually faces yet another expensive software transportation. The problem of the proposed research is to create a design methodology that helps designers to more precisely control both portability and hardware-specific programming details. The proposed research emphasizes programming for scientific applications. We completed our study of the parallelizability of a subsystem of the NASA Earth Radiation Budget Experiment (ERBE) data processing system. This work is summarized in section two. A more detailed description is provided in Appendix A ('Programming Practices to Support Eventual Parallelism'). Mr. Chrisman, a graduate student, wrote and successfully defended a Ph.D. dissertation proposal which describes our research associated with the issues of software portability and high performance. The list of research tasks are specified in the proposal. The proposal 'A Design Methodology for Portable Software on Parallel Computers' is summarized in section three and is provided in its entirety in Appendix B. We are currently studying a proposed subsystem of the NASA Clouds and the Earth's Radiant Energy System (CERES) data processing system. This software is the proof-of-concept for the Ph.D. dissertation. We have implemented and measured

  10. Application of software technology to a future spacecraft computer design

    Science.gov (United States)

    Labaugh, R. J.

    1980-01-01

    A study was conducted to determine how major improvements in spacecraft computer systems can be obtained from recent advances in hardware and software technology. Investigations into integrated circuit technology indicated that the CMOS/SOS chip set being developed for the Air Force Avionics Laboratory at Wright Patterson had the best potential for improving the performance of spaceborne computer systems. An integral part of the chip set is the bit slice arithmetic and logic unit. The flexibility allowed by microprogramming, combined with the software investigations, led to the specification of a baseline architecture and instruction set.

  11. Structure and software tools of AIDA.

    Science.gov (United States)

    Duisterhout, J S; Franken, B; Witte, F

    1987-01-01

    AIDA consists of a set of software tools to allow for fast development and easy-to-maintain Medical Information Systems. AIDA supports all aspects of such a system both during development and operation. It contains tools to build and maintain forms for interactive data entry and on-line input validation, a database management system including a data dictionary and a set of run-time routines for database access, and routines for querying the database and output formatting. Unlike an application generator, the user of AIDA may select parts of the tools to fulfill his needs and program other subsystems not developed with AIDA. The AIDA software uses as host language the ANSI-standard programming language MUMPS, an interpreted language embedded in an integrated database and programming environment. This greatly facilitates the portability of AIDA applications. The database facilities supported by AIDA are based on a relational data model. This data model is built on top of the MUMPS database, the so-called global structure. This relational model overcomes the restrictions of the global structure regarding string length. The global structure is especially powerful for sorting purposes. Using MUMPS as a host language allows the user an easy interface between user-defined data validation checks or other user-defined code and the AIDA tools. AIDA has been designed primarily for prototyping and for the construction of Medical Information Systems in a research environment which requires a flexible approach. The prototyping facility of AIDA operates terminal independent and is even to a great extent multi-lingual. Most of these features are table-driven; this allows on-line changes in the use of terminal type and language, but also causes overhead. AIDA has a set of optimizing tools by which it is possible to build a faster, but (of course) less flexible code from these table definitions. By separating the AIDA software in a source and a run-time version, one is able to write

  12. SOFTWARE AGENT AND CLOUD COMPUTING: A BRIEF REVIEW

    Directory of Open Access Journals (Sweden)

    Firas D. Ahmed

    2016-02-01

    Full Text Available The merging of interests between Cloud applications which necessary require an intelligent software agent with elastic, dynamic, with independent behavior ability and multi-agent systems that need consistent distributed infrastructures can be resulted with new effective applications and systems. Cloud computing services offered by using large-scale infrastructures with elastic services and high-performance capability since these resources could be adjusted to end user and application needs. Cloud systems and infrastructures are offered by a service-oriented interface that provides computing resources using X-as-a-service model to introduce cloud services on the pay-per-use model. Agent-based system is significant for the improving the use of software agents for boosting cloud service composition, service discovery, negotiation mechanism and several domains. Integrating these two computing paradigms enables cloud-computing systems to become more elastic, autonomic, and intelligent service's capability. Meanwhile, scalable systems with high-performance on the cloud are capable of providing MASs with a consistent and large-scale computing infrastructure on which to execute large-scale systems with flexibility. The significant of this paper is introducing and discussing potential benefits of integration multi-agent technology with Cloud computing systems. This reviewing will lead having in attention the aim of developing high-performance and sophisticated applications with intelligent using software agents and Cloud paradigm.

  13. Computer-assisted qualitative data analysis software: a review.

    Science.gov (United States)

    Banner, Davina J; Albarrran, John W

    2009-01-01

    Over recent decades, qualitative research has become accepted as a uniquely valuable methodological approach for generating knowledge, particularly in relation to promoting understanding of patients' experiences and responses to illness. Within cardiovascular nursing such qualitative approaches have been widely adopted to systematically investigate a number of phenomena. Contemporary qualitative research practice comprises a diverse range of disciplines and approaches. Computer-aided qualitative data analysis software represents an important facet of this increasingly sophisticated movement. Such software offers an efficient means through which to manage and organize data while supporting rigorous data analysis. The increasing use of qualitative data analysis software has stimulated wide discussion. This research column includes a review of some of the advantages and debates related to the use and integration of qualitative data analysis software.

  14. Approximator: Predicting Interruptibility in Software Development with Commodity Computers

    DEFF Research Database (Denmark)

    Tell, Paolo; Jalaliniya, Shahram; Andersen, Kristian S. M.

    2015-01-01

    Assessing the presence and availability of a remote colleague is key in coordination in global software development but is not easily done using existing computer-mediated channels. Previous research has shown that automated estimation of interruptibility is feasible and can achieve a precision...

  15. A community Q&A for HEP Software and Computing ?

    CERN Multimedia

    CERN. Geneva

    2016-01-01

    How often do you use StackOverflow or ServerFault to find information in your daily work? Would you be interested in a community Q&A site for HEP Software and Computing, for instance a dedicated StackExchange site? I looked into this question...

  16. Versatile Software for the Computation of Kinetic Profiles of Intricate ...

    African Journals Online (AJOL)

    NJD

    Simkine 2 is a user-friendly software package developed with the Delphi 5 programming language, for the computation of complex reaction profiles based on a proposed kinetic scheme for the chemical system. The programme applies the semi-implicit extrapolation method (SIEM), which utilizes the implicit midpoint rule ...

  17. Simkine 2 – Versatile Software for the Computation of Kinetic ...

    African Journals Online (AJOL)

    Simkine 2 is a user-friendly software package developed with the Delphi 5 programming language, for the computation of complex reaction profiles based on a proposed kinetic scheme for the chemical system. The programme applies the semi-implicit extrapolation method (SIEM), which utilizes the implicit midpoint rule ...

  18. Legal Protection of Computer Software: An Industrial Survey.

    Science.gov (United States)

    Miller, Richard I.

    A survey was commissioned to establish a baseline on the present modes of legal protection employed by the computer software industry. Questionnaires were distributed to 308 member companies of the Association of Data Processing Service Organizations (ADAPSO), and 116 responses were returned. Results indicated that (1) business executives…

  19. Development of the Law of Computer Software Protection.

    Science.gov (United States)

    Nimtz, Robert O.

    1979-01-01

    Traces the history of the development of the law dealing with the protection of computer software. The available forms of protection are the patent, copyright, and trade secret laws. Available from Business Manager, P. O. Box 2600, Arlington, Virginia 22202; sc $1.25. (Author/IRT)

  20. A Software Framework for Multimodal Human-Computer Interaction Systems

    NARCIS (Netherlands)

    Shen, Jie; Pantic, Maja

    2009-01-01

    This paper describes a software framework we designed and implemented for the development and research in the area of multimodal human-computer interface. The proposed framework is based on publish / subscribe architecture, which allows developers and researchers to conveniently configure, test and

  1. StructuralComponents : A software system for conceptual structural design

    NARCIS (Netherlands)

    Van de Weerd, B.; Rolvink, A.; Coenders, J.L.

    2012-01-01

    Conceptual design is the starting point of the design process. The conceptual design stage comprises the formation of several ideas or design concepts to meet the imposed constraints. StructuralComponents is a software application that attempts to provide the designing engineer with a suitable set

  2. Human-Centered Software Engineering: Software Engineering Architectures, Patterns, and Sodels for Human Computer Interaction

    Science.gov (United States)

    Seffah, Ahmed; Vanderdonckt, Jean; Desmarais, Michel C.

    The Computer-Human Interaction and Software Engineering (CHISE) series of edited volumes originated from a number of workshops and discussions over the latest research and developments in the field of Human Computer Interaction (HCI) and Software Engineering (SE) integration, convergence and cross-pollination. A first volume in this series (CHISE Volume I - Human-Centered Software Engineering: Integrating Usability in the Development Lifecycle) aims at bridging the gap between the field of SE and HCI, and addresses specifically the concerns of integrating usability and user-centered systems design methods and tools into the software development lifecycle and practices. This has been done by defining techniques, tools and practices that can fit into the entire software engineering lifecycle as well as by defining ways of addressing the knowledge and skills needed, and the attitudes and basic values that a user-centered development methodology requires. The first volume has been edited as Vol. 8 in the Springer HCI Series (Seffah, Gulliksen and Desmarais, 2005).

  3. Mission Management Computer Software for RLV-TD

    Science.gov (United States)

    Manju, C. R.; Joy, Josna Susan; Vidya, L.; Sheenarani, I.; Sruthy, C. N.; Viswanathan, P. C.; Dinesh, Sudin; Jayalekshmy, L.; Karuturi, Kesavabrahmaji; Sheema, E.; Syamala, S.; Unnikrishnan, S. Manju; Ali, S. Akbar; Paramasivam, R.; Sheela, D. S.; Shukkoor, A. Abdul; Lalithambika, V. R.; Mookiah, T.

    2017-12-01

    The Mission Management Computer (MMC) software is responsible for the autonomous navigation, sequencing, guidance and control of the Re-usable Launch Vehicle (RLV), through lift-off, ascent, coasting, re-entry, controlled descent and splashdown. A hard real-time system has been designed for handling the mission requirements in an integrated manner and for meeting the stringent timing constraints. Redundancy management and fault-tolerance techniques are also built into the system, in order to achieve a successful mission even in presence of component failures. This paper describes the functions and features of the components of the MMC software which has accomplished the successful RLV-Technology Demonstrator mission.

  4. Mission Management Computer Software for RLV-TD

    Science.gov (United States)

    Manju, C. R.; Joy, Josna Susan; Vidya, L.; Sheenarani, I.; Sruthy, C. N.; Viswanathan, P. C.; Dinesh, Sudin; Jayalekshmy, L.; Karuturi, Kesavabrahmaji; Sheema, E.; Syamala, S.; Unnikrishnan, S. Manju; Ali, S. Akbar; Paramasivam, R.; Sheela, D. S.; Shukkoor, A. Abdul; Lalithambika, V. R.; Mookiah, T.

    2017-10-01

    The Mission Management Computer (MMC) software is responsible for the autonomous navigation, sequencing, guidance and control of the Re-usable Launch Vehicle (RLV), through lift-off, ascent, coasting, re-entry, controlled descent and splashdown. A hard real-time system has been designed for handling the mission requirements in an integrated manner and for meeting the stringent timing constraints. Redundancy management and fault-tolerance techniques are also built into the system, in order to achieve a successful mission even in presence of component failures. This paper describes the functions and features of the components of the MMC software which has accomplished the successful RLV-Technology Demonstrator mission.

  5. Software for Distributed Computation on Medical Databases: A Demonstration Project

    Directory of Open Access Journals (Sweden)

    Balasubramanian Narasimhan

    2017-05-01

    Full Text Available Bringing together the information latent in distributed medical databases promises to personalize medical care by enabling reliable, stable modeling of outcomes with rich feature sets (including patient characteristics and treatments received. However, there are barriers to aggregation of medical data, due to lack of standardization of ontologies, privacy concerns, proprietary attitudes toward data, and a reluctance to give up control over end use. Aggregation of data is not always necessary for model fitting. In models based on maximizing a likelihood, the computations can be distributed, with aggregation limited to the intermediate results of calculations on local data, rather than raw data. Distributed fitting is also possible for singular value decomposition. There has been work on the technical aspects of shared computation for particular applications, but little has been published on the software needed to support the "social networking" aspect of shared computing, to reduce the barriers to collaboration. We describe a set of software tools that allow the rapid assembly of a collaborative computational project, based on the flexible and extensible R statistical software and other open source packages, that can work across a heterogeneous collection of database environments, with full transparency to allow local officials concerned with privacy protections to validate the safety of the method. We describe the principles, architecture, and successful test results for the site-stratified Cox model and rank-k singular value decomposition.

  6. Software Safety Risk in Legacy Safety-Critical Computer Systems

    Science.gov (United States)

    Hill, Janice L.; Baggs, Rhoda

    2007-01-01

    Safety Standards contain technical and process-oriented safety requirements. Technical requirements are those such as "must work" and "must not work" functions in the system. Process-Oriented requirements are software engineering and safety management process requirements. Address the system perspective and some cover just software in the system > NASA-STD-8719.13B Software Safety Standard is the current standard of interest. NASA programs/projects will have their own set of safety requirements derived from the standard. Safety Cases: a) Documented demonstration that a system complies with the specified safety requirements. b) Evidence is gathered on the integrity of the system and put forward as an argued case. [Gardener (ed.)] c) Problems occur when trying to meet safety standards, and thus make retrospective safety cases, in legacy safety-critical computer systems.

  7. [Design and application of computer-assisted software for teaching and research of population genetics].

    Science.gov (United States)

    Gao, Jing; Pan, Shen-Yuan; Cao, Jing

    2008-05-01

    The computer-assisted software for population genetics' teaching and research is designed with Microsoft Visual Basic 6.0. It takes all factors affecting population genetic structure into account. By way of setting various conditional parameters, calculating theoretically, simulating with computer and choosing the states of show, the system displays the results in charts and tables accurately. It also presents distribution pattern, statistical characteristics and the change of population gene frequencies and genotype frequencies in the inter-generation visually. The software is a mature system for teaching and researches on the aspects of its functions, operations and interface.

  8. Shade matching assisted by digital photography and computer software.

    Science.gov (United States)

    Schropp, Lars

    2009-04-01

    To evaluate the efficacy of digital photographs and graphic computer software for color matching compared to conventional visual matching. The shade of a tab from a shade guide (Vita 3D-Master Guide) placed in a phantom head was matched to a second guide of the same type by nine observers. This was done for twelve selected shade tabs (tests). The shade-matching procedure was performed visually in a simulated clinic environment and with digital photographs, and the time spent for both procedures was recorded. An alternative arrangement of the shade tabs was used in the digital photographs. In addition, a graphic software program was used for color analysis. Hue, chroma, and lightness values of the test tab and all tabs of the second guide were derived from the digital photographs. According to the CIE L*C*h* color system, the color differences between the test tab and tabs of the second guide were calculated. The shade guide tab that deviated least from the test tab was determined to be the match. Shade matching performance by means of graphic software was compared with the two visual methods and tested by Chi-square tests (alpha= 0.05). Eight of twelve test tabs (67%) were matched correctly by the computer software method. This was significantly better (p < 0.02) than the performance of the visual shade matching methods conducted in the simulated clinic (32% correct match) and with photographs (28% correct match). No correlation between time consumption for the visual shade matching methods and frequency of correct match was observed. Shade matching assisted by digital photographs and computer software was significantly more reliable than by conventional visual methods.

  9. Enhancing the Understanding of Computer Networking Courses through Software Tools

    OpenAIRE

    Dafalla, Z. I.; Balaji, R. D.

    2015-01-01

    Computer networking is an important specialization in Information and Communication Technologies. However imparting the right knowledge to students can be a challenging task due to the fact that there is not enough time to deliver lengthy labs during normal lecture hours. Augmenting the use of physical machines with software tools help the students to learn beyond the limited lab sessions within the environment of higher Institutions of learning throughout the world. The Institutions focus mo...

  10. Improving Software Performance in the Compute Unified Device Architecture

    OpenAIRE

    Alexandru PIRJAN

    2010-01-01

    This paper analyzes several aspects regarding the improvement of software performance for applications written in the Compute Unified Device Architecture CUDA). We address an issue of great importance when programming a CUDA application: the Graphics Processing Unit’s (GPU’s) memory management through ranspose ernels. We also benchmark and evaluate the performance for progressively optimizing a transposing matrix application in CUDA. One particular interest was to research how well the op...

  11. Software

    Energy Technology Data Exchange (ETDEWEB)

    Macedo, R.; Budd, G.; Ross, E.; Wells, P.

    2010-07-15

    The software section of this journal presented new software programs that have been developed to help in the exploration and development of hydrocarbon resources. Software provider IHS Inc. has made additions to its geological and engineering analysis software tool, IHS PETRA, a product used by geoscientists and engineers to visualize, analyze and manage well production, well log, drilling, reservoir, seismic and other related information. IHS PETRA also includes a directional well module and a decline curve analysis module to improve analysis capabilities in unconventional reservoirs. Petris Technology Inc. has developed a software to help manage the large volumes of data. PetrisWinds Enterprise (PWE) helps users find and manage wellbore data, including conventional wireline and MWD core data; analysis core photos and images; waveforms and NMR; and external files documentation. Ottawa-based Ambercore Software Inc. has been collaborating with Nexen on the Petroleum iQ software for steam assisted gravity drainage (SAGD) producers. Petroleum iQ integrates geology and geophysics data with engineering data in 3D and 4D. Calgary-based Envirosoft Corporation has developed a software that reduces the costly and time-consuming effort required to comply with Directive 39 of the Alberta Energy Resources Conservation Board. The product includes an emissions modelling software. Houston-based Seismic Micro-Technology (SMT) has developed the Kingdom software that features the latest in seismic interpretation. Holland-based Joa Oil and Gas and Calgary-based Computer Modelling Group have both supplied the petroleum industry with advanced reservoir simulation software that enables reservoir interpretation. The 2010 software survey included a guide to new software applications designed to facilitate petroleum exploration, drilling and production activities. Oil and gas producers can use the products for a range of functions, including reservoir characterization and accounting. In

  12. Asynchronous replica exchange software for grid and heterogeneous computing

    Science.gov (United States)

    Gallicchio, Emilio; Xia, Junchao; Flynn, William F.; Zhang, Baofeng; Samlalsingh, Sade; Mentes, Ahmet; Levy, Ronald M.

    2015-11-01

    Parallel replica exchange sampling is an extended ensemble technique often used to accelerate the exploration of the conformational ensemble of atomistic molecular simulations of chemical systems. Inter-process communication and coordination requirements have historically discouraged the deployment of replica exchange on distributed and heterogeneous resources. Here we describe the architecture of a software (named ASyncRE) for performing asynchronous replica exchange molecular simulations on volunteered computing grids and heterogeneous high performance clusters. The asynchronous replica exchange algorithm on which the software is based avoids centralized synchronization steps and the need for direct communication between remote processes. It allows molecular dynamics threads to progress at different rates and enables parameter exchanges among arbitrary sets of replicas independently from other replicas. ASyncRE is written in Python following a modular design conducive to extensions to various replica exchange schemes and molecular dynamics engines. Applications of the software for the modeling of association equilibria of supramolecular and macromolecular complexes on BOINC campus computational grids and on the CPU/MIC heterogeneous hardware of the XSEDE Stampede supercomputer are illustrated. They show the ability of ASyncRE to utilize large grids of desktop computers running the Windows, MacOS, and/or Linux operating systems as well as collections of high performance heterogeneous hardware devices.

  13. DSM Software for Computing Synthetic Seismograms in Transversely Isotropic Spherically Symmetric Media and Its Application

    Science.gov (United States)

    Kawai, K.; Takeuchi, N.; Geller, R. J.

    2002-12-01

    The existence of anisotropy has been suggested in many regions in the Earth. Determining the anisotropic seismic velocity structure of the Earth can contribute to our understanding of geodynamics and rheology. Inversion of observed seismic waveforms is a promising approach for determining the Earth's anisotropic structure, but development of computational algorithms and software for computing synthetic seismograms in anisotropic media is required. Software for computing seismic waveforms in isotropic media based on the Direct Solution Method (DSM; Geller and Ohminato 1994, GJI) has previously been developed and is being used in data analysis, but DSM software for computing synthetic seismograms for anisotropic media has not yet been developed. In this study, we derive algorithms and develop software for computing synthetics for transversely isotropic spherically symmetric media. Our derivation follows previous work for isotropic media (Takeuchi et al. 1996, GRL; Cummins et al. 1997, GJI). The displacement is represented using spherical harmonics for the lateral dependence and linear spline functions for the vertical dependence of the trial functions. The numerical operators derived using these trial functions are then replaced by optimally accurate operators (Geller and Takeuchi 1995, GJI; Takeuchi and Geller 2002, GJI, submitted). Although the number of elastic constants increases from 2 to 5, the numerical operators are basically identical to those for the isotropic case. Our derivation does not require approximations that treat the anisotropic or laterally heterogeneous structure as an infinitesimal perturbation to the isotropic structure. Only spherically symmetric models are considered in this paper, but when our methods can be extended to the 3-D case to permit computation of synthetic seismograms with the same accuracy as for spherically symmetric isotropic models. We present computational examples such as accuracy checks and also some applications to

  14. The Phenix software for automated determination of macromolecular structures.

    Science.gov (United States)

    Adams, Paul D; Afonine, Pavel V; Bunkóczi, Gábor; Chen, Vincent B; Echols, Nathaniel; Headd, Jeffrey J; Hung, Li-Wei; Jain, Swati; Kapral, Gary J; Grosse Kunstleve, Ralf W; McCoy, Airlie J; Moriarty, Nigel W; Oeffner, Robert D; Read, Randy J; Richardson, David C; Richardson, Jane S; Terwilliger, Thomas C; Zwart, Peter H

    2011-09-01

    X-ray crystallography is a critical tool in the study of biological systems. It is able to provide information that has been a prerequisite to understanding the fundamentals of life. It is also a method that is central to the development of new therapeutics for human disease. Significant time and effort are required to determine and optimize many macromolecular structures because of the need for manual interpretation of complex numerical data, often using many different software packages, and the repeated use of interactive three-dimensional graphics. The Phenix software package has been developed to provide a comprehensive system for macromolecular crystallographic structure solution with an emphasis on automation. This has required the development of new algorithms that minimize or eliminate subjective input in favor of built-in expert-systems knowledge, the automation of procedures that are traditionally performed by hand, and the development of a computational framework that allows a tight integration between the algorithms. The application of automated methods is particularly appropriate in the field of structural proteomics, where high throughput is desired. Features in Phenix for the automation of experimental phasing with subsequent model building, molecular replacement, structure refinement and validation are described and examples given of running Phenix from both the command line and graphical user interface. Copyright © 2011 Elsevier Inc. All rights reserved.

  15. Mahotas: Open source software for scriptable computer vision

    Directory of Open Access Journals (Sweden)

    Luis Pedro Coelho

    2013-07-01

    Full Text Available Mahotas is a computer vision library for Python. It contains traditional image processing functionality such as filtering and morphological operations as well as more modern computer vision functions for feature computation, including interest point detection and local descriptors. The interface is in Python, a dynamic programming language, which is appropriate for fast development, but the algorithms are implemented in C++ and are tuned for speed. The library is designed to fit in with the scientific software ecosystem in this language and can leverage the existing infrastructure developed in that language. Mahotas is released under a liberal open source license (MIT License and is available from http://github.com/luispedro/mahotas and from the Python Package Index (http://pypi.python.org/pypi/mahotas. Tutorials and full API documentation are available online at http://mahotas.readthedocs.org/.

  16. Cloud Computing as an Enabler of Agile Global Software Development

    Directory of Open Access Journals (Sweden)

    Maureen Tanner

    2016-05-01

    Full Text Available Agile global software development (AGSD is an increasingly prevalent software development strategy, as organizations hope to realize the benefits of accessing a larger resource pool of skilled labor, at a potentially reduced cost, while at the same time delivering value incrementally and iteratively. However, the distributed nature of AGSD creates geographic, temporal, socio-cultural distances that challenge collaboration between project stakeholders. The Cloud Computing (CC service models of Infrastructure as a Service (IaaS, Platform as a Service (PaaS, and Software as a Service (SaaS are similar to the aspirant qualities of AGSD as they provide services that are globally accessible, efficient, and stable, with lower predictable operating costs that scale to meet the computational demand. This study focused on the 12 agile principles upon which all agile methodologies are based, therein potentially increasing the potential for the findings to be generalized. Domestication Theory was used to assist in understanding how cloud technologies were appropriated in support of AGSD. The research strategy took the form of case study research. The findings suggest that some of the challenges in applying the agile principles in AGSD may be overcome by using CC.

  17. Accuracy study of new computer-assisted orthopedic surgery software

    Energy Technology Data Exchange (ETDEWEB)

    Sidon, Eli [Department of Orthopaedic Surgery, Beilinson-Rabin Medical Center, Petach Tikva (Israel); Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv (Israel); Steinberg, Ely L., E-mail: steinberge@tasmc.health.gov.il [Department of Orthopaedic Surgery, Tel-Aviv Medical Center, Tel Aviv (Israel); Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv (Israel)

    2012-12-15

    Purpose: The new computerized system is based on image analysis and designed to aid in orthopedic surgeries by virtual trajectory of the guide wire, intra-operative planning and various measurements. Validation of the accuracy and safety of any computer-aided surgery system is essential before implementing it clinically. We examined the accuracy of guide-wire length and angle measurements and fusion of multiple adjacent images (panoramic view image, PVI{sup ®}) of the new software. Methods: This is a 2-part study. Part I: twenty guide wires were drilled to various depths in a synthetic femur model and the results obtained by the software measurements were compared with manual measurements by a caliper and a depth gauge. Part II: a sawbone femur shaft was osteotomized and various inclinations of >10° to the varus or valgus angles were tested. The manually obtained measurements of angles and lengths were compared to the new computerized system software PVI. Results: There was a significant positive linear correlation between all groups of the computerized length and the control measurements (r > 0.983, p < 0.01). There was no significant difference among different distances, angles or positions from the image intensifier. There was a significant positive linear correlation between the angle and length measurement on the PVI and the control measurement (r > 0.993, p < 0.01). Conclusions: The new computerized software has high reliability in performing measurements of length using an aiming, positioning and referring device intra-operatively.

  18. IEEE Computer Society/Software Engineering Institute Watts S. Humphrey Software Process Achievement (SPA) Award 2016: Nationwide

    Science.gov (United States)

    2017-04-05

    REV-03.18.2016.0 IEEE Computer Society/Software Engineering Institute Watts S. Humphrey Software Process Achievement (SPA) Award 2016...20503. 1. AGENCY USE ONLY (Leave Blank) 2. REPORT DATE April 2017 3. REPORT TYPE AND DATES COVERED Final 4. TITLE AND SUBTITLE IEEE Computer

  19. Software Defects, Scientific Computation and the Scientific Method

    CERN Multimedia

    CERN. Geneva

    2011-01-01

    Computation has rapidly grown in the last 50 years so that in many scientific areas it is the dominant partner in the practice of science. Unfortunately, unlike the experimental sciences, it does not adhere well to the principles of the scientific method as espoused by, for example, the philosopher Karl Popper. Such principles are built around the notions of deniability and reproducibility. Although much research effort has been spent on measuring the density of software defects, much less has been spent on the more difficult problem of measuring their effect on the output of a program. This talk explores these issues with numerous examples suggesting how this situation might be improved to match the demands of modern science. Finally it develops a theoretical model based on an amalgam of statistical mechanics and Hartley/Shannon information theory which suggests that software systems have strong implementation independent behaviour and supports the widely observed phenomenon that defects clust...

  20. COMPUTATIONAL MODELLING OF BUFFETING EFFECTS USING OPENFOAM SOFTWARE PACKAGE

    Directory of Open Access Journals (Sweden)

    V. T. Kalugin

    2015-01-01

    Full Text Available In this paper, the preliminary results of computational modeling of an aircraft with the airbrake deployed are presented. The calculations were performed with OpenFOAM software package. The results outlined are a part of a research project to optimise aircraft performance using a perforated airbrake. Within this stage of the project OpenFOAM software package with hybrid RANS-LES approach was tested in respect to a given configuration of the aircraft, airbrake and then has been compared with the test data. For the worst case the amplitude of the peak force acting on the tail fin can be up to 6 times higher than the average value without airbrake deployed. To reduce unsteady loads acting on the tailfin, perforation of the airbrake was proposed.

  1. Scientific Computation Application Partnerships in Materials and Chemical Sciences, Charge Transfer and Charge Transport in Photoactivated Systems, Developing Electron-Correlated Methods for Excited State Structure and Dynamics in the NWChem Software Suite

    Energy Technology Data Exchange (ETDEWEB)

    Cramer, Christopher J. [Univ. of Minnesota, Minneapolis, MN (United States)

    2017-11-12

    Charge transfer and charge transport in photoactivated systems are fundamental processes that underlie solar energy capture, solar energy conversion, and photoactivated catalysis, both organometallic and enzymatic. We developed methods, algorithms, and software tools needed for reliable treatment of the underlying physics for charge transfer and charge transport, an undertaking with broad applicability to the goals of the fundamental-interaction component of the Department of Energy Office of Basic Energy Sciences and the exascale initiative of the Office of Advanced Scientific Computing Research.

  2. Improving Software Performance in the Compute Unified Device Architecture

    Directory of Open Access Journals (Sweden)

    Alexandru PIRJAN

    2010-01-01

    Full Text Available This paper analyzes several aspects regarding the improvement of software performance for applications written in the Compute Unified Device Architecture CUDA. We address an issue of great importance when programming a CUDA application: the Graphics Processing Unit’s (GPU’s memory management through ranspose ernels. We also benchmark and evaluate the performance for progressively optimizing a transposing matrix application in CUDA. One particular interest was to research how well the optimization techniques, applied to software application written in CUDA, scale to the latest generation of general-purpose graphic processors units (GPGPU, like the Fermi architecture implemented in the GTX480 and the previous architecture implemented in GTX280. Lately, there has been a lot of interest in the literature for this type of optimization analysis, but none of the works so far (to our best knowledge tried to validate if the optimizations can apply to a GPU from the latest Fermi architecture and how well does the Fermi architecture scale to these software performance improving techniques.

  3. Using a software-defined computer in teaching the basics of computer architecture and operation

    Science.gov (United States)

    Kosowska, Julia; Mazur, Grzegorz

    2017-08-01

    The paper describes the concept and implementation of SDC_One software-defined computer designed for experimental and didactic purposes. Equipped with extensive hardware monitoring mechanisms, the device enables the students to monitor the computer's operation on bus transfer cycle or instruction cycle basis, providing the practical illustration of basic aspects of computer's operation. In the paper, we describe the hardware monitoring capabilities of SDC_One and some scenarios of using it in teaching the basics of computer architecture and microprocessor operation.

  4. High Energy Physics Forum for Computational Excellence: Working Group Reports (I. Applications Software II. Software Libraries and Tools III. Systems)

    Energy Technology Data Exchange (ETDEWEB)

    Habib, Salman [Argonne National Lab. (ANL), Argonne, IL (United States); Roser, Robert [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); LeCompte, Tom [Argonne National Lab. (ANL), Argonne, IL (United States); Marshall, Zach [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Borgland, Anders [SLAC National Accelerator Lab., Menlo Park, CA (United States); Viren, Brett [Brookhaven National Lab. (BNL), Upton, NY (United States); Nugent, Peter [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Asai, Makato [SLAC National Accelerator Lab., Menlo Park, CA (United States); Bauerdick, Lothar [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Finkel, Hal [Argonne National Lab. (ANL), Argonne, IL (United States); Gottlieb, Steve [Indiana Univ., Bloomington, IN (United States); Hoeche, Stefan [SLAC National Accelerator Lab., Menlo Park, CA (United States); Sheldon, Paul [Vanderbilt Univ., Nashville, TN (United States); Vay, Jean-Luc [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Elmer, Peter [Princeton Univ., NJ (United States); Kirby, Michael [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Patton, Simon [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Potekhin, Maxim [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Yanny, Brian [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Calafiura, Paolo [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Dart, Eli [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Gutsche, Oliver [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Izubuchi, Taku [Brookhaven National Lab. (BNL), Upton, NY (United States); Lyon, Adam [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Petravick, Don [Univ. of Illinois, Urbana-Champaign, IL (United States). National Center for Supercomputing Applications (NCSA)

    2015-10-29

    Computing plays an essential role in all aspects of high energy physics. As computational technology evolves rapidly in new directions, and data throughput and volume continue to follow a steep trend-line, it is important for the HEP community to develop an effective response to a series of expected challenges. In order to help shape the desired response, the HEP Forum for Computational Excellence (HEP-FCE) initiated a roadmap planning activity with two key overlapping drivers -- 1) software effectiveness, and 2) infrastructure and expertise advancement. The HEP-FCE formed three working groups, 1) Applications Software, 2) Software Libraries and Tools, and 3) Systems (including systems software), to provide an overview of the current status of HEP computing and to present findings and opportunities for the desired HEP computational roadmap. The final versions of the reports are combined in this document, and are presented along with introductory material.

  5. High Energy Physics Forum for Computational Excellence: Working Group Reports (I. Applications Software II. Software Libraries and Tools III. Systems)

    Energy Technology Data Exchange (ETDEWEB)

    Habib, Salman [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Roser, Robert [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States)

    2015-10-28

    Computing plays an essential role in all aspects of high energy physics. As computational technology evolves rapidly in new directions, and data throughput and volume continue to follow a steep trend-line, it is important for the HEP community to develop an effective response to a series of expected challenges. In order to help shape the desired response, the HEP Forum for Computational Excellence (HEP-FCE) initiated a roadmap planning activity with two key overlapping drivers -- 1) software effectiveness, and 2) infrastructure and expertise advancement. The HEP-FCE formed three working groups, 1) Applications Software, 2) Software Libraries and Tools, and 3) Systems (including systems software), to provide an overview of the current status of HEP computing and to present findings and opportunities for the desired HEP computational roadmap. The final versions of the reports are combined in this document, and are presented along with introductory material.

  6. Mathematical structures for computer graphics

    CERN Document Server

    Janke, Steven J

    2014-01-01

    A comprehensive exploration of the mathematics behind the modeling and rendering of computer graphics scenes Mathematical Structures for Computer Graphics presents an accessible and intuitive approach to the mathematical ideas and techniques necessary for two- and three-dimensional computer graphics. Focusing on the significant mathematical results, the book establishes key algorithms used to build complex graphics scenes. Written for readers with various levels of mathematical background, the book develops a solid foundation for graphics techniques and fills in relevant grap

  7. Impact of new computing systems on computational mechanics and flight-vehicle structures technology

    Science.gov (United States)

    Noor, A. K.; Storaasli, O. O.; Fulton, R. E.

    1984-01-01

    Advances in computer technology which may have an impact on computational mechanics and flight vehicle structures technology were reviewed. The characteristics of supersystems, highly parallel systems, and small systems are summarized. The interrelations of numerical algorithms and software with parallel architectures are discussed. A scenario for future hardware/software environment and engineering analysis systems is presented. Research areas with potential for improving the effectiveness of analysis methods in the new environment are identified.

  8. Computer software helps cut costs on Delta series rockets

    Energy Technology Data Exchange (ETDEWEB)

    Cuda, S.; Laudergan, J.; Sorenson, E.; Kessler, B. [Boeing Co., Huntington Beach, California (United States)

    1999-08-01

    The Boeing company used computer simulation to reduce the cost of developing composite payload fairings for their Delta family of launch vehicles. Through part reduction and structural optimization, the new composite payload fairings cost nearly 50% less than alternative metallic assemblies. (authors)

  9. Enhancement/upgrade of Engine Structures Technology Best Estimator (EST/BEST) Software System

    Science.gov (United States)

    Shah, Ashwin

    2003-01-01

    This report describes the work performed during the contract period and the capabilities included in the EST/BEST software system. The developed EST/BEST software system includes the integrated NESSUS, IPACS, COBSTRAN, and ALCCA computer codes required to perform the engine cycle mission and component structural analysis. Also, the interactive input generator for NESSUS, IPACS, and COBSTRAN computer codes have been developed and integrated with the EST/BEST software system. The input generator allows the user to create input from scratch as well as edit existing input files interactively. Since it has been integrated with the EST/BEST software system, it enables the user to modify EST/BEST generated files and perform the analysis to evaluate the benefits. Appendix A gives details of how to use the newly added features in the EST/BEST software system.

  10. The Future of Software Engineering for High Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Pope, G [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2015-07-16

    DOE ASCR requested that from May through mid-July 2015 a study group identify issues and recommend solutions from a software engineering perspective transitioning into the next generation of High Performance Computing. The approach used was to ask some of the DOE complex experts who will be responsible for doing this work to contribute to the study group. The technique used was to solicit elevator speeches: a short and concise write up done as if the author was a speaker with only a few minutes to convince a decision maker of their top issues. Pages 2-18 contain the original texts of the contributed elevator speeches and end notes identifying the 20 contributors. The study group also ranked the importance of each topic, and those scores are displayed with each topic heading. A perfect score (and highest priority) is three, two is medium priority, and one is lowest priority. The highest scoring topic areas were software engineering and testing resources; the lowest scoring area was compliance to DOE standards. The following two paragraphs are an elevator speech summarizing the contributed elevator speeches. Each sentence or phrase in the summary is hyperlinked to its source via a numeral embedded in the text. A risk one liner has also been added to each topic to allow future risk tracking and mitigation.

  11. A Survey of Software Infrastructures and Frameworks for Ubiquitous Computing

    Directory of Open Access Journals (Sweden)

    Christoph Endres

    2005-01-01

    Full Text Available In this survey, we discuss 29 software infrastructures and frameworks which support the construction of distributed interactive systems. They range from small projects with one implemented prototype to large scale research efforts, and they come from the fields of Augmented Reality (AR, Intelligent Environments, and Distributed Mobile Systems. In their own way, they can all be used to implement various aspects of the ubiquitous computing vision as described by Mark Weiser [60]. This survey is meant as a starting point for new projects, in order to choose an existing infrastructure for reuse, or to get an overview before designing a new one. It tries to provide a systematic, relatively broad (and necessarily not very deep overview, while pointing to relevant literature for in-depth study of the systems discussed.

  12. Copyright Protection for Computer Software: Is There a Need for More Protection?

    Science.gov (United States)

    Ku, Linlin

    Because the computer industry's expansion has been much faster than has the development of laws protecting computer software and since the practice of software piracy seems to be alive and well, the issue of whether existing laws can provide effective protection for software needs further discussion. Three bodies of law have been used to protect…

  13. Computer Software: Copyright and Licensing Considerations for Schools and Libraries. ERIC Digest.

    Science.gov (United States)

    Reed, Mary Hutchings

    This digest notes that the terms and conditions of computer software package license agreements control the use of software in schools and libraries, and examines the implications of computer software license agreements for classroom use and for library lending policies. Guidelines are provided for interpreting the Copyright Act, and insuring the…

  14. Computer models of concrete structures

    OpenAIRE

    Cervenka, Vladimir; Eligehausen, Rolf; Pukl, Radomir

    1991-01-01

    The application of the nonlinear finite element analysis of concrete structures as a design tool is discussed. A computer program for structures in plane stress state is described and examples of its application in the research of fastening technique and in engineering practice are shown.

  15. Data structures, computer graphics, and pattern recognition

    CERN Document Server

    Klinger, A; Kunii, T L

    1977-01-01

    Data Structures, Computer Graphics, and Pattern Recognition focuses on the computer graphics and pattern recognition applications of data structures methodology.This book presents design related principles and research aspects of the computer graphics, system design, data management, and pattern recognition tasks. The topics include the data structure design, concise structuring of geometric data for computer aided design, and data structures for pattern recognition algorithms. The survey of data structures for computer graphics systems, application of relational data structures in computer gr

  16. Computer-aided design in power engineering. Application of software tools

    Energy Technology Data Exchange (ETDEWEB)

    Stojkovic, Zlatan

    2012-07-01

    Demonstrates the use software tools in the practice of design in the field of power systems. Presents many applications in the design in the field of power systems. Useful for educative purposes and practical work. This textbooks demonstrates the application of software tools in solving a series of problems from the field of designing power system structures and systems. It contains four chapters: The first chapter leads the reader through all the phases necessary in the procedures of computer aided modeling and simulation. It guides through the complex problems presenting on the basis of eleven original examples. The second chapter presents application of software tools in power system calculations of power systems equipment design. Several design example calculations are carried out using engineering standards like MATLAB, EMTP/ATP, Excel and Access, AutoCAD and Simulink. The third chapters focuses on the graphical documentation using a collection of software tools (AutoCAD, EPLAN, SIMARIS SIVACON, SIMARIS DESIGN) which enable the complete automation of the development of graphical documentation of a power systems. In the fourth chapter, the application of software tools in the project management in power systems is discussed. Here, the emphasis is put on the standard software MS Excel and MS Project.

  17. Providing structural modules with self-integrity monitoring software user's manual

    Science.gov (United States)

    1990-01-01

    National Aeronautics and Space Administration (NASA) Contract NAS7-961 (A Small Business Innovation and Research (SBIR) contract from NASA) involved research dealing with remote structural damage detection using the concept of substructures. Several approaches were developed. The main two were: (1) the module (substructure) transfer function matrix (MTFM) approach; and (2) modal strain energy distribution method (MSEDM). Either method can be used with a global structure; however, the focus was on substructures. As part of the research contract, computer software was to be developed which would implement the developed methods. This was done and it was used to process all the finite element generated numerical data for the research. The software was written for the IBM AT personal computer. Copies of it were placed on floppy disks. This report serves as a user's manual for the two sets of damage detection software. Sections 2.0 and 3.0 discuss the use of the MTFM and MSEDM software, respectively.

  18. Data systems and computer science: Software Engineering Program

    Science.gov (United States)

    Zygielbaum, Arthur I.

    1991-01-01

    An external review of the Integrated Technology Plan for the Civil Space Program is presented. This review is specifically concerned with the Software Engineering Program. The goals of the Software Engineering Program are as follows: (1) improve NASA's ability to manage development, operation, and maintenance of complex software systems; (2) decrease NASA's cost and risk in engineering complex software systems; and (3) provide technology to assure safety and reliability of software in mission critical applications.

  19. Meteosat Second Generation station: processing software and computing arquitecture; Estacion de Recepcion de Imagenes del Satelite Meteosat Segunda generacion: Arquitectura Informatica y Software de Proceso

    Energy Technology Data Exchange (ETDEWEB)

    Martin, L.; Cony, M.; Navarro, A. A.; Zarzalejo, L. F.; Polo, J.

    2010-05-01

    The Renewable Energy Division of CIEMAT houses a specific station for receiving the Meteosat Second Generation images, which is of interest in the works being carried out concerning the solar radiation derived from satellite images. The complexity, the huge amount of information being received and the particular characteristics of the MSG images encouraged the design and development of a specific computer structure and the associated software as well, for a better and more suitable use of the images. This document describes the mentioned structure and software. (Author) 8 refs.

  20. Geometric algebra and information geometry for quantum computational software

    Science.gov (United States)

    Cafaro, Carlo

    2017-03-01

    The art of quantum algorithm design is highly nontrivial. Grover's search algorithm constitutes a masterpiece of quantum computational software. In this article, we use methods of geometric algebra (GA) and information geometry (IG) to enhance the algebraic efficiency and the geometrical significance of the digital and analog representations of Grover's algorithm, respectively. Specifically, GA is used to describe the Grover iterate and the discretized iterative procedure that exploits quantum interference to amplify the probability amplitude of the target-state before measuring the query register. The transition from digital to analog descriptions occurs via Stone's theorem which relates the (unitary) Grover iterate to a suitable (Hermitian) Hamiltonian that controls Schrodinger's quantum mechanical evolution of a quantum state towards the target state. Once the discrete-to-continuos transition is completed, IG is used to interpret Grover's iterative procedure as a geodesic path on the manifold of the parametric density operators of pure quantum states constructed from the continuous approximation of the parametric quantum output state in Grover's algorithm. Finally, we discuss the dissipationless nature of quantum computing, recover the quadratic speedup relation, and identify the superfluity of the Walsh-Hadamard operation from an IG perspective with emphasis on statistical mechanical considerations.

  1. CLaSS Computer Literacy Software: From Design to Implementation - A Three Year Student Evaluation

    Directory of Open Access Journals (Sweden)

    Ian Cole

    2006-12-01

    Full Text Available Both computer literacy and information retrieval techniques are required to undertake studies in higher education in the United Kingdom. This paper considers the research, development and the 3-year student evaluation of a piece of learning technology in computer and information literacy (CLaSS software. Students completed a questionnaire to examine their own assessment of knowledge and competence in computer and information literacy and based on this assessment CLaSS software was created to assist nursing students with computer and information literacy. This paper draws on existing literature and applies a specific learning model to the software while considering software engineering and user-centered design methodologies. The technical processes involved in designing and creating the software are briefly considered with software development data analysis discussed. A 3-year student evaluation of the software after it's release was undertaken to consider the long-term validity and usefulness of this software with the results analysed and discussed.

  2. Good Counsel: A Computer Software-User's Guide to Legal Liability.

    Science.gov (United States)

    Evans, Thomas S.; Getman, George J.

    1984-01-01

    Various legal methods, such as copyright, trade secret, and patent law, protect proprietary rights in computer software. Users of this software must respect these rights or face infringement lawsuits. (TE)

  3. In Forming Software: Systems, Structuralism, Demythification

    Directory of Open Access Journals (Sweden)

    Edward A. Shanken

    2014-05-01

    Full Text Available In the mid-1960s, Marshall McLuhan prophesied that electronic media were creating an increasingly interconnected global village. Such pronouncements popularized the idea that the era of machine-age technology was drawing to a close, ushering in a new era of information technology. This shift finds parallels in a wave of major art performances and exhibitions between 1966-1970, including nine evenings: theatre and engineering at the New York Armory, spearheaded by Robert Rauschenberg, Billy Klüver, and Robert Whitman in 1966; The Machine: As Seen at the End of the Mechanical Age, curated by Pontus Hultén at the Museum of Modern Art in New York (MOMA in 1968; Cybernetic Serendipity, curated by Jasia Reichardt at the Institute of Contemporary Art in London in 1968; and Software, Information Technology: Its New Meaning for Art, curated by Jack Burnham at the Jewish Museum in New York.

  4. Prototype Software for Automated Structural Analysis of Systems

    DEFF Research Database (Denmark)

    Jørgensen, A.; Izadi-Zamanabadi, Roozbeh; Kristensen, M.

    2004-01-01

    In this paper we present a prototype software tool that is developed to analyse the structural model of automated systems in order to identify redundant information that is hence utilized for Fault detection and Isolation (FDI) purposes. The dedicated algorithms in this software tool use a tri......-partite graph that represents the structural model of the system. A component-based approach has been used to address issues such as system complexity and recon¯gurability possibilities....

  5. Prototype Software for Automated Structural Analysis of Systems

    DEFF Research Database (Denmark)

    Jørgensen, A.; Izadi-Zamanabadi, Roozbeh; Kristensen, M.

    2004-01-01

    In this paper we present a prototype software tool that is developed to analyse the structural model of automated systems in order to identify redundant information that is hence utilized for Fault detection and Isolation (FDI) purposes. The dedicated algorithms in this software tool use a tri......-partite graph that represents the structural model of the system. A component-based approach has been used to address issues such as system complexity and reconfigurability possibilities....

  6. Why People Copy Software and Create Computer Viruses: Individual Characteristics or Situational Factors?

    Science.gov (United States)

    Harrington, Susan J.

    1989-01-01

    Examines individual and situational variables as factors in the behavior of computer users, namely copying of software and distribution of computer viruses. The results indicate that situational variables explain much of the variance in copying software, while the individual variable of sex is strongly related to computer virus distribution. (17…

  7. 34 CFR 464.42 - What limit applies to purchasing computer hardware and software?

    Science.gov (United States)

    2010-07-01

    ... computer hardware and software? Not more than ten percent of funds received under any grant under this part may be used to purchase computer hardware or software. (Authority: 20 U.S.C. 1208aa(f)) ... 34 Education 3 2010-07-01 2010-07-01 false What limit applies to purchasing computer hardware and...

  8. Software structure for broadband wireless sensor network system

    Science.gov (United States)

    Kwon, Hyeokjun; Oh, Sechang; Yoon, Hargsoon; Varadan, Vijay K.

    2010-04-01

    Zigbee Sensor Network system has been investigating for monitoring and analyzing the data measured from a lot of sensors because the Zigbee Sensor Network has several advantages of low power consumption, compact size, and multi-node connection. However, it has a disadvantage not to be able to monitor the data measured from sensors at the remote area such as other room that is located at other city. This paper describes the software structure to compensate the defect with combining the Zigbee Sensor Network and wireless LAN technology for remote monitoring of measured sensor data. The software structure has both benefits of Zigbee Sensor Network and the advantage of wireless LAN. The software structure has three main software structures. The first software structure consists of the function in order to acquire the data from sensors and the second software structure is to gather the sensor data through wireless Zigbee and to send the data to Monitoring system by using wireless LAN. The second part consists of Linux packages software based on 2440 CPU (Samsung corp.), which has ARM9 core. The Linux packages include bootloader, device drivers, kernel, and applications, and the applications are TCP/IP server program, the program interfacing with Zigbee RF module, and wireless LAN program. The last part of software structure is to receive the sensor data through TCP/IP client program from Wireless Gate Unit and to display graphically measured data by using MATLAB program; the sensor data is measured on 100Hz sampling rate and the measured data has 10bit data resolution. The wireless data transmission rate per each channel is 1.6kbps.

  9. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan : ASC software quality engineering practices Version 3.0.

    Energy Technology Data Exchange (ETDEWEB)

    Turgeon, Jennifer L.; Minana, Molly A.; Hackney, Patricia; Pilch, Martin M.

    2009-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in the US Department of Energy/National Nuclear Security Agency (DOE/NNSA) Quality Criteria, Revision 10 (QC-1) as 'conformance to customer requirements and expectations'. This quality plan defines the SNL ASC Program software quality engineering (SQE) practices and provides a mapping of these practices to the SNL Corporate Process Requirement (CPR) 001.3.6; 'Corporate Software Engineering Excellence'. This plan also identifies ASC management's and the software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals. This SNL ASC Software Quality Plan establishes the signatories commitments to improving software products by applying cost-effective SQE practices. This plan enumerates the SQE practices that comprise the development of SNL ASC's software products and explains the project teams opportunities for tailoring and implementing the practices.

  10. HCI∧2 framework: a software framework for multimodal human-computer interaction systems.

    Science.gov (United States)

    Shen, Jie; Pantic, Maja

    2013-12-01

    This paper presents a novel software framework for the development and research in the area of multimodal human-computer interface (MHCI) systems. The proposed software framework, which is called the HCI∧2 Framework, is built upon publish/subscribe (P/S) architecture. It implements a shared-memory-based data transport protocol for message delivery and a TCP-based system management protocol. The latter ensures that the integrity of system structure is maintained at runtime. With the inclusion of bridging modules, the HCI∧2 Framework is interoperable with other software frameworks including Psyclone and ActiveMQ. In addition to the core communication middleware, we also present the integrated development environment (IDE) of the HCI∧2 Framework. It provides a complete graphical environment to support every step in a typical MHCI system development process, including module development, debugging, packaging, and management, as well as the whole system management and testing. The quantitative evaluation indicates that our framework outperforms other similar tools in terms of average message latency and maximum data throughput under a typical single PC scenario. To demonstrate HCI∧2 Framework's capabilities in integrating heterogeneous modules, we present several example modules working with a variety of hardware and software. We also present an example of a full system developed using the proposed HCI∧2 Framework, which is called the CamGame system and represents a computer game based on hand-held marker(s) and low-cost camera(s).

  11. Organizational Structure Considerations for Software Development Projects.

    Science.gov (United States)

    1982-12-01

    decil-e :h- optimal quantity of an input to utilize i the sof- waze development process. Programmer labor will be used as a rspresenta-:ivs inpu-i. It...analysis of the importance of several elements of organizational structure, and of how they can use this knowledge to make decisions which will have a... use :his knowledge m ake decision: which wil. have a: positive impact On the success c- thar projects. The structural elements discussed are

  12. PHENIX: building new software for automated crystallographic structure determination.

    Science.gov (United States)

    Adams, Paul D; Grosse-Kunstleve, Ralf W; Hung, Li Wei; Ioerger, Thomas R; McCoy, Airlie J; Moriarty, Nigel W; Read, Randy J; Sacchettini, James C; Sauter, Nicholas K; Terwilliger, Thomas C

    2002-11-01

    Structural genomics seeks to expand rapidly the number of protein structures in order to extract the maximum amount of information from genomic sequence databases. The advent of several large-scale projects worldwide leads to many new challenges in the field of crystallographic macromolecular structure determination. A novel software package called PHENIX (Python-based Hierarchical ENvironment for Integrated Xtallography) is therefore being developed. This new software will provide the necessary algorithms to proceed from reduced intensity data to a refined molecular model and to facilitate structure solution for both the novice and expert crystallographer.

  13. CRYSNET manual. Informal report. [Hardware and software of crystallographic computing network

    Energy Technology Data Exchange (ETDEWEB)

    None,

    1976-07-01

    This manual describes the hardware and software which together make up the crystallographic computing network (CRYSNET). The manual is intended as a users' guide and also provides general information for persons without any experience with the system. CRYSNET is a network of intelligent remote graphics terminals that are used to communicate with the CDC Cyber 70/76 computing system at the Brookhaven National Laboratory (BNL) Central Scientific Computing Facility. Terminals are in active use by four research groups in the field of crystallography. A protein data bank has been established at BNL to store in machine-readable form atomic coordinates and other crystallographic data for macromolecules. The bank currently includes data for more than 20 proteins. This structural information can be accessed at BNL directly by the CRYSNET graphics terminals. More than two years of experience has been accumulated with CRYSNET. During this period, it has been demonstrated that the terminals, which provide access to a large, fast third-generation computer, plus stand-alone interactive graphics capability, are useful for computations in crystallography, and in a variety of other applications as well. The terminal hardware, the actual operations of the terminals, and the operations of the BNL Central Facility are described in some detail, and documentation of the terminal and central-site software is given. (RWR)

  14. New computing systems, future computing environment, and their implications on structural analysis and design

    Science.gov (United States)

    Noor, Ahmed K.; Housner, Jerrold M.

    1993-01-01

    Recent advances in computer technology that are likely to impact structural analysis and design of flight vehicles are reviewed. A brief summary is given of the advances in microelectronics, networking technologies, and in the user-interface hardware and software. The major features of new and projected computing systems, including high performance computers, parallel processing machines, and small systems, are described. Advances in programming environments, numerical algorithms, and computational strategies for new computing systems are reviewed. The impact of the advances in computer technology on structural analysis and the design of flight vehicles is described. A scenario for future computing paradigms is presented, and the near-term needs in the computational structures area are outlined.

  15. BOOK REVIEW: Computational Atomic Structure

    Science.gov (United States)

    Post, Douglass E.

    1998-02-01

    The primary purpose of `Computational Atomic Structure' is to give a potential user of the Multi-Configuration Hartree-Fock (MCHF) Atomic Structure Package an outline of the physics and computational methods in the package, guidance on how to use the package, and information on how to interpret and use the computational results. The book is successful in all three aspects. In addition, the book provides a good overview and review of the physics of atomic structure that would be useful to the plasma physicist interested in refreshing his knowledge of atomic structure and quantum mechanics. While most of the subjects are covered in greater detail in other sources, the book is reasonably self-contained, and, in most cases, the reader can understand the basic material without recourse to other sources. The MCHF package is the standard package for computing atomic structure and wavefunctions for single or multielectron ions and atoms. It is available from a number of ftp sites. When the code was originally written in FORTRAN 77, it could only be run on large mainframes. With the advances in computer technology, the suite of codes can now be compiled and run on present day workstations and personal computers and is thus available for use by any physicist, even those with extremely modest computing resources. Sample calculations in interactive mode are included in the book to illustrate the input needed for the code, what types of results and information the code can produce, and whether the user has installed the code correctly. The user can also specify the calculational level, from simple Hartree-Fock to multiconfiguration Hartree-Fock. The MCHF method begins by finding approximate wavefunctions for the bound states of an atomic system. This involves minimizing the energy of the bound state using a variational technique. Once the wavefunctions have been determined, other atomic properties, such as the transition rates, can be determined. The book begins with an

  16. Digital computer structure and design

    CERN Document Server

    Townsend, R

    2014-01-01

    Digital Computer Structure and Design, Second Edition discusses switching theory, counters, sequential circuits, number representation, and arithmetic functions The book also describes computer memories, the processor, data flow system of the processor, the processor control system, and the input-output system. Switching theory, which is purely a mathematical concept, centers on the properties of interconnected networks of ""gates."" The theory deals with binary functions of 1 and 0 which can change instantaneously from one to the other without intermediate values. The binary number system is

  17. Development of innovative computer software to facilitate the setup and computation of water quality index.

    Science.gov (United States)

    Nabizadeh, Ramin; Valadi Amin, Maryam; Alimohammadi, Mahmood; Naddafi, Kazem; Mahvi, Amir Hossein; Yousefzadeh, Samira

    2013-04-26

    Developing a water quality index which is used to convert the water quality dataset into a single number is the most important task of most water quality monitoring programmes. As the water quality index setup is based on different local obstacles, it is not feasible to introduce a definite water quality index to reveal the water quality level. In this study, an innovative software application, the Iranian Water Quality Index Software (IWQIS), is presented in order to facilitate calculation of a water quality index based on dynamic weight factors, which will help users to compute the water quality index in cases where some parameters are missing from the datasets. A dataset containing 735 water samples of drinking water quality in different parts of the country was used to show the performance of this software using different criteria parameters. The software proved to be an efficient tool to facilitate the setup of water quality indices based on flexible use of variables and water quality databases.

  18. 78 FR 47014 - Configuration Management Plans for Digital Computer Software Used in Safety Systems of Nuclear...

    Science.gov (United States)

    2013-08-02

    ... (IEEE) Standard 828-2005, ``IEEE Standard for Software Configuration Management Plans,'' issued in 2005... RG 1.169 endorses IEEE Std. 828-2005, ``IEEE Standard for Software Configuration Management Plans... COMMISSION Configuration Management Plans for Digital Computer Software Used in Safety Systems of Nuclear...

  19. The Use Of Computer Algebra Software in Teaching Intermediate and College Algebra.

    Science.gov (United States)

    Stephens, Larry J.; Konvalina, John

    1999-01-01

    Compares two groups of students in an intermediate algebra course and two groups of students in a college algebra course with regard to the use/non-use of computer algebra software in the courses. Indicates that in both courses, students using the software outperformed the students not using the software on a common final exam. (Author/ASK)

  20. The Dark Side of Software Engineering Evil on Computing Projects

    CERN Document Server

    Rost, Johann

    2010-01-01

    Betrayal! Corruption! Software engineering? Industry experts Johann Rost and Robert L. Glass explore the seamy underbelly of software engineering in this timely report on and analysis of the prevalance of subversion, lying, hacking, and espionage on every level of software project management. Based on the authors' original research and augmented by frank discussion and insights from other well-respected figures, The Dark Side of Software Engineering goes where other management studies fear to tread -- a corporate environment where schedules are fabricated, trust is betrayed, millions of dollar

  1. 48 CFR 27.404-2 - Limited rights data and restricted computer software.

    Science.gov (United States)

    2010-10-01

    ... restricted computer software. 27.404-2 Section 27.404-2 Federal Acquisition Regulations System FEDERAL ACQUISITION REGULATION GENERAL CONTRACTING REQUIREMENTS PATENTS, DATA, AND COPYRIGHTS Rights in Data and Copyrights 27.404-2 Limited rights data and restricted computer software. (a) General. The basic clause at 52...

  2. Ethics and Equity in the Intellectual Marketplace: The Computer Software Perspective in Education.

    Science.gov (United States)

    Wholeben, Brent Edward

    1987-01-01

    Discusses the issues of ethics, equity, and values in the use of computers in education. Conceptual foundations of computer software are described; responsibilities of the software author, marketer, and user are reviewed; and the conflicts between personal property rights and individual liberty are explored. (LRW)

  3. Current practice in software development for computational neuroscience and how to improve it.

    Science.gov (United States)

    Gewaltig, Marc-Oliver; Cannon, Robert

    2014-01-01

    Almost all research work in computational neuroscience involves software. As researchers try to understand ever more complex systems, there is a continual need for software with new capabilities. Because of the wide range of questions being investigated, new software is often developed rapidly by individuals or small groups. In these cases, it can be hard to demonstrate that the software gives the right results. Software developers are often open about the code they produce and willing to share it, but there is little appreciation among potential users of the great diversity of software development practices and end results, and how this affects the suitability of software tools for use in research projects. To help clarify these issues, we have reviewed a range of software tools and asked how the culture and practice of software development affects their validity and trustworthiness. We identified four key questions that can be used to categorize software projects and correlate them with the type of product that results. The first question addresses what is being produced. The other three concern why, how, and by whom the work is done. The answers to these questions show strong correlations with the nature of the software being produced, and its suitability for particular purposes. Based on our findings, we suggest ways in which current software development practice in computational neuroscience can be improved and propose checklists to help developers, reviewers, and scientists to assess the quality of software and whether particular pieces of software are ready for use in research.

  4. Current practice in software development for computational neuroscience and how to improve it.

    Directory of Open Access Journals (Sweden)

    Marc-Oliver Gewaltig

    2014-01-01

    Full Text Available Almost all research work in computational neuroscience involves software. As researchers try to understand ever more complex systems, there is a continual need for software with new capabilities. Because of the wide range of questions being investigated, new software is often developed rapidly by individuals or small groups. In these cases, it can be hard to demonstrate that the software gives the right results. Software developers are often open about the code they produce and willing to share it, but there is little appreciation among potential users of the great diversity of software development practices and end results, and how this affects the suitability of software tools for use in research projects. To help clarify these issues, we have reviewed a range of software tools and asked how the culture and practice of software development affects their validity and trustworthiness. We identified four key questions that can be used to categorize software projects and correlate them with the type of product that results. The first question addresses what is being produced. The other three concern why, how, and by whom the work is done. The answers to these questions show strong correlations with the nature of the software being produced, and its suitability for particular purposes. Based on our findings, we suggest ways in which current software development practice in computational neuroscience can be improved and propose checklists to help developers, reviewers, and scientists to assess the quality of software and whether particular pieces of software are ready for use in research.

  5. Assessment of Computer Simulation Software and Process Data for High Pressure Die Casting of Magnesium

    Energy Technology Data Exchange (ETDEWEB)

    Sabau, Adrian S [ORNL; Hatfield, Edward C [ORNL; Dinwiddie, Ralph Barton [ORNL; Kuwana, Kazunori [University of Kentucky; Viti, Valerio [University of Kentucky, Lexington; Hassan, Mohamed I [University of Kentucky, Lexington; Saito, Kozo [University of Kentucky

    2007-09-01

    Computer software for the numerical simulation of solidification and mold filling is an effective design tool for cast structural automotive magnesium components. A review of commercial software capabilities and their validation procedures was conducted. Aside form the software assessment, the program addressed five main areas: lubricant degradation, lubricant application, gate atomization, and heat transfer at metal mold interfaces. A test stand for lubricant application was designed. A sensor was used for the direct measurement of heat fluxes during lubricant application and casting solidification in graphite molds. Spray experiments were conducted using pure deionized water and commercial die lubricants. The results show that the sensor can be used with confidence for measuring heat fluxes under conditions specific to the die lube application. The data on heat flux was presented in forms suitable for use in HPDC simulation software. Severe jet breakup and atomization phenomena are likely to occur due to high gate velocities in HPDC. As a result of gate atomization, droplet flow affects the mold filling pattern, air entrapment, skin formation, and ensuing defects. Warm water analogue dies were designed for obtaining experimental data on mold filling phenomena. Data on break-up jet length, break-up pattern, velocities, and droplet size distribution were obtained experimentally and was used to develop correlations for jet break-up phenomena specific to die casting gate configurations.

  6. Introductory Molecular Orbital Theory: An Honors General Chemistry Computational Lab as Implemented Using Three-Dimensional Modeling Software

    Science.gov (United States)

    Ruddick, Kristie R.; Parrill, Abby L.; Petersen, Richard L.

    2012-01-01

    In this study, a computational molecular orbital theory experiment was implemented in a first-semester honors general chemistry course. Students used the GAMESS (General Atomic and Molecular Electronic Structure System) quantum mechanical software (as implemented in ChemBio3D) to optimize the geometry for various small molecules. Extended Huckel…

  7. Collective network for computer structures

    Energy Technology Data Exchange (ETDEWEB)

    Blumrich, Matthias A [Ridgefield, CT; Coteus, Paul W [Yorktown Heights, NY; Chen, Dong [Croton On Hudson, NY; Gara, Alan [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Hoenicke, Dirk [Ossining, NY; Takken, Todd E [Brewster, NY; Steinmacher-Burow, Burkhard D [Wernau, DE; Vranas, Pavlos M [Bedford Hills, NY

    2011-08-16

    A system and method for enabling high-speed, low-latency global collective communications among interconnected processing nodes. The global collective network optimally enables collective reduction operations to be performed during parallel algorithm operations executing in a computer structure having a plurality of the interconnected processing nodes. Router devices ate included that interconnect the nodes of the network via links to facilitate performance of low-latency global processing operations at nodes of the virtual network and class structures. The global collective network may be configured to provide global barrier and interrupt functionality in asynchronous or synchronized manner. When implemented in a massively-parallel supercomputing structure, the global collective network is physically and logically partitionable according to needs of a processing algorithm.

  8. 77 FR 50724 - Developing Software Life Cycle Processes for Digital Computer Software Used in Safety Systems of...

    Science.gov (United States)

    2012-08-22

    ... revision 2 of RG 1.168, ``Verification, Validation, Reviews, and Audits for Digital Computer Software used... NRC-2012- 0195. You may submit comments by any of the following methods: Federal Rulemaking Web Site... possesses and are publicly available, by any of the following methods: Federal Rulemaking Web Site: Go to...

  9. CAMEO (Computer-Aided Management of Emergency Operations) Software Suite

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — CAMEO is the umbrella name for a system of software applications used widely to plan for and respond to chemical emergencies. All of the programs in the suite work...

  10. Computer Center: Software Review: The DynaPulse 200M.

    Science.gov (United States)

    Pankiewicz, Philip R., Ed.

    1995-01-01

    Reviews the DynaPulse 200M Education Edition microcomputer-based laboratory, which combines interactive software with curriculum and medical instrumentation to teach students about the cardiovascular system. (MKR)

  11. RNA secondary structure prediction using highly parallel computers.

    Science.gov (United States)

    Nakaya, A; Yamamoto, K; Yonezawa, A

    1995-12-01

    An RNA secondary structure prediction method using a highly parallel computer is reported. We focus on finding thermodynamically stable structures of a single-stranded RNA molecule. Our approach is based on a parallel combinatorial method which calculates the free energy of a molecule as the sum of the free energies of all the physically possible hydrogen bonds. Our parallel algorithm finds many highly stable structures all at once, while most of the conventional prediction methods find only the most stable structure. The important idea in our algorithm is search tree pruning, with dynamic load balancing across the processor elements in a parallel computer. Software tools for visualization and classification of secondary structures are also presented using the sequence of cadang-cadang coconut viroid as an example. Our software system runs on CM-5.

  12. Certainty in Stockpile Computing: Recommending a Verification and Validation Program for Scientific Software

    Energy Technology Data Exchange (ETDEWEB)

    Lee, J.R.

    1998-11-01

    As computing assumes a more central role in managing the nuclear stockpile, the consequences of an erroneous computer simulation could be severe. Computational failures are common in other endeavors and have caused project failures, significant economic loss, and loss of life. This report examines the causes of software failure and proposes steps to mitigate them. A formal verification and validation program for scientific software is recommended and described.

  13. Why Software Engineers are Reluctant to Use Structures Methods ...

    African Journals Online (AJOL)

    This paper looks into the reasons for an apparent boycott of theoretically proven structured techniques and methodologies by most developers. It begins by discussing the fact that when a methodology for developing software contradicts with the way human brains solves problems, the result is stress and thence burnout to ...

  14. Whole earth modeling: developing and disseminating scientific software for computational geophysics.

    Science.gov (United States)

    Kellogg, L. H.

    2016-12-01

    Historically, a great deal of specialized scientific software for modeling and data analysis has been developed by individual researchers or small groups of scientists working on their own specific research problems. As the magnitude of available data and computer power has increased, so has the complexity of scientific problems addressed by computational methods, creating both a need to sustain existing scientific software, and expand its development to take advantage of new algorithms, new software approaches, and new computational hardware. To that end, communities like the Computational Infrastructure for Geodynamics (CIG) have been established to support the use of best practices in scientific computing for solid earth geophysics research and teaching. Working as a scientific community enables computational geophysicists to take advantage of technological developments, improve the accuracy and performance of software, build on prior software development, and collaborate more readily. The CIG community, and others, have adopted an open-source development model, in which code is developed and disseminated by the community in an open fashion, using version control and software repositories like Git. One emerging issue is how to adequately identify and credit the intellectual contributions involved in creating open source scientific software. The traditional method of disseminating scientific ideas, peer reviewed publication, was not designed for review or crediting scientific software, although emerging publication strategies such software journals are attempting to address the need. We are piloting an integrated approach in which authors are identified and credited as scientific software is developed and run. Successful software citation requires integration with the scholarly publication and indexing mechanisms as well, to assign credit, ensure discoverability, and provide provenance for software.

  15. Software development methodology for computer based I&C systems of prototype fast breeder reactor

    Energy Technology Data Exchange (ETDEWEB)

    Manimaran, M., E-mail: maran@igcar.gov.in; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2015-10-15

    Highlights: • Software development methodology adopted for computer based I&C systems of PFBR is detailed. • Constraints imposed as part of software requirements and coding phase are elaborated. • Compliance to safety and security requirements are described. • Usage of CASE (Computer Aided Software Engineering) tools during software design, analysis and testing phase are explained. - Abstract: Prototype Fast Breeder Reactor (PFBR) is sodium cooled reactor which is in the advanced stage of construction in Kalpakkam, India. Versa Module Europa bus based Real Time Computer (RTC) systems are deployed for Instrumentation & Control of PFBR. RTC systems have to perform safety functions within the stipulated time which calls for highly dependable software. Hence, well defined software development methodology is adopted for RTC systems starting from the requirement capture phase till the final validation of the software product. V-model is used for software development. IEC 60880 standard and AERB SG D-25 guideline are followed at each phase of software development. Requirements documents and design documents are prepared as per IEEE standards. Defensive programming strategies are followed for software development using C language. Verification and validation (V&V) of documents and software are carried out at each phase by independent V&V committee. Computer aided software engineering tools are used for software modelling, checking for MISRA C compliance and to carry out static and dynamic analysis. Various software metrics such as cyclomatic complexity, nesting depth and comment to code are checked. Test cases are generated using equivalence class partitioning, boundary value analysis and cause and effect graphing techniques. System integration testing is carried out wherein functional and performance requirements of the system are monitored.

  16. The software life cycle

    CERN Document Server

    Ince, Darrel

    1990-01-01

    The Software Life Cycle deals with the software lifecycle, that is, what exactly happens when software is developed. Topics covered include aspects of software engineering, structured techniques of software development, and software project management. The use of mathematics to design and develop computer systems is also discussed. This book is comprised of 20 chapters divided into four sections and begins with an overview of software engineering and software development, paying particular attention to the birth of software engineering and the introduction of formal methods of software develop

  17. The work breakdown structure in software project management

    Science.gov (United States)

    Tausworthe, R. C.

    1980-01-01

    A work breakdown structure (WBS) is defined as an enumeration of all work activities in hierarchic refinement of detail which organizes work to be done into short manageable tasks with quantifiable inputs, outputs, schedules, and assigned responsibilities. Some of the characteristics and benefits of the WBS are reviewed, and ways in which these can be developed and applied in software implementation projects are discussed. Although the material is oriented principally toward new-software production tasks, many of the concepts are applicable to continuing maintenance and operations tasks.

  18. Optimal reliability allocation for large software projects through soft computing techniques

    DEFF Research Database (Denmark)

    Madsen, Henrik; Albeanu, Grigore; Popentiu-Vladicescu, Florin

    2012-01-01

    Software architecture is considered as a critical design methodology for the development of complex software. As an important step in software quality assurance, the optimal reliability allocation for software projects can be obtained by minimizing the total cost of achieving the target reliability....... Firstly, a review on existing soft computing approaches to optimization is given. The main section extends the results considering self-organizing migrating algorithms for solving intuitionistic-fuzzy optimization problems attached to complex fault-tolerant software architectures which proved...

  19. Computer organization and design the hardware/software interface

    CERN Document Server

    Patterson, David A

    2013-01-01

    The 5th edition of Computer Organization and Design moves forward into the post-PC era with new examples, exercises, and material highlighting the emergence of mobile computing and the cloud. This generational change is emphasized and explored with updated content featuring tablet computers, cloud infrastructure, and the ARM (mobile computing devices) and x86 (cloud computing) architectures. Because an understanding of modern hardware is essential to achieving good performance and energy efficiency, this edition adds a new concrete example, "Going Faster," used throughout the text to demonstrate extremely effective optimization techniques. Also new to this edition is discussion of the "Eight Great Ideas" of computer architecture. As with previous editions, a MIPS processor is the core used to present the fundamentals of hardware technologies, assembly language, computer arithmetic, pipelining, memory hierarchies and I/O. Optimization techniques featured throughout the text. It covers parallelism in depth with...

  20. High performance computing in structural determination by electron cryomicroscopy.

    Science.gov (United States)

    Fernández, J J

    2008-10-01

    Computational advances have significantly contributed to the current role of electron cryomicroscopy (cryoEM) in structural biology. The needs for computational power are constantly growing with the increasing complexity of algorithms and the amount of data needed to push the resolution limits. High performance computing (HPC) is becoming paramount in cryoEM to cope with those computational needs. Since the nineties, different HPC strategies have been proposed for some specific problems in cryoEM and, in fact, some of them are already available in common software packages. Nevertheless, the literature is scattered in the areas of computer science and structural biology. In this communication, the HPC approaches devised for the computation-intensive tasks in cryoEM (single particles and tomography) are retrospectively reviewed and the future trends are discussed. Moreover, the HPC capabilities available in the most common cryoEM packages are surveyed, as an evidence of the importance of HPC in addressing the future challenges.

  1. MHTool User's Guide - Software for Manufactured Housing Structural Design

    Energy Technology Data Exchange (ETDEWEB)

    W. D. Richins

    2005-07-01

    Since the late 1990s, the Department of Energy's Idaho National Laboratory (INL) has worked with the US Department of Housing and Urban Development (HUD), the Manufactured Housing Institute (MHI), the National Institute of Standards and Technology (NIST), the National Science Foundation (NSF), and an industry committee to measure the response of manufactured housing to both artificial and natural wind loads and to develop a computational desktop tool to optimize the structural performance of manufactured housing to HUD Code loads. MHTool is the result of an 8-year intensive testing and verification effort using single and double section homes. MHTool is the first fully integrated structural analysis software package specifically designed for manufactured housing. To use MHTool, industry design engineers will enter information (geometries, materials, connection types, etc.) describing the structure of a manufactured home, creating a base model. Windows, doors, and interior walls can be added to the initial design. Engineers will input the loads required by the HUD Code (wind, snow loads, interior live loads, etc.) and run an embedded finite element solver to find walls or connections where stresses are either excessive or very low. The designer could, for example, substitute a less expensive and easier to install connection in areas with very low stress, then re-run the analysis for verification. If forces and stresses are still within HUD Code requirements, construction costs would be saved without sacrificing quality. Manufacturers can easily change geometries or component properties to optimize designs of various floor plans then submit MHTool input and output in place of calculations for DAPIA review. No change in the regulatory process is anticipated. MHTool, while not yet complete, is now ready for demonstration. The pre-BETA version (Build-16) was displayed at the 2005 National Congress & Expo for Manufactured & Modular Housing. Additional base models

  2. Software to compute infinitesimal symmetries of exterior differenial systems, with applications

    NARCIS (Netherlands)

    Kersten, P.H.M.

    1989-01-01

    A description is given of a software package to compute symmetries of partial differential equations, using computer algebra. As an application, the computation of higher-order symmetries of the classical Boussinesq equation is given leading to the recursion operator for symmetries in a

  3. Understanding Computation of Impulse Response in Microwave Software Tools

    Science.gov (United States)

    Potrebic, Milka M.; Tosic, Dejan V.; Pejovic, Predrag V.

    2010-01-01

    In modern microwave engineering curricula, the introduction of the many new topics in microwave industrial development, or of software tools for design and simulation, sometimes results in students having an inadequate understanding of the fundamental theory. The terminology for and the explanation of algorithms for calculating impulse response in…

  4. Advances, interactions, and future developments in the CNS, Phenix, and Rosetta structural biology software systems.

    Science.gov (United States)

    Adams, Paul D; Baker, David; Brunger, Axel T; Das, Rhiju; DiMaio, Frank; Read, Randy J; Richardson, David C; Richardson, Jane S; Terwilliger, Thomas C

    2013-01-01

    Advances in our understanding of macromolecular structure come from experimental methods, such as X-ray crystallography, and also computational analysis of the growing number of atomic models obtained from such experiments. The later analyses have made it possible to develop powerful tools for structure prediction and optimization in the absence of experimental data. In recent years, a synergy between these computational methods for crystallographic structure determination and structure prediction and optimization has begun to be exploited. We review some of the advances in the algorithms used for crystallographic structure determination in the Phenix and Crystallography & NMR System software packages and describe how methods from ab initio structure prediction and refinement in Rosetta have been applied to challenging crystallographic problems. The prospects for future improvement of these methods are discussed.

  5. Joint Logistics Commanders’ Bienniel Software Workshop (3rd) on Post Deployment Software Support (PDSS) for Mission-Critical Computer Software Held in Winter Park, Florida on 31 October-4 November 1983. Volume 1. Executive Summary

    Science.gov (United States)

    1984-06-01

    computer resources after its initial development and deployment. Figure 2 from the Monterey conferences presents an idealization of the computer software development cycle...Verification. The iterative process of determining whether the product of each step of the computer software development process fulfills all requirements... Software Development yl 1-3 3 Th-1rd Software Workshop Malnagement16 LIST OF TABLES I Orlando I Panel Objectives 1-5 2 Administrative Organization17 3

  6. Software for computing plant biomass—BIOPAK users guide.

    Science.gov (United States)

    Joseph E. Means; Heather A. Hansen; Greg J. Koerper; Paul B Alaback; Mark W. Klopsch

    1994-01-01

    BIOPAK is a menu-driven package of computer programs for IBM-compatible personal computers that calculates the biomass, area, height, length, or volume of plant components (leaves, branches, stem, crown, and roots). The routines were written in FoxPro, Fortran, and C.BIOPAK was created to facilitate linking of a diverse array of vegetation datasets with the...

  7. Software survey: VOSviewer, a computer program for bibliometric mapping

    NARCIS (Netherlands)

    N.J.P. van Eck (Nees Jan); L. Waltman (Ludo)

    2010-01-01

    textabstractWe present VOSviewer, a freely available computer program that we have developed for constructing and viewing bibliometric maps. Unlike most computer programs that are used for bibliometric mapping, VOSviewer pays special attention to the graphical representation of bibliometric maps.

  8. Software and Hardware Utilization in Computer Medicine Education.

    Science.gov (United States)

    Pitts, Gerald N.; Bateman, Barry L.

    Computers are currently being used to perform medical tasks such as: (1) taking medical histories; (2) patient care and health-unit care management; (3) clinical and laboratory work; (4) physiological signal monitoring; and (5) multiphasic screening. In a survey of over 200 institutions, over 339 computer language applications were found, many of…

  9. The cognitive dynamics of computer science cost-effective large scale software development

    CERN Document Server

    De Gyurky, Szabolcs Michael; John Wiley & Sons

    2006-01-01

    This book has three major objectives: To propose an ontology for computer software; To provide a methodology for development of large software systems to cost and schedule that is based on the ontology; To offer an alternative vision regarding the development of truly autonomous systems.

  10. Comparison of the Effectiveness of Stylewriter and Microsoft Word Computer Software to Improve English Writing Skills

    Science.gov (United States)

    Prvinchandar, Sunita; Ayub, Ahmad Fauzi Mohd

    2014-01-01

    This study compared the effectiveness of two types of computer software for improving the English writing skills of pupils in a Malaysian primary school. Sixty students who participated in the seven-week training course were divided into two groups, with the experimental group using the StyleWriter software and the control group using the…

  11. Incorporating Computer-Aided Software in the Undergraduate Chemical Engineering Core Courses

    Science.gov (United States)

    Alnaizy, Raafat; Abdel-Jabbar, Nabil; Ibrahim, Taleb H.; Husseini, Ghaleb A.

    2014-01-01

    Introductions of computer-aided software and simulators are implemented during the sophomore-year of the chemical engineering (ChE) curriculum at the American University of Sharjah (AUS). Our faculty concurs that software integration within the curriculum is beneficial to our students, as evidenced by the positive feedback received from industry…

  12. A Middleware Platform for Providing Mobile and Embedded Computing Instruction to Software Engineering Students

    Science.gov (United States)

    Mattmann, C. A.; Medvidovic, N.; Malek, S.; Edwards, G.; Banerjee, S.

    2012-01-01

    As embedded software systems have grown in number, complexity, and importance in the modern world, a corresponding need to teach computer science students how to effectively engineer such systems has arisen. Embedded software systems, such as those that control cell phones, aircraft, and medical equipment, are subject to requirements and…

  13. Learning Vocabulary in a Foreign Language: A Computer Software Based Model Attempt

    Science.gov (United States)

    Yelbay Yilmaz, Yasemin

    2015-01-01

    This study aimed at devising a vocabulary learning software that would help learners learn and retain vocabulary items effectively. Foundation linguistics and learning theories have been adapted to the foreign language vocabulary learning context using a computer software named Parole that was designed exclusively for this study. Experimental…

  14. Conversion of DST Group Shape Optimisation Software for Increased Portability across Computing Platforms

    Science.gov (United States)

    2016-05-01

    UNCLASSIFIED UNCLASSIFIED Conversion of DST Group Shape Optimisation Software for Increased Portability across Computing Platforms Robert Kaye...and Witold Waldman Aerospace Division Defence Science and Technology Group DST-Group-TR-3251 ABSTRACT The DST Group shape optimisation ...2016 APPROVED FOR PUBLIC RELEASE UNCLASSIFIED UNCLASSIFIED Conversion of DST Group Shape Optimisation Software for Increased Portability

  15. Educator Liability for Copyright Infringement of Computer Software for Instructional Purposes after "Basic Books vs. Kinko's."

    Science.gov (United States)

    Douvanis, Costas J.

    1993-01-01

    Examines the nature of copyright, the current status of computer software in the context of copyright, and the applicability of the "Fair Use Doctrine" and "Classroom Guidelines" to instructional software and video as determined by relevant statute and case law. (Author/AEF)

  16. HCI^2 Framework: A software framework for multimodal human-computer interaction systems

    NARCIS (Netherlands)

    Shen, Jie; Pantic, Maja

    2013-01-01

    This paper presents a novel software framework for the development and research in the area of multimodal human-computer interface (MHCI) systems. The proposed software framework, which is called the HCI∧2 Framework, is built upon publish/subscribe (P/S) architecture. It implements a

  17. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1: ASC software quality engineering practices, Version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  18. Computational Science And Engineering Software Sustainability And Productivity (CSESSP) Challenges Workshop Report

    Data.gov (United States)

    Networking and Information Technology Research and Development, Executive Office of the President — This report details the challenges and opportunities discussed at the NITRD sponsored multi-agency workshop on Computational Science and Engineering Software...

  19. 76 FR 22918 - In the Matter of Certain Handheld Electronic Computing Devices, Related Software, and Components...

    Science.gov (United States)

    2011-04-25

    ... From the Federal Register Online via the Government Publishing Office ] INTERNATIONAL TRADE COMMISSION In the Matter of Certain Handheld Electronic Computing Devices, Related Software, and Components Thereof; Notice of Investigation AGENCY: U.S. International Trade Commission. ACTION: Notice. SUMMARY...

  20. A Dedicated Computational Platform for Cellular Monte Carlo T-CAD Software Tools

    Science.gov (United States)

    2015-07-14

    AFRL-OSR-VA-TR-2015-0176 A dedicated computational platform for Cellular Monte Carlo T- CAD software tools Marco Saraniti ARIZONA STATE UNIVERSITY...TITLE AND SUBTITLE A Dedicated Computational Platform for Cellular Monte Carlo T- CAD Software Tools 5a. CONTRACT NUMBER 5b. GRANT NUMBER FA9550-14...with an optimized architecture for the Cellular Monte Carlo particle-based T- CAD simulation tools developed by our group. Such code is used for the

  1. PSPP: a protein structure prediction pipeline for computing clusters.

    Directory of Open Access Journals (Sweden)

    Michael S Lee

    Full Text Available BACKGROUND: Protein structures are critical for understanding the mechanisms of biological systems and, subsequently, for drug and vaccine design. Unfortunately, protein sequence data exceed structural data by a factor of more than 200 to 1. This gap can be partially filled by using computational protein structure prediction. While structure prediction Web servers are a notable option, they often restrict the number of sequence queries and/or provide a limited set of prediction methodologies. Therefore, we present a standalone protein structure prediction software package suitable for high-throughput structural genomic applications that performs all three classes of prediction methodologies: comparative modeling, fold recognition, and ab initio. This software can be deployed on a user's own high-performance computing cluster. METHODOLOGY/PRINCIPAL FINDINGS: The pipeline consists of a Perl core that integrates more than 20 individual software packages and databases, most of which are freely available from other research laboratories. The query protein sequences are first divided into domains either by domain boundary recognition or Bayesian statistics. The structures of the individual domains are then predicted using template-based modeling or ab initio modeling. The predicted models are scored with a statistical potential and an all-atom force field. The top-scoring ab initio models are annotated by structural comparison against the Structural Classification of Proteins (SCOP fold database. Furthermore, secondary structure, solvent accessibility, transmembrane helices, and structural disorder are predicted. The results are generated in text, tab-delimited, and hypertext markup language (HTML formats. So far, the pipeline has been used to study viral and bacterial proteomes. CONCLUSIONS: The standalone pipeline that we introduce here, unlike protein structure prediction Web servers, allows users to devote their own computing assets to process a

  2. High Interactivity Visualization Software for Large Computational Data Sets Project

    Data.gov (United States)

    National Aeronautics and Space Administration — We propose to develop a collection of computer tools and libraries called SciViz that enable researchers to visualize large scale data sets on HPC resources remotely...

  3. A Visual Basic Software for Computing Fisher's Exact Probability

    OpenAIRE

    Haseeb Ahmad Khan

    2003-01-01

    Fisher's exact test (FET) is an important statistical method for testing association between two groups. However, the computations involved in FET are extremely tedious and time consuming due to multi-step factorial calculations after the construction of numerous 2x2 tables depending on the smallest cell value. A Visual-Basic computer program, CalcFisher, has been developed to handle the complexities of FET resorting the techniques of looping subroutines and logarithmic conversions. The softw...

  4. Software Requirements for a System to Compute Mean Failure Cost

    Energy Technology Data Exchange (ETDEWEB)

    Aissa, Anis Ben [University of Tunis, Belvedere, Tunisia; Abercrombie, Robert K [ORNL; Sheldon, Frederick T [ORNL; Mili, Ali [New Jersey Insitute of Technology

    2010-01-01

    In earlier works, we presented a computational infrastructure that allows an analyst to estimate the security of a system in terms of the loss that each stakeholder. We also demonstrated this infrastructure through the results of security breakdowns for the ecommerce case. In this paper, we illustrate this infrastructure by an application that supports the computation of the Mean Failure Cost (MFC) for each stakeholder.

  5. Software survey: VOSviewer, a computer program for bibliometric mapping

    OpenAIRE

    van Eck, Nees Jan; Waltman, Ludo

    2010-01-01

    textabstractWe present VOSviewer, a freely available computer program that we have developed for constructing and viewing bibliometric maps. Unlike most computer programs that are used for bibliometric mapping, VOSviewer pays special attention to the graphical representation of bibliometric maps. The functionality of VOSviewer is especially useful for displaying large bibliometric maps in an easy-to-interpret way. The paper consists of three parts. In the first part, an overview of VOSviewer'...

  6. Bringing Legacy Visualization Software to Modern Computing Devices via Application Streaming

    Science.gov (United States)

    Fisher, Ward

    2014-05-01

    Planning software compatibility across forthcoming generations of computing platforms is a problem commonly encountered in software engineering and development. While this problem can affect any class of software, data analysis and visualization programs are particularly vulnerable. This is due in part to their inherent dependency on specialized hardware and computing environments. A number of strategies and tools have been designed to aid software engineers with this task. While generally embraced by developers at 'traditional' software companies, these methodologies are often dismissed by the scientific software community as unwieldy, inefficient and unnecessary. As a result, many important and storied scientific software packages can struggle to adapt to a new computing environment; for example, one in which much work is carried out on sub-laptop devices (such as tablets and smartphones). Rewriting these packages for a new platform often requires significant investment in terms of development time and developer expertise. In many cases, porting older software to modern devices is neither practical nor possible. As a result, replacement software must be developed from scratch, wasting resources better spent on other projects. Enabled largely by the rapid rise and adoption of cloud computing platforms, 'Application Streaming' technologies allow legacy visualization and analysis software to be operated wholly from a client device (be it laptop, tablet or smartphone) while retaining full functionality and interactivity. It mitigates much of the developer effort required by other more traditional methods while simultaneously reducing the time it takes to bring the software to a new platform. This work will provide an overview of Application Streaming and how it compares against other technologies which allow scientific visualization software to be executed from a remote computer. We will discuss the functionality and limitations of existing application streaming

  7. Computer Aided Design of Transformer Station Grounding System Using CDEGS Software

    Directory of Open Access Journals (Sweden)

    S. Nikolovski

    2004-01-01

    Full Text Available This paper presents a computer-aided design of a transformer station grounding system. Fault conditions in a transformer station can produce huge damage to transformer station equipment if the grounding system is not designed properly. A well designed grounding system is a very important part of the project for transformer station design as a whole. This paper analyses a procedure for transformer grounding system design and spatial distribution of touch and step voltage on the ground surface level, using the CDEGS (Current Distribution Electromagnetic Interference Grounding and Soil Structure Analysis software. Spatial distribution is needed for checking and finding dangerous step and touch voltages above and around the transformer station. Apparent earth resistivity data is measured and analyzed using the RESAP module of the CDEGS software. Because of the very high current flow into the grounding system during a single line to ground fault or a three phase fault in the transformer station, very high and dangerous potentials can be induced on the metallic structures including the fence, which can cause dangerous situations for people and animals near the station and for the personnel inside the station. The PLOT module of CDEGS is used to view the results of the scalar potential, step and touch voltage on the surface. Graphic displays include equipotent contour lines and potential profiles (gradients in 3D and 2D perspective and apparent soil resistivity (Wm versus inter electrode spacing (m. The results of alternative grid designs may be displayed simultaneously for the purpose of comparison.

  8. Finite difference computing with PDEs a modern software approach

    CERN Document Server

    Langtangen, Hans Petter

    2017-01-01

    This book is open access under a CC BY 4.0 license. This easy-to-read book introduces the basics of solving partial differential equations by means of finite difference methods. Unlike many of the traditional academic works on the topic, this book was written for practitioners. Accordingly, it especially addresses: the construction of finite difference schemes, formulation and implementation of algorithms, verification of implementations, analyses of physical behavior as implied by the numerical solutions, and how to apply the methods and software to solve problems in the fields of physics and biology.

  9. Free and open-source software application for the evaluation of coronary computed tomography angiography images.

    Science.gov (United States)

    Hadlich, Marcelo Souza; Oliveira, Gláucia Maria Moraes; Feijóo, Raúl A; Azevedo, Clerio F; Tura, Bernardo Rangel; Ziemer, Paulo Gustavo Portela; Blanco, Pablo Javier; Pina, Gustavo; Meira, Márcio; Souza e Silva, Nelson Albuquerque de

    2012-10-01

    The standardization of images used in Medicine in 1993 was performed using the DICOM (Digital Imaging and Communications in Medicine) standard. Several tests use this standard and it is increasingly necessary to design software applications capable of handling this type of image; however, these software applications are not usually free and open-source, and this fact hinders their adjustment to most diverse interests. To develop and validate a free and open-source software application capable of handling DICOM coronary computed tomography angiography images. We developed and tested the ImageLab software in the evaluation of 100 tests randomly selected from a database. We carried out 600 tests divided between two observers using ImageLab and another software sold with Philips Brilliance computed tomography appliances in the evaluation of coronary lesions and plaques around the left main coronary artery (LMCA) and the anterior descending artery (ADA). To evaluate intraobserver, interobserver and intersoftware agreements, we used simple and kappa statistics agreements. The agreements observed between software applications were generally classified as substantial or almost perfect in most comparisons. The ImageLab software agreed with the Philips software in the evaluation of coronary computed tomography angiography tests, especially in patients without lesions, with lesions 70% in the ADA was lower, but this is also observed when the anatomical reference standard is used.

  10. Computing with words to feasibility study of software projects

    Directory of Open Access Journals (Sweden)

    Marieta Peña Abreu

    2017-02-01

    Full Text Available Objective: This paper proposes a method to analyze the technical, commercial and social feasibility of software projects in environments of uncertainty. It allows working with multiple experts and multiple criteria and facilitates decision-making. Method: The proposal contains two phases, first the necessary information is collected and in second place projects are evaluated using 2-tuple linguistic representation model. The experts are selected by analyzing their curricular synthesis. The evaluation criteria are defined using the technique Focus Group and weighted in the interval (0,1 according to their importance. three domains are offered to express the preferences: numeric, interval-valued and linguistic. For aggregation extended arithmetic mean and weighted average extended are used, preventing the loss of information. A 2-tuple (feasibility, precision is obtained as a result for each project. Results: The evaluation of P1 project was a very high feasibility with -0,33 of precision. The P2 project obtained a high feasibility with 0,38 of precision and P3 project achieved a medium feasibility with -0,21 of precision. Conclusions: This method is favorable for software projects feasibility analysis with presence of multiple experts and criteria, in environments of uncertainty. It tries heterogeneous assessments without loss of information. Their results are consistent and useful for decision makers.

  11. Training Software in Artificial-Intelligence Computing Techniques

    Science.gov (United States)

    Howard, Ayanna; Rogstad, Eric; Chalfant, Eugene

    2005-01-01

    The Artificial Intelligence (AI) Toolkit is a computer program for training scientists, engineers, and university students in three soft-computing techniques (fuzzy logic, neural networks, and genetic algorithms) used in artificial-intelligence applications. The program promotes an easily understandable tutorial interface, including an interactive graphical component through which the user can gain hands-on experience in soft-computing techniques applied to realistic example problems. The tutorial provides step-by-step instructions on the workings of soft-computing technology, whereas the hands-on examples allow interaction and reinforcement of the techniques explained throughout the tutorial. In the fuzzy-logic example, a user can interact with a robot and an obstacle course to verify how fuzzy logic is used to command a rover traverse from an arbitrary start to the goal location. For the genetic algorithm example, the problem is to determine the minimum-length path for visiting a user-chosen set of planets in the solar system. For the neural-network example, the problem is to decide, on the basis of input data on physical characteristics, whether a person is a man, woman, or child. The AI Toolkit is compatible with the Windows 95,98, ME, NT 4.0, 2000, and XP operating systems. A computer having a processor speed of at least 300 MHz, and random-access memory of at least 56MB is recommended for optimal performance. The program can be run on a slower computer having less memory, but some functions may not be executed properly.

  12. Improving package structure of object-oriented software using multi-objective optimization and weighted class connections

    Directory of Open Access Journals (Sweden)

    Amarjeet

    2017-07-01

    Full Text Available The software maintenance activities performed without following the original design decisions about the package structure usually deteriorate the quality of software modularization, leading to decay of the quality of the system. One of the main reasons for such structural deterioration is inappropriate grouping of source code classes in software packages. To improve such grouping/modular-structure, previous researchers formulated the software remodularization problem as an optimization problem and solved it using search-based meta-heuristic techniques. These optimization approaches aimed at improving the quality metrics values of the structure without considering the original package design decisions, often resulting into a totally new software modularization. The entirely changed software modularization becomes costly to realize as well as difficult to understand for the developers/maintainers. To alleviate this issue, we propose a multi-objective optimization approach to improve the modularization quality of an object-oriented system with minimum possible movement of classes between existing packages of original software modularization. The optimization is performed using NSGA-II, a widely-accepted multi-objective evolutionary algorithm. In order to ensure minimum modification of original package structure, a new approach of computing class relations using weighted strengths has been proposed here. The weights of relations among different classes are computed on the basis of the original package structure. A new objective function has been formulated using these weighted class relations. This objective function drives the optimization process toward better modularization quality simultaneously ensuring preservation of original structure. To evaluate the results of the proposed approach, a series of experiments are conducted over four real-worlds and two random software applications. The experimental results clearly indicate the effectiveness

  13. Computer organization and design the hardware/software interface

    CERN Document Server

    Patterson, David A

    2009-01-01

    The classic textbook for computer systems analysis and design, Computer Organization and Design, has been thoroughly updated to provide a new focus on the revolutionary change taking place in industry today: the switch from uniprocessor to multicore microprocessors. This new emphasis on parallelism is supported by updates reflecting the newest technologies with examples highlighting the latest processor designs, benchmarking standards, languages and tools. As with previous editions, a MIPS processor is the core used to present the fundamentals of hardware technologies, assembly language, compu

  14. Human-Computer Interaction Software: Lessons Learned, Challenges Ahead

    Science.gov (United States)

    1989-01-01

    appropriate next step. the history of HCI over the last decades, makes it much easier and productive for Because humans and computers are not duringwhich...ivnice et.Lsrivo Clrd.CiiisP ware produict. The cost of hardware and 430. Bulder , CO 80309: CStiet getirrAl~tilder.niloradii.edtI softwsare in future

  15. What's New in Software? Integrated Computer Programs and Daily Living.

    Science.gov (United States)

    Hedley, Carolyn N.

    1989-01-01

    Various kinds of electronic information media can now be integrated to plan educational programs, through use of computer videodiscs, hypercards, and hypertexts. Discussed are the components of integrative technology, including audio technology, video technology, and electronic text and graphics, and possibilities for interfacing the various…

  16. Page 1 Computer science and technology 221 software ...

    Indian Academy of Sciences (India)

    technology and the far-reaching social, economic, and industrial changes that this technology is bringing about in advanced industrialized countries. The coming together of computing and communications has been the single most important cause for the revolutionary impact of information technology. In this background ...

  17. QuBiLS-MIDAS: a parallel free-software for molecular descriptors computation based on multilinear algebraic maps.

    Science.gov (United States)

    García-Jacas, César R; Marrero-Ponce, Yovani; Acevedo-Martínez, Liesner; Barigye, Stephen J; Valdés-Martiní, José R; Contreras-Torres, Ernesto

    2014-07-05

    The present report introduces the QuBiLS-MIDAS software belonging to the ToMoCoMD-CARDD suite for the calculation of three-dimensional molecular descriptors (MDs) based on the two-linear (bilinear), three-linear, and four-linear (multilinear or N-linear) algebraic forms. Thus, it is unique software that computes these tensor-based indices. These descriptors, establish relations for two, three, and four atoms by using several (dis-)similarity metrics or multimetrics, matrix transformations, cutoffs, local calculations and aggregation operators. The theoretical background of these N-linear indices is also presented. The QuBiLS-MIDAS software was developed in the Java programming language and employs the Chemical Development Kit library for the manipulation of the chemical structures and the calculation of the atomic properties. This software is composed by a desktop user-friendly interface and an Abstract Programming Interface library. The former was created to simplify the configuration of the different options of the MDs, whereas the library was designed to allow its easy integration to other software for chemoinformatics applications. This program provides functionalities for data cleaning tasks and for batch processing of the molecular indices. In addition, it offers parallel calculation of the MDs through the use of all available processors in current computers. The studies of complexity of the main algorithms demonstrate that these were efficiently implemented with respect to their trivial implementation. Lastly, the performance tests reveal that this software has a suitable behavior when the amount of processors is increased. Therefore, the QuBiLS-MIDAS software constitutes a useful application for the computation of the molecular indices based on N-linear algebraic maps and it can be used freely to perform chemoinformatics studies. Copyright © 2014 Wiley Periodicals, Inc.

  18. RNAstructure: software for RNA secondary structure prediction and analysis

    Directory of Open Access Journals (Sweden)

    Mathews David H

    2010-03-01

    Full Text Available Abstract Background To understand an RNA sequence's mechanism of action, the structure must be known. Furthermore, target RNA structure is an important consideration in the design of small interfering RNAs and antisense DNA oligonucleotides. RNA secondary structure prediction, using thermodynamics, can be used to develop hypotheses about the structure of an RNA sequence. Results RNAstructure is a software package for RNA secondary structure prediction and analysis. It uses thermodynamics and utilizes the most recent set of nearest neighbor parameters from the Turner group. It includes methods for secondary structure prediction (using several algorithms, prediction of base pair probabilities, bimolecular structure prediction, and prediction of a structure common to two sequences. This contribution describes new extensions to the package, including a library of C++ classes for incorporation into other programs, a user-friendly graphical user interface written in JAVA, and new Unix-style text interfaces. The original graphical user interface for Microsoft Windows is still maintained. Conclusion The extensions to RNAstructure serve to make RNA secondary structure prediction user-friendly. The package is available for download from the Mathews lab homepage at http://rna.urmc.rochester.edu/RNAstructure.html.

  19. Paradigm of Legal Protection of Computer Software Contracts in the United States: Brief Overview of “Principles of the Law of Software Contracts”

    Science.gov (United States)

    Furuya, Haruhisa; Hiratsuka, Mitsuyoshi

    This article overviews the historical transition of legal protection of Computer software contracts in the Unite States and presents how it should function under Uniform Commercial Code and its amended Article 2B, Uniform Computer Information Transactions Act, and also recently-approved “Principles of the Law of Software Contracts”.

  20. 48 CFR 1852.227-19 - Commercial computer software-Restricted rights.

    Science.gov (United States)

    2010-10-01

    ... Provisions and Clauses 1852.227-19 Commercial computer software—Restricted rights. (a) As prescribed in 1827... 48 Federal Acquisition Regulations System 6 2010-10-01 2010-10-01 true Commercial computer software-Restricted rights. 1852.227-19 Section 1852.227-19 Federal Acquisition Regulations System NATIONAL...

  1. Computer Game Theories for Designing Motivating Educational Software: A Survey Study

    Science.gov (United States)

    Ang, Chee Siang; Rao, G. S. V. Radha Krishna

    2008-01-01

    The purpose of this study is to evaluate computer game theories for educational software. We propose a framework for designing engaging educational games based on contemporary game studies which includes ludology and narratology. Ludology focuses on the study of computer games as play and game activities, while narratology revolves around the…

  2. The shift to Cloud Computing : The impact of disruptive technology on the enterprise software business ecosystem

    NARCIS (Netherlands)

    Nieuwenhuis, Lambert J.M.; Ehrenhard, Michel L.; Prause, Lars

    2017-01-01

    The rapid diffusion of Cloud Computing influences the way enterprise software is developed, distributed, and implemented. This uptake of Cloud Computing has profound implications for the IT industry and related industries, as it does not only affect the vendors' business models but also the other

  3. Computer Security: How to succeed in software deployment

    CERN Multimedia

    Computer Security Team

    2014-01-01

    The summer student period has ended and we would like to congratulate all those who successfully accomplished their project! In particular, well done to those who managed to develop and deploy sophisticated web applications in the short summer season. Unfortunately, not all web applications made the final cut, moved into production and became visible on the Internet. We had to reject some... let me explain why.   Making a web application visible on the Internet requires an opening in the CERN outer perimeter firewall. Such a request is usually made through the CERN WebReq web interface. As standard procedure, the CERN Computer Security team reviews every request and performs a security assessment. This is where you, your supervisee and the Computer Security team all start to get frustrated. Many summer students delivered awesome web applications with great new functions and a good “look and feel” following precise use cases, using modern web technologies, dashboards, integr...

  4. Computer modeling and software development for unsteady chemical technological systems

    Directory of Open Access Journals (Sweden)

    Dolganov Igor

    2016-01-01

    Full Text Available The paper deals with mathematical modeling in transient conditions to create a computer system that can reflect the behavior of real industrial plants. Such systems can respond to complex and pressing questions about the stability of the industrial facilities and the time spent on transients passing through the unstable regimes. In addition, such systems have a kind of intelligence and predictive ability, as they consider partial integral and differential systems of equations that are based on physical and chemical nature of the processes occurring in devices of technological systems.

  5. Computer organization and design the hardware/software interface

    CERN Document Server

    Patterson, David A

    2011-01-01

    This Fourth Revised Edition of Computer Organization and Design includes a complete set of updated and new exercises, along with improvements and changes suggested by instructors and students. Focusing on the revolutionary change taking place in industry today--the switch from uniprocessor to multicore microprocessors--this classic textbook has a modern and up-to-date focus on parallelism in all its forms. Examples highlighting multicore and GPU processor designs are supported with performance and benchmarking data. As with previous editions, a MIPS processor is the core used to pres

  6. Coherent information structure in complex computation.

    Science.gov (United States)

    Lizier, Joseph T; Prokopenko, Mikhail; Zomaya, Albert Y

    2012-09-01

    We have recently presented a framework for the information dynamics of distributed computation that locally identifies the component operations of information storage, transfer, and modification. We have observed that while these component operations exist to some extent in all types of computation, complex computation is distinguished in having coherent structure in its local information dynamics profiles. In this article, we conjecture that coherent information structure is a defining feature of complex computation, particularly in biological systems or artificially evolved computation that solves human-understandable tasks. We present a methodology for studying coherent information structure, consisting of state-space diagrams of the local information dynamics and a measure of structure in these diagrams. The methodology identifies both clear and "hidden" coherent structure in complex computation, most notably reconciling conflicting interpretations of the complexity of the Elementary Cellular Automata rule 22.

  7. Advances in Multimedia, Software Engineering and Computing Vol.1 : Proceedings of the 2011 MSEC International Conference on Multimedia, Software Engineering and Computing

    CERN Document Server

    Lin, Sally

    2012-01-01

    MSEC2011 is an integrated conference concentrating its focus upon Multimedia ,Software Engineering, Computing and Education. In the proceeding, you can learn much more knowledge about Multimedia, Software Engineering ,Computing and Education of researchers all around the world. The main role of the proceeding is to be used as an exchange pillar for researchers who are working in the mentioned field. In order to meet high standard of Springer, AISC series ,the organization committee has made their efforts to do the following things. Firstly, poor quality paper has been refused after reviewing course by anonymous referee experts. Secondly, periodically review meetings have been held around the reviewers about five times for exchanging reviewing suggestions. Finally, the conference organization had several preliminary sessions before the conference. Through efforts of different people and departments, the conference will be successful and fruitful.

  8. Advances in Multimedia, Software Engineering and Computing Vol.2 : Proceedings of the 2011 MSEC International Conference on Multimedia, Software Engineering and Computing

    CERN Document Server

    Lin, Sally

    2012-01-01

    MSEC2011 is an integrated conference concentrating its focus upon Multimedia ,Software Engineering, Computing and Education. In the proceeding, you can learn much more knowledge about Multimedia, Software Engineering, Computing and Education of researchers all around the world. The main role of the proceeding is to be used as an exchange pillar for researchers who are working in the mentioned field. In order to meet high standard of Springer, AISC series ,the organization committee has made their efforts to do the following things. Firstly, poor quality paper has been refused after reviewing course by anonymous referee experts. Secondly, periodically review meetings have been held around the reviewers about five times for exchanging reviewing suggestions. Finally, the conference organization had several preliminary sessions before the conference. Through efforts of different people and departments, the conference will be successful and fruitful.

  9. 78 FR 47805 - Test Documentation for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2013-08-06

    ... COMMISSION Test Documentation for Digital Computer Software Used in Safety Systems of Nuclear Power Plants..., ``Test Documentation for Digital Computer Software Used in Safety Systems of Nuclear Power Plants.'' This... quality in the software used in safety systems of nuclear power plants. ADDRESSES: Please refer to Docket...

  10. Metagram Software - A New Perspective on the Art of Computation.

    Science.gov (United States)

    1981-10-01

    registration Unity is initially 1, and nolds succesZive prodicts. The information Pred--n is a structure that gives n as first value and tnereafter values...and small scales. And 38 Emil " . the phenomenon occurs in the education of every child. The possibility of such creation is implicit in human

  11. Computational Ecology and Software (http://www.iaees.org/publications/journals/ces/online-version.asp

    Directory of Open Access Journals (Sweden)

    ces@iaees.org

    Full Text Available Computational Ecology and Software ISSN 2220-721X URL: http://www.iaees.org/publications/journals/ces/online-version.asp RSS: http://www.iaees.org/publications/journals/ces/rss.xml E-mail: ces@iaees.org Editor-in-Chief: WenJun Zhang Aims and Scope COMPUTATIONAL ECOLOGY AND SOFTWARE (ISSN 2220-721X is an open access, peer-reviewed online journal that considers scientific articles in all different areas of computational ecology. It is the transactions of the International Society of Computational Ecology. The journal is concerned with the ecological researches, constructions and applications of theories and methods of computational sciences including computational mathematics, computational statistics and computer science. It features the simulation, approximation, prediction, recognition, and classification of ecological issues. Intensive computation is one of the major stresses of the journal. The journal welcomes research articles, short communications, review articles, perspectives, and book reviews. The journal also supports the activities of the International Society of Computational Ecology. The topics to be covered by CES include, but are not limited to: •Computation intensive methods, numerical and optimization methods, differential and difference equation modeling and simulation, prediction, recognition, classification, statistical computation (Bayesian computing, randomization, bootstrapping, Monte Carlo techniques, stochastic process, etc., agent-based modeling, individual-based modeling, artificial neural networks, knowledge based systems, machine learning, genetic algorithms, data exploration, network analysis and computation, databases, ecological modeling and computation using Geographical Information Systems, satellite imagery, and other computation intensive theories and methods. •Artificial ecosystems, artificial life, complexity of ecosystems and virtual reality. •The development, evaluation and validation of software and

  12. Computer software configuration management plan for 200 East/West Liquid Effluent Facilities

    Energy Technology Data Exchange (ETDEWEB)

    Graf, F.A. Jr.

    1995-02-27

    This computer software management configuration plan covers the control of the software for the monitor and control system that operates the Effluent Treatment Facility and its associated truck load in station and some key aspects of the Liquid Effluent Retention Facility that stores condensate to be processed. Also controlled is the Treated Effluent Disposal System`s pumping stations and monitors waste generator flows in this system as well as the Phase Two Effluent Collection System.

  13. Exploring the effects of computer software for teaching reading and writing skills in young children

    OpenAIRE

    Vilaseca Momplet, Rosa María; Basil, Carme; Reyes, S.

    2013-01-01

    This article describes computer software for teaching reading and writing skills grounded on theoretical principles that promote motivating, self-initiated and discovery learning. This software is based on error-free learning and allows children to explore textual material successfully without the need for prior reading and writing skills. Twenty children aged four and five were divided into an experimental group and a control group. Their reading and writing skills were assessed at the initi...

  14. Development of a computer-aided design software for dental splint in orthognathic surgery

    OpenAIRE

    Xiaojun Chen; Xing Li; Lu Xu; Yi Sun; Constantinus Politis; Jan Egger

    2016-01-01

    In the orthognathic surgery, dental splints are important and necessary to help the surgeon reposition the maxilla or mandible. However, the traditional methods of manual design of dental splints are difficult and time-consuming. The research on computer-aided design software for dental splints is rarely reported. Our purpose is to develop a novel special software named EasySplint to design the dental splints conveniently and efficiently. The design can be divided into two steps, which are th...

  15. Review Essay: Guidance in the World of Computer-Assisted Qualitative Data Analysis Software (CAQDAS) Programs

    OpenAIRE

    Áine Humble

    2015-01-01

    This review discusses Christina SILVER and Ann Lewins' book, "Using Software in Qualitative Research: A Step-by-Step Guide" (2nd ed.). This book is an impressive undertaking, with online supplemental material in the form of three data sets consisting of many different types of data, detailed instructions for seven CAQDAS (Computer-Assisted Qualitative Data Analysis Software) programs, and full-color reproductions of illustrations from the book. The 14 chapters in the book cover a wide range o...

  16. Software for computer-aided study of intake and exhaust systems for engines

    Energy Technology Data Exchange (ETDEWEB)

    Payri, F.; Benajes, J.; Chust, M.D. (Universidad Politecnica de Valencia (ES))

    1991-01-01

    In this work are described the main characteristics of a software designed to globally analyze the behavior of intake and exhaust systems of alternating engines with multiple cylinders and the effect produced by each element on the engine performances and the emitted noise. The software consists in a model of waves action, which computes the fluid behavior through the collectors using particular boundary conditions for each singular element of the system: filter, compressor, carburettor, valves, pipes, mufflers, junctions, etc. Moreover, the software contains interactive modules to introduce the data and to graphically output the results. Finally we present results showing the model accuracy and examples of input and outputs. 29 refs..

  17. Surveillance Analysis Computer System (SACS): Software requirements specification (SRS). Revision 2

    Energy Technology Data Exchange (ETDEWEB)

    Glasscock, J.A.

    1995-03-08

    This document is the primary document establishing requirements for the Surveillance Analysis Computer System (SACS) database, an Impact Level 3Q system. SACS stores information on tank temperatures, surface levels, and interstitial liquid levels. This information is retrieved by the customer through a PC-based interface and is then available to a number of other software tools. The software requirements specification (SRS) describes the system requirements for the SACS Project, and follows the Standard Engineering Practices (WHC-CM-6-1), Software Practices (WHC-CM-3-10) and Quality Assurance (WHC-CM-4-2, QR 19.0) policies.

  18. Agile Development of Various Computational Power Adaptive Web-Based Mobile-Learning Software Using Mobile Cloud Computing

    Science.gov (United States)

    Zadahmad, Manouchehr; Yousefzadehfard, Parisa

    2016-01-01

    Mobile Cloud Computing (MCC) aims to improve all mobile applications such as m-learning systems. This study presents an innovative method to use web technology and software engineering's best practices to provide m-learning functionalities hosted in a MCC-learning system as service. Components hosted by MCC are used to empower developers to create…

  19. Efficient multi-objective calibration of a computationally intensive hydrologic model with parallel computing software in Python

    Science.gov (United States)

    With enhanced data availability, distributed watershed models for large areas with high spatial and temporal resolution are increasingly used to understand water budgets and examine effects of human activities and climate change/variability on water resources. Developing parallel computing software...

  20. Computational Software for Fitting Seismic Data to Epidemic-Type Aftershock Sequence Models

    Science.gov (United States)

    Chu, A.

    2014-12-01

    Modern earthquake catalogs are often analyzed using spatial-temporal point process models such as the epidemic-type aftershock sequence (ETAS) models of Ogata (1998). My work introduces software to implement two of ETAS models described in Ogata (1998). To find the Maximum-Likelihood Estimates (MLEs), my software provides estimates of the homogeneous background rate parameter and the temporal and spatial parameters that govern triggering effects by applying the Expectation-Maximization (EM) algorithm introduced in Veen and Schoenberg (2008). Despite other computer programs exist for similar data modeling purpose, using EM-algorithm has the benefits of stability and robustness (Veen and Schoenberg, 2008). Spatial shapes that are very long and narrow cause difficulties in optimization convergence and problems with flat or multi-modal log-likelihood functions encounter similar issues. My program uses a robust method to preset a parameter to overcome the non-convergence computational issue. In addition to model fitting, the software is equipped with useful tools for examining modeling fitting results, for example, visualization of estimated conditional intensity, and estimation of expected number of triggered aftershocks. A simulation generator is also given with flexible spatial shapes that may be defined by the user. This open-source software has a very simple user interface. The user may execute it on a local computer, and the program also has potential to be hosted online. Java language is used for the software's core computing part and an optional interface to the statistical package R is provided.

  1. An Optimal Path Computation Architecture for the Cloud-Network on Software-Defined Networking

    Directory of Open Access Journals (Sweden)

    Hyunhun Cho

    2015-05-01

    Full Text Available Legacy networks do not open the precise information of the network domain because of scalability, management and commercial reasons, and it is very hard to compute an optimal path to the destination. According to today’s ICT environment change, in order to meet the new network requirements, the concept of software-defined networking (SDN has been developed as a technological alternative to overcome the limitations of the legacy network structure and to introduce innovative concepts. The purpose of this paper is to propose the application that calculates the optimal paths for general data transmission and real-time audio/video transmission, which consist of the major services of the National Research & Education Network (NREN in the SDN environment. The proposed SDN routing computation (SRC application is designed and applied in a multi-domain network for the efficient use of resources, selection of the optimal path between the multi-domains and optimal establishment of end-to-end connections.

  2. The LHCb software and computing upgrade for Run 3: opportunities and challenges

    Science.gov (United States)

    Bozzi, C.; Roiser, S.; LHCb Collaboration

    2017-10-01

    The LHCb detector will be upgraded for the LHC Run 3 and will be readout at 30 MHz, corresponding to the full inelastic collision rate, with major implications on the full software trigger and offline computing. If the current computing model and software framework are kept, the data storage capacity and computing power required to process data at this rate, and to generate and reconstruct equivalent samples of simulated events, will exceed the current capacity by at least one order of magnitude. A redesign of the software framework, including scheduling, the event model, the detector description and the conditions database, is needed to fully exploit the computing power of multi-, many-core architectures, and coprocessors. Data processing and the analysis model will also change towards an early streaming of different data types, in order to limit storage resources, with further implications for the data analysis workflows. Fast simulation options will allow to obtain a reasonable parameterization of the detector response in considerably less computing time. Finally, the upgrade of LHCb will be a good opportunity to review and implement changes in the domains of software design, test and review, and analysis workflow and preservation. In this contribution, activities and recent results in all the above areas are presented.

  3. Teaching structure: student use of software tools for understanding macromolecular structure in an undergraduate biochemistry course.

    Science.gov (United States)

    Jaswal, Sheila S; O'Hara, Patricia B; Williamson, Patrick L; Springer, Amy L

    2013-01-01

    Because understanding the structure of biological macromolecules is critical to understanding their function, students of biochemistry should become familiar not only with viewing, but also with generating and manipulating structural representations. We report a strategy from a one-semester undergraduate biochemistry course to integrate use of structural representation tools into both laboratory and homework activities. First, early in the course we introduce the use of readily available open-source software for visualizing protein structure, coincident with modules on amino acid and peptide bond properties. Second, we use these same software tools in lectures and incorporate images and other structure representations in homework tasks. Third, we require a capstone project in which teams of students examine a protein-nucleic acid complex and then use the software tools to illustrate for their classmates the salient features of the structure, relating how the structure helps explain biological function. To ensure engagement with a range of software and database features, we generated a detailed template file that can be used to explore any structure, and that guides students through specific applications of many of the software tools. In presentations, students demonstrate that they are successfully interpreting structural information, and using representations to illustrate particular points relevant to function. Thus, over the semester students integrate information about structural features of biological macromolecules into the larger discussion of the chemical basis of function. Together these assignments provide an accessible introduction to structural representation tools, allowing students to add these methods to their biochemical toolboxes early in their scientific development. © 2013 by The International Union of Biochemistry and Molecular Biology.

  4. ATLAS Experience with HEP Software at the Argonne Leadership Computing Facility

    CERN Document Server

    LeCompte, T; The ATLAS collaboration; Benjamin, D

    2014-01-01

    A number of HEP software packages used by the ATLAS experiment, including GEANT4, ROOT and ALPGEN, have been adapted to run on the IBM Blue Gene supercomputers at the Argonne Leadership Computing Facility. These computers use a non-x86 architecture and have a considerably less rich operating environment than in common use in HEP, but also represent a computing capacity an order of magnitude beyond what ATLAS is presently using via the LCG. The status and potential for making use of leadership-class computing, including the status of integration with the ATLAS production system, is discussed.

  5. QRev—Software for computation and quality assurance of acoustic doppler current profiler moving-boat streamflow measurements—Technical manual for version 2.8

    Science.gov (United States)

    Mueller, David S.

    2016-06-21

    The software program, QRev applies common and consistent computational algorithms combined with automated filtering and quality assessment of the data to improve the quality and efficiency of streamflow measurements and helps ensure that U.S. Geological Survey streamflow measurements are consistent, accurate, and independent of the manufacturer of the instrument used to make the measurement. Software from different manufacturers uses different algorithms for various aspects of the data processing and discharge computation. The algorithms used by QRev to filter data, interpolate data, and compute discharge are documented and compared to the algorithms used in the manufacturers’ software. QRev applies consistent algorithms and creates a data structure that is independent of the data source. QRev saves an extensible markup language (XML) file that can be imported into databases or electronic field notes software. This report is the technical manual for version 2.8 of QRev.

  6. Enhanced Bayesian modelling in BAPS software for learning genetic structures of populations

    Directory of Open Access Journals (Sweden)

    Sirén Jukka

    2008-12-01

    Full Text Available Abstract Background During the most recent decade many Bayesian statistical models and software for answering questions related to the genetic structure underlying population samples have appeared in the scientific literature. Most of these methods utilize molecular markers for the inferences, while some are also capable of handling DNA sequence data. In a number of earlier works, we have introduced an array of statistical methods for population genetic inference that are implemented in the software BAPS. However, the complexity of biological problems related to genetic structure analysis keeps increasing such that in many cases the current methods may provide either inappropriate or insufficient solutions. Results We discuss the necessity of enhancing the statistical approaches to face the challenges posed by the ever-increasing amounts of molecular data generated by scientists over a wide range of research areas and introduce an array of new statistical tools implemented in the most recent version of BAPS. With these methods it is possible, e.g., to fit genetic mixture models using user-specified numbers of clusters and to estimate levels of admixture under a genetic linkage model. Also, alleles representing a different ancestry compared to the average observed genomic positions can be tracked for the sampled individuals, and a priori specified hypotheses about genetic population structure can be directly compared using Bayes' theorem. In general, we have improved further the computational characteristics of the algorithms behind the methods implemented in BAPS facilitating the analyses of large and complex datasets. In particular, analysis of a single dataset can now be spread over multiple computers using a script interface to the software. Conclusion The Bayesian modelling methods introduced in this article represent an array of enhanced tools for learning the genetic structure of populations. Their implementations in the BAPS software are

  7. Selection Criteria for Computer Software and Hardware: A Case Study of Six University Libraries in Nigeria

    Directory of Open Access Journals (Sweden)

    Udoh-Ilomechine Queenette

    2011-12-01

    Full Text Available This paper investigates the criteria used in the selection of computer hardware and software in six university libraries in Nigeria. Six (6 copies of a questionnaire were sent to selected librarians in Edo and Delta states, Nigeria. All copies of the questionnaire were retrieved. The data collected were analyzed. The fin dings reveal that the respondents took into consideration such factors as memory, speed, capacity, durability, costs, reliability and standardization, brand and manufacturer, warranty, and scalability of the system before procuring computer hardware. The respondents also take into consideration the reliability and track record of the vendor, service and technical support, previews or sample sections, compatibility with other program s being used, product cost, and data migration before procuring computer software. It is also noteworthy that the respondents have encountered with electricity failure, improper implementation, and difficulties to get qualified personnel to maintain and/or repair computer hardware when it was broken down.

  8. Development of innovative computer software to facilitate the setup and computation of water quality index

    OpenAIRE

    Nabizadeh, Ramin; Valadi Amin, Maryam; Alimohammadi, Mahmood; Naddafi, Kazem; Mahvi, Amir Hossein; Yousefzadeh, Samira

    2013-01-01

    Background Developing a water quality index which is used to convert the water quality dataset into a single number is the most important task of most water quality monitoring programmes. As the water quality index setup is based on different local obstacles, it is not feasible to introduce a definite water quality index to reveal the water quality level. Findings In this study, an innovative software application, the Iranian Water Quality Index Software (IWQIS), is presented in order to faci...

  9. A Decision Framework for Selecting Licensing Rights for Noncommercial Computer Software in the DoD Environment

    Science.gov (United States)

    2011-07-01

    Funding Source Who is going to fund or has funded the noncommercial computer software development and to what extent? Funding Source refers to the...1. The DoD agency may need to provide all of the funds for the noncommercial computer software development . This Source of Funds example supports...The DoD agency may need to provide some funding, along with the supplier, who is also providing funding for the noncommercial computer software development . This

  10. A surface measure for probabilistic structural computations.

    Science.gov (United States)

    Schmidt, J P; Chen, C C; Cooper, J L; Altman, R B

    1998-01-01

    Computing three-dimensional structures from sparse experimental constraints requires method for combining heterogeneous sources of information, such as distances, angles, and measures of total volume, shape, and surface. For some types of information, such as distances between atoms, numerous methods are available for computing structures that satisfy the provided constraints. It is more difficult, however, to use information about the degree to which an atom is on the surface or buried as a useful constraint during structure computations. Surface measures have been used as accept/reject criteria for previously computed structures, but this is not an efficient strategy. In this paper, we investigate the efficacy of applying a surface measure in the computation of molecular structure, using a method of probabilistic least square computations which facilitates the introduction of multiple, noisy, heterogeneous data sources. For this purpose, we introduce a simple purely geometrical measure of surface proximity called maximal conic view (MCV). MCV is efficiently computable and differentiable, and is hence well suited to driving a structural optimization method based, in part, on surface data. As an initial validation, we show that MCV correlates well with known measures for total exposed surface area. We use this measure in our experiments to show that information about surface proximity (derived from theory or experiment, for example) can be added to a set of distance measurements to increase significantly the quality of the computed structure. In particular, when 30 to 50 percent of all possible short-range distances are provided, the addition of surface information improves the quality of the computed structure (as measured by RMS fit) by as much as 80 percent. Our results demonstrate that knowledge of which atoms are on the surface and which are buried can be used as a powerful constraint in estimating molecular structure.

  11. Development of high performance casting analysis software by coupled parallel computation

    Directory of Open Access Journals (Sweden)

    Sang Hyun CHO

    2007-08-01

    Full Text Available Up to now, so much casting analysis software has been continuing to develop the new access way to real casting processes. Those include the melt flow analysis, heat transfer analysis for solidification calculation, mechanical property predictions and microstructure predictions. These trials were successful to obtain the ideal results comparing with real situations, so that CAE technologies became inevitable to design or develop new casting processes. But for manufacturing fields, CAE technologies are not so frequently being used because of their difficulties in using the software or insufficient computing performances. To introduce CAE technologies to manufacturing field, the high performance analysis is essential to shorten the gap between product designing time and prototyping time. The software code optimization can be helpful, but it is not enough, because the codes developed by software experts are already optimized enough. As an alternative proposal for high performance computations, the parallel computation technologies are eagerly being applied to CAE technologies to make the analysis time shorter. In this research, SMP (Shared Memory Processing and MPI (Message Passing Interface (1 methods for parallelization were applied to commercial software "Z-Cast" to calculate the casting processes. In the code parallelizing processes, the network stabilization, core optimization were also carried out under Microsoft Windows platform and their performances and results were compared with those of normal linear analysis codes.

  12. Development of a Radiation Dose Reporting Software for X-ray Computed Tomography (CT)

    Science.gov (United States)

    Ding, Aiping

    software provides a better CT dose assessment by using anatomically realistic patient phantoms. All the related organ doses are incorporated into a standardized database compiled using Microsoft Structured Query Language (SQL) server 2008. Organ doses from contiguous axial or helical scans defined by a specific protocol can be rapidly obtained from the database. A next-generation software architecture and Active Server Pages (ASP) .NET framework are adopted to create a browser-hosted application to improve the user interactivity and reporting functionality including scanning parameter selection and organ dose reporting. “VirtualDose” has been developed as a web-based CT dose reporting platform to facilitate several important features including: (1) easy access via Internet; (2) no need for installation on the local computer; (3) a user-friendly, dynamic, browser-hosted graphical user interface; (4) use of advanced patient models for the adult male and female, pregnant women, children, and obese patient models; (5) adoption of modern CT scanners and protocols, as well as the most recent ICRP 103 effective dose algorithms; and (6) flexibility to manage and easily upgrade without impacting user’s usage.

  13. CSA06 Computing, Software and Analysis challenge at the Spanish Tier-1 and Tier-2 sites

    CERN Document Server

    Alcaraz, J; Cabrillo, Iban Jose; Colino, Nicanor; Cuevas-Maestro, J; Delgado Peris, Antonio; Fernandez Menendez, Javier; Flix, Jose; García-Abia, Pablo; González-Caballero, I; Hernández, Jose M; Marco, Rafael; Martinez Ruiz del Arbol, Pablo; Matorras, Francisco; Merino, Gonzalo; Rodríguez-Calonge, F J; Vizan Garcia, Jesus Manuel

    2007-01-01

    This note describes the participation of the Spanish centres PIC, CIEMAT and IFCA as Tier-1 and Tier-2 sites in the CMS CSA06 Computing, Software and Analysis challenge. A number of the facilities, services and workflows have been demonstrated at the 2008 25% scale. Very valuable experience has been gained running the complex computing system under realistic conditions at a significant scale. The focus of this note is on presenting achieved results, operational experience and lessons learnt during the challenge.

  14. Real-Time Software Vulnerabilities in Cloud Computing : Challenges and Mitigation Techniques

    OpenAIRE

    Okonoboh, Matthias Aifuobhokhan; Tekkali, Sudhakar

    2011-01-01

    Context: Cloud computing is rapidly emerging in the area of distributed computing. In the meantime, many organizations also attributed the technology to be associated with several business risks which are yet to be resolved. These challenges include lack of adequate security, privacy and legal issues, resource allocation, control over data, system integrity, risk assessment, software vulnerabilities and so on which all have compromising effect in cloud environment. Organizations based their w...

  15. Computer applications for engineering/structural analysis

    Energy Technology Data Exchange (ETDEWEB)

    Zaslawsky, M.; Samaddar, S.K.

    1991-01-01

    Analysts and organizations have a tendency to lock themselves into specific codes with the obvious consequences of not addressing the real problem and thus reaching the wrong conclusion. This paper discusses the role of the analyst in selecting computer codes. The participation and support of a computation division in modifying the source program, configuration management, and pre- and post-processing of codes are among the subjects discussed. Specific examples illustrating the computer code selection process are described in the following problem areas: soil structure interaction, structural analysis of nuclear reactors, analysis of waste tanks where fluid structure interaction is important, analysis of equipment, structure-structure interaction, analysis of the operation of the superconductor supercollider which includes friction and transient temperature, and 3D analysis of the 10-meter telescope being built in Hawaii. Validation and verification of computer codes and their impact on the selection process are also discussed.

  16. Computational predictions of zinc oxide hollow structures

    Science.gov (United States)

    Tuoc, Vu Ngoc; Huan, Tran Doan; Thao, Nguyen Thi

    2018-03-01

    Nanoporous materials are emerging as potential candidates for a wide range of technological applications in environment, electronic, and optoelectronics, to name just a few. Within this active research area, experimental works are predominant while theoretical/computational prediction and study of these materials face some intrinsic challenges, one of them is how to predict porous structures. We propose a computationally and technically feasible approach for predicting zinc oxide structures with hollows at the nano scale. The designed zinc oxide hollow structures are studied with computations using the density functional tight binding and conventional density functional theory methods, revealing a variety of promising mechanical and electronic properties, which can potentially find future realistic applications.

  17. Multiresolution Computation of Conformal Structures of Surfaces

    Directory of Open Access Journals (Sweden)

    Xianfeng Gu

    2003-10-01

    Full Text Available An efficient multiresolution method to compute global conformal structures of nonzero genus triangle meshes is introduced. The homology, cohomology groups of meshes are computed explicitly, then a basis of harmonic one forms and a basis of holomorphic one forms are constructed. A progressive mesh is generated to represent the original surface at different resolutions. The conformal structure is computed for the coarse level first, then used as the estimation for that of the finer level, by using conjugate gradient method it can be refined to the conformal structure of the finer level.

  18. Designing of a Computer Software for Detection of Approximal Caries in Posterior Teeth

    Science.gov (United States)

    Valizadeh, Solmaz; Goodini, Mostafa; Ehsani, Sara; Mohseni, Hadis; Azimi, Fateme; Bakhshandeh, Hooman

    2015-01-01

    Background: Radiographs, adjunct to clinical examination are always valuable complementary methods for dental caries detection. Recently, progressing in digital imaging system provides possibility of software designing for automatically dental caries detection. Objectives: The aim of this study was to develop and assess the function of diagnostic computer software designed for evaluation of approximal caries in posterior teeth. This software should be able to indicate the depth and location of caries on digital radiographic images. Materials and Methods: Digital radiographs were obtained of 93 teeth including 183 proximal surfaces. These images were used as a database for designing the software and training the software designer. In the design phase, considering the summed density of pixels in rows and columns of the images, the teeth were separated from each other and the unnecessary regions; for example, the root area in the alveolar bone was eliminated. Therefore, based on summed intensities, each image was segmented such that each segment contained only one tooth. Subsequently, based on the fuzzy logic, a well-known data-clustering algorithm named fuzzy c-means (FCM) was applied to the images to cluster or segment each tooth. This algorithm is referred to as a soft clustering method, which assigns data elements to one or more clusters with a specific membership function. Using the extracted clusters, the tooth border was determined and assessed for cavity. The results of histological analysis were used as the gold standard for comparison with the results obtained from the software. Depth of caries was measured, and finally Intraclass Correlation Coefficient (ICC) and Bland-Altman plot were used to show the agreement between the methods. Results: The software diagnosed 60% of enamel caries. The ICC (for detection of enamel caries) between the computer software and histological analysis results was determined as 0.609 (95% confidence interval [CI] = 0

  19. Designing of a Computer Software for Detection of Approximal Caries in Posterior Teeth.

    Science.gov (United States)

    Valizadeh, Solmaz; Goodini, Mostafa; Ehsani, Sara; Mohseni, Hadis; Azimi, Fateme; Bakhshandeh, Hooman

    2015-10-01

    Radiographs, adjunct to clinical examination are always valuable complementary methods for dental caries detection. Recently, progressing in digital imaging system provides possibility of software designing for automatically dental caries detection. The aim of this study was to develop and assess the function of diagnostic computer software designed for evaluation of approximal caries in posterior teeth. This software should be able to indicate the depth and location of caries on digital radiographic images. Digital radiographs were obtained of 93 teeth including 183 proximal surfaces. These images were used as a database for designing the software and training the software designer. In the design phase, considering the summed density of pixels in rows and columns of the images, the teeth were separated from each other and the unnecessary regions; for example, the root area in the alveolar bone was eliminated. Therefore, based on summed intensities, each image was segmented such that each segment contained only one tooth. Subsequently, based on the fuzzy logic, a well-known data-clustering algorithm named fuzzy c-means (FCM) was applied to the images to cluster or segment each tooth. This algorithm is referred to as a soft clustering method, which assigns data elements to one or more clusters with a specific membership function. Using the extracted clusters, the tooth border was determined and assessed for cavity. The results of histological analysis were used as the gold standard for comparison with the results obtained from the software. Depth of caries was measured, and finally Intraclass Correlation Coefficient (ICC) and Bland-Altman plot were used to show the agreement between the methods. The software diagnosed 60% of enamel caries. The ICC (for detection of enamel caries) between the computer software and histological analysis results was determined as 0.609 (95% confidence interval [CI] = 0.159-0.849) (P = 0.006). Also, the computer program diagnosed 97% of

  20. Visualization techniques for the analysis of software behavior and related structures

    OpenAIRE

    Trümper, Jonas

    2014-01-01

    Software maintenance encompasses any changes made to a software system after its initial deployment and is thereby one of the key phases in the typical software-engineering lifecycle. In software maintenance, we primarily need to understand structural and behavioral aspects, which are difficult to obtain, e.g., by code reading. Software analysis is therefore a vital tool for maintaining these systems: It provides - the preferably automated - means to extract and evaluate information from thei...

  1. Impact of anti-virus software on computer virus dynamical behavior

    Science.gov (United States)

    Sun, Mei; Li, Dandan; Han, Dun; Jia, Changsheng

    2014-11-01

    The impact of anti-virus software on the spreading of computer virus is investigated via developing a mathematical model in this paper. Considering the anti-virus software may not be effective, as it may be an outdated version, and then the computers may be infected with a reduced incidence rate. According to the method of next generation matrix, the basic reproduction number is derived. By introducing appropriate Lyapunov function and the Routh stability criterion, acquiring the stability conditions of the virus-free equilibrium and virus equilibrium. The effect of anti-virus software and disconnecting rate on the spreading of virus are also analyzed. When combined with the numerical results, a set of suggestions are put forward for eradicating virus effectively.

  2. A computer-aided software-tool for sustainable process synthesis-intensification

    DEFF Research Database (Denmark)

    Kumar Tula, Anjan; Babi, Deenesh K.; Bottlaender, Jack

    2017-01-01

    and determine within the design space, the more sustainable processes. In this paper, an integrated computer-aided software-tool that searches the design space for hybrid/intensified more sustainable process options is presented. Embedded within the software architecture are process synthesis......Currently, the process industry is moving towards the design of innovative, more sustainable processes that show improvements in both economic and environmental factors. The design space of unit operations that can be combined to generate process flowsheet alternatives considering known unit...... constraints while also matching the design targets, they are therefore more sustainable than the base case. The application of the software-tool to the production of biodiesel is presented, highlighting the main features of the computer-aided, multi-stage, multi-scale methods that are able to determine more...

  3. The future of commodity computing and many-core versus the interests of HEP software

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    As the mainstream computing world has shifted from multi-core to many-core platforms, the situation for software developers has changed as well. With the numerous hardware and software options available, choices balancing programmability and performance are becoming a significant challenge. The expanding multiplicative dimensions of performance offer a growing number of possibilities that need to be assessed and addressed on several levels of abstraction. This paper reviews the major tradeoffs forced upon the software domain by the changing landscape of parallel technologies – hardware and software alike. Recent developments, paradigms and techniques are considered with respect to their impact on the rather traditional HEP programming models. Other considerations addressed include aspects of efficiency and reasonably achievable targets for the parallelization of large scale HEP workloads.

  4. SCRAM: Software configuration and management for the LHC Computing Grid project

    CERN Document Server

    Wellisch, J P; Ashby, S

    2003-01-01

    Recently SCRAM (Software Configuration And Management) has been adopted by the applications area of the LHC computing grid project as baseline configuration management and build support infrastructure tool. SCRAM is a software engineering tool, that supports the configuration management and management processes for software development. It resolves the issues of configuration definition, assembly break-down, build, project organization, run-time environment, installation, distribution, deployment, and source code distribution. It was designed with a focus on supporting a distributed, multi-project development work-model. We will describe the underlying technology, and the solutions SCRAM offers to the above software engineering processes, while taking a users view of the system under configuration management.

  5. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan part 2 mappings for the ASC software quality engineering practices, version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR001.3.2 and CPR001.3.6 and to a Department of Energy document, ''ASCI Software Quality Engineering: Goals, Principles, and Guidelines''. This document also identifies ASC management and software project teams' responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  6. Journal of Chemical Education: Software: Abstract of "The Computer-Based Laboratory."

    Science.gov (United States)

    Krause, Daniel C.; Divis, Lynne M., Ed.

    1988-01-01

    Describes a chemistry laboratory software package for interfacing the Apple IIe for high school and introductory college courses. Topics include: thermistor calibration, phase change, heat of reaction, freezing point depression, Beer's law, and color decay in crystal violet. Explains the computer interface and the tools needed. (MVL)

  7. Technology survey of computer software as applicable to the MIUS project

    Science.gov (United States)

    Fulbright, B. E.

    1975-01-01

    Existing computer software, available from either governmental or private sources, applicable to modular integrated utility system program simulation is surveyed. Several programs and subprograms are described to provide a consolidated reference, and a bibliography is included. The report covers the two broad areas of design simulation and system simulation.

  8. Goethe Gossips with Grass: Using Computer Chatting Software in an Introductory Literature Course.

    Science.gov (United States)

    Fraser, Catherine C.

    1999-01-01

    Students in a third-year introduction to German literature course chatted over networked computers, using "FirstClass" software. A brief description of the course design is provided with detailed information on how the three chat sessions were organized. (Author/VWL)

  9. Comparing Results from Constant Comparative and Computer Software Methods: A Reflection about Qualitative Data Analysis

    Science.gov (United States)

    Putten, Jim Vander; Nolen, Amanda L.

    2010-01-01

    This study compared qualitative research results obtained by manual constant comparative analysis with results obtained by computer software analysis of the same data. An investigated about issues of trustworthiness and accuracy ensued. Results indicated that the inductive constant comparative data analysis generated 51 codes and two coding levels…

  10. 77 FR 50727 - Configuration Management Plans for Digital Computer Software Used in Safety Systems of Nuclear...

    Science.gov (United States)

    2012-08-22

    ... revision 2 of RG 1.168, ``Verification, Validation, Reviews, and Audits for Digital Computer Software used... NRC-2012- 0195. You may submit comments by any of the following methods: Federal Rulemaking Web Site... possesses and are publicly available, by any of the following methods: Federal Rulemaking Web Site: Go to...

  11. DVS-SOFTWARE: An Effective Tool for Applying Highly Parallelized Hardware To Computational Geophysics

    Science.gov (United States)

    Herrera, I.; Herrera, G. S.

    2015-12-01

    Most geophysical systems are macroscopic physical systems. The behavior prediction of such systems is carried out by means of computational models whose basic models are partial differential equations (PDEs) [1]. Due to the enormous size of the discretized version of such PDEs it is necessary to apply highly parallelized super-computers. For them, at present, the most efficient software is based on non-overlapping domain decomposition methods (DDM). However, a limiting feature of the present state-of-the-art techniques is due to the kind of discretizations used in them. Recently, I. Herrera and co-workers using 'non-overlapping discretizations' have produced the DVS-Software which overcomes this limitation [2]. The DVS-software can be applied to a great variety of geophysical problems and achieves very high parallel efficiencies (90%, or so [3]). It is therefore very suitable for effectively applying the most advanced parallel supercomputers available at present. In a parallel talk, in this AGU Fall Meeting, Graciela Herrera Z. will present how this software is being applied to advance MOD-FLOW. Key Words: Parallel Software for Geophysics, High Performance Computing, HPC, Parallel Computing, Domain Decomposition Methods (DDM)REFERENCES [1]. Herrera Ismael and George F. Pinder, Mathematical Modelling in Science and Engineering: An axiomatic approach", John Wiley, 243p., 2012. [2]. Herrera, I., de la Cruz L.M. and Rosas-Medina A. "Non Overlapping Discretization Methods for Partial, Differential Equations". NUMER METH PART D E, 30: 1427-1454, 2014, DOI 10.1002/num 21852. (Open source) [3]. Herrera, I., & Contreras Iván "An Innovative Tool for Effectively Applying Highly Parallelized Software To Problems of Elasticity". Geofísica Internacional, 2015 (In press)

  12. ELSI: A unified software interface for Kohn-Sham electronic structure solvers

    Science.gov (United States)

    Yu, Victor Wen-zhe; Corsetti, Fabiano; García, Alberto; Huhn, William P.; Jacquelin, Mathias; Jia, Weile; Lange, Björn; Lin, Lin; Lu, Jianfeng; Mi, Wenhui; Seifitokaldani, Ali; Vázquez-Mayagoitia, Álvaro; Yang, Chao; Yang, Haizhao; Blum, Volker

    2018-01-01

    Solving the electronic structure from a generalized or standard eigenproblem is often the bottleneck in large scale calculations based on Kohn-Sham density-functional theory. This problem must be addressed by essentially all current electronic structure codes, based on similar matrix expressions, and by high-performance computation. We here present a unified software interface, ELSI, to access different strategies that address the Kohn-Sham eigenvalue problem. Currently supported algorithms include the dense generalized eigensolver library ELPA, the orbital minimization method implemented in libOMM, and the pole expansion and selected inversion (PEXSI) approach with lower computational complexity for semilocal density functionals. The ELSI interface aims to simplify the implementation and optimal use of the different strategies, by offering (a) a unified software framework designed for the electronic structure solvers in Kohn-Sham density-functional theory; (b) reasonable default parameters for a chosen solver; (c) automatic conversion between input and internal working matrix formats, and in the future (d) recommendation of the optimal solver depending on the specific problem. Comparative benchmarks are shown for system sizes up to 11,520 atoms (172,800 basis functions) on distributed memory supercomputing architectures.

  13. The structural robustness of multiprocessor computing system

    Directory of Open Access Journals (Sweden)

    N. Andronaty

    1996-03-01

    Full Text Available The model of the multiprocessor computing system on the base of transputers which permits to resolve the question of valuation of a structural robustness (viability, survivability is described.

  14. Tensegrity structures - Computational and experimental tensegrity mechanics

    Science.gov (United States)

    Kuhl, Detlef; Lim, Yi Chung; Long, David S.

    2017-07-01

    The present paper deals with tensegrity structures. We review the definition of tensegrity structures, and describe both experimental and computational form finding methods. Also described are the numerical methods for the simulation of prestress induced stiffness, and the static and dynamic structural analyses. Furthermore, we present laboratory models and measurement methods for identifying the realized geometry and prestress state. Finally, computationally and experimentally obtained geometries and prestress states are compared, a representative realization of a real world tensegrity tower is shown and the modeling of biological cells as tensegrity structures is adressed.

  15. The Computer-based Health Evaluation Software (CHES: a software for electronic patient-reported outcome monitoring

    Directory of Open Access Journals (Sweden)

    Holzner Bernhard

    2012-11-01

    Full Text Available Abstract Background Patient-reported Outcomes (PROs capturing e.g., quality of life, fatigue, depression, medication side-effects or disease symptoms, have become important outcome parameters in medical research and daily clinical practice. Electronic PRO data capture (ePRO with software packages to administer questionnaires, storing data, and presenting results has facilitated PRO assessment in hospital settings. Compared to conventional paper-pencil versions of PRO instruments, ePRO is more economical with regard to staff resources and time, and allows immediate presentation of results to the medical staff. The objective of our project was to develop software (CHES – Computer-based Health Evaluation System for ePRO in hospital settings and at home with a special focus on the presentation of individual patient’s results. Methods Following the Extreme Programming development approach architecture was not fixed up-front, but was done in close, continuous collaboration with software end users (medical staff, researchers and patients to meet their specific demands. Developed features include sophisticated, longitudinal charts linking patients’ PRO data to clinical characteristics and to PRO scores from reference populations, a web-interface for questionnaire administration, and a tool for convenient creating and editing of questionnaires. Results By 2012 CHES has been implemented at various institutions in Austria, Germany, Switzerland, and the UK and about 5000 patients participated in ePRO (with around 15000 assessments in total. Data entry is done by the patients themselves via tablet PCs with a study nurse or an intern approaching patients and supervising questionnaire completion. Discussion During the last decade several software packages for ePRO have emerged for different purposes. Whereas commercial products are available primarily for ePRO in clinical trials, academic projects have focused on data collection and presentation in daily

  16. The Computer-based Health Evaluation Software (CHES): a software for electronic patient-reported outcome monitoring

    Science.gov (United States)

    2012-01-01

    Background Patient-reported Outcomes (PROs) capturing e.g., quality of life, fatigue, depression, medication side-effects or disease symptoms, have become important outcome parameters in medical research and daily clinical practice. Electronic PRO data capture (ePRO) with software packages to administer questionnaires, storing data, and presenting results has facilitated PRO assessment in hospital settings. Compared to conventional paper-pencil versions of PRO instruments, ePRO is more economical with regard to staff resources and time, and allows immediate presentation of results to the medical staff. The objective of our project was to develop software (CHES – Computer-based Health Evaluation System) for ePRO in hospital settings and at home with a special focus on the presentation of individual patient’s results. Methods Following the Extreme Programming development approach architecture was not fixed up-front, but was done in close, continuous collaboration with software end users (medical staff, researchers and patients) to meet their specific demands. Developed features include sophisticated, longitudinal charts linking patients’ PRO data to clinical characteristics and to PRO scores from reference populations, a web-interface for questionnaire administration, and a tool for convenient creating and editing of questionnaires. Results By 2012 CHES has been implemented at various institutions in Austria, Germany, Switzerland, and the UK and about 5000 patients participated in ePRO (with around 15000 assessments in total). Data entry is done by the patients themselves via tablet PCs with a study nurse or an intern approaching patients and supervising questionnaire completion. Discussion During the last decade several software packages for ePRO have emerged for different purposes. Whereas commercial products are available primarily for ePRO in clinical trials, academic projects have focused on data collection and presentation in daily clinical practice and

  17. Primary Health Care Software-A Computer Based Data Management System

    Directory of Open Access Journals (Sweden)

    Tuli K

    1990-01-01

    Full Text Available Realising the duplication and time consumption in the usual manual system of data collection necessitated experimentation with computer based management system for primary health care in the primary health centers. The details of the population as available in the existing manual system were used for computerizing the data. Software was designed for data entry and analysis. It was written in Dbase III plus language. It was so designed that a person with no knowledge about computer could use it, A cost analysis was done and the computer system was found more cost effective than the usual manual system.

  18. Effect of anti-virus software on infectious nodes in computer network: A mathematical model

    Science.gov (United States)

    Mishra, Bimal Kumar; Pandey, Samir Kumar

    2012-07-01

    An e-epidemic model of malicious codes in the computer network through vertical transmission is formulated. We have observed that if the basic reproduction number is less than unity, the infected proportion of computer nodes disappear and malicious codes die out and also the malicious codes-free equilibrium is globally asymptotically stable which leads to its eradication. Effect of anti-virus software on the removal of the malicious codes from the computer network is critically analyzed. Analysis and simulation results show some managerial insights that are helpful for the practice of anti-virus in information sharing networks.

  19. Software designs of image processing tasks with incremental refinement of computation.

    Science.gov (United States)

    Anastasia, Davide; Andreopoulos, Yiannis

    2010-08-01

    Software realizations of computationally-demanding image processing tasks (e.g., image transforms and convolution) do not currently provide graceful degradation when their clock-cycles budgets are reduced, e.g., when delay deadlines are imposed in a multitasking environment to meet throughput requirements. This is an important obstacle in the quest for full utilization of modern programmable platforms' capabilities since worst-case considerations must be in place for reasonable quality of results. In this paper, we propose (and make available online) platform-independent software designs performing bitplane-based computation combined with an incremental packing framework in order to realize block transforms, 2-D convolution and frame-by-frame block matching. The proposed framework realizes incremental computation: progressive processing of input-source increments improves the output quality monotonically. Comparisons with the equivalent nonincremental software realization of each algorithm reveal that, for the same precision of the result, the proposed approach can lead to comparable or faster execution, while it can be arbitrarily terminated and provide the result up to the computed precision. Application examples with region-of-interest based incremental computation, task scheduling per frame, and energy-distortion scalability verify that our proposal provides significant performance scalability with graceful degradation.

  20. KD-ACP: A Software Framework for Social Computing in Emergency Management

    Directory of Open Access Journals (Sweden)

    Bin Chen

    2015-01-01

    Full Text Available This paper addresses the application of a computational theory and related techniques for studying emergency management in social computing. We propose a novel software framework called KD-ACP. The framework provides a systematic and automatic platform for scientists to study the emergency management problems in three aspects: modelling the society in emergency scenario as the artificial society; investigating the emergency management problems by the repeat computational experiments; parallel execution between artificial society and the actual society managed by the decisions from computational experiments. The software framework is composed of a series of tools. These tools are categorized into three parts corresponding to “A,” “C,” and “P,” respectively. Using H1N1 epidemic in Beijing city as the case study, the modelling and data generating of Beijing city, experiments with settings of H1N1, and intervention measures and parallel execution by situation tool are implemented by KD-ACP. The results output by the software framework shows that the emergency response decisions can be tested to find a more optimal one through the computational experiments. In the end, the advantages of the KD-ACP and the future work are summarized in the conclusion.

  1. Mapping Social Network to Software Architecture to Detect Structure Clashes in Agile Software Development

    NARCIS (Netherlands)

    Amrit, Chintan Amrit; van Hillegersberg, Jos

    2007-01-01

    Software development is rarely an individual effort and generally involves teams of developers collaborating together in order to generate reliable code. Such collaborations require proper communication and regular coordination among the team members. In addition, coordination is required to sort

  2. FiD: a software for ab initio structural identification of product ions from tandem mass spectrometric data.

    Science.gov (United States)

    Heinonen, Markus; Rantanen, Ari; Mielikäinen, Taneli; Kokkonen, Juha; Kiuru, Jari; Ketola, Raimo A; Rousu, Juho

    2008-10-01

    We present FiD (Fragment iDentificator), a software tool for the structural identification of product ions produced with tandem mass spectrometric measurement of low molecular weight organic compounds. Tandem mass spectrometry (MS/MS) has proven to be an indispensable tool in modern, cell-wide metabolomics and fluxomics studies. In such studies, the structural information of the MS(n) product ions is usually needed in the downstream analysis of the measurement data. The manual identification of the structures of MS(n) product ions is, however, a nontrivial task requiring expertise, and calls for computer assistance. Commercial software tools, such as Mass Frontier and ACD/MS Fragmenter, rely on fragmentation rule databases for the identification of MS(n) product ions. FiD, on the other hand, conducts a combinatorial search over all possible fragmentation paths and outputs a ranked list of alternative structures. This gives the user an advantage in situations where the MS/MS data of compounds with less well-known fragmentation mechanisms are processed. FiD software implements two fragmentation models, the single-step model that ignores intermediate fragmentation states and the multi-step model, which allows for complex fragmentation pathways. The software works for MS/MS data produced both in positive- and negative-ion modes. The software has an easy-to-use graphical interface with built-in visualization capabilities for structures of product ions and fragmentation pathways. In our experiments involving amino acids and sugar-phosphates, often found, e.g., in the central carbon metabolism of yeasts, FiD software correctly predicted the structures of product ions on average in 85% of the cases. The FiD software is free for academic use and is available for download from www.cs.helsinki.fi/group/sysfys/software/fragid.

  3. ODEion--a software module for structural identification of ordinary differential equations.

    Science.gov (United States)

    Gennemark, Peter; Wedelin, Dag

    2014-02-01

    In the systems biology field, algorithms for structural identification of ordinary differential equations (ODEs) have mainly focused on fixed model spaces like S-systems and/or on methods that require sufficiently good data so that derivatives can be accurately estimated. There is therefore a lack of methods and software that can handle more general models and realistic data. We present ODEion, a software module for structural identification of ODEs. Main characteristic features of the software are: • The model space is defined by arbitrary user-defined functions that can be nonlinear in both variables and parameters, such as for example chemical rate reactions. • ODEion implements computationally efficient algorithms that have been shown to efficiently handle sparse and noisy data. It can run a range of realistic problems that previously required a supercomputer. • ODEion is easy to use and provides SBML output. We describe the mathematical problem, the ODEion system itself, and provide several examples of how the system can be used. Available at: http://www.odeidentification.org.

  4. Computer-assisted neurosurgery system: Wayne State University hardware and software configuration.

    Science.gov (United States)

    Zamorano, L; Jiang, Z; Kadi, A M

    1994-01-01

    Computer-assisted neurosurgery uses the latest technological advancements in imaging, computers, mechanics, and electronics to improve the accuracy and reduce the invasiveness and risk of neurosurgical procedures. We describe the Wayne State University, Detroit, Michigan, computer-assisted neurosurgical system with the emphasis on software and discuss the theory guiding the development of this system and its application in real-time position tracking systems. Our system consists of the Neurological Surgery Planning System (NSPS) software which we developed at our medical center and three types of position tracking systems: the Zamorano-Dujovny (Z-D) are digitizer for frame-based procedures, an articulated arm, and an infrared-based digitizer for frameless procedures. The NSPS software is designed to offer neurosurgeons a safe and accurate method to approach intracranial lesions by preoperatively planning a surgical trajectory. Software consisting of the most advanced technologies in computer vision, computer imaging/graphics, and stereotactic numeric analysis forms the core of the system. Capabilities for correlating data from imaging studies to facilitate image reconstruction, image mapping, and three-dimensional (3D) visualization of target volumes enable the neurosurgeon to simulate surgical procedures into a preoperative protocol to be used during surgery, both to follow the preplanned trajectory and to track the position of surgical instruments in real-time on the computer monitor. The tracking systems position and orient the surgical instruments relative to the patient's head. With these devices, the display of the surgical instruments together with the virtual images create an excellent intraoperative tool.

  5. SU-F-I-43: A Software-Based Statistical Method to Compute Low Contrast Detectability in Computed Tomography Images

    Energy Technology Data Exchange (ETDEWEB)

    Chacko, M; Aldoohan, S [University of Oklahoma Health Sciences Center, Oklahoma City, OK (United States)

    2016-06-15

    Purpose: The low contrast detectability (LCD) of a CT scanner is its ability to detect and display faint lesions. The current approach to quantify LCD is achieved using vendor-specific methods and phantoms, typically by subjectively observing the smallest size object at a contrast level above phantom background. However, this approach does not yield clinically applicable values for LCD. The current study proposes a statistical LCD metric using software tools to not only to assess scanner performance, but also to quantify the key factors affecting LCD. This approach was developed using uniform QC phantoms, and its applicability was then extended under simulated clinical conditions. Methods: MATLAB software was developed to compute LCD using a uniform image of a QC phantom. For a given virtual object size, the software randomly samples the image within a selected area, and uses statistical analysis based on Student’s t-distribution to compute the LCD as the minimal Hounsfield Unit’s that can be distinguished from the background at the 95% confidence level. Its validity was assessed by comparison with the behavior of a known QC phantom under various scan protocols and a tissue-mimicking phantom. The contributions of beam quality and scattered radiation upon the computed LCD were quantified by using various external beam-hardening filters and phantom lengths. Results: As expected, the LCD was inversely related to object size under all scan conditions. The type of image reconstruction kernel filter and tissue/organ type strongly influenced the background noise characteristics and therefore, the computed LCD for the associated image. Conclusion: The proposed metric and its associated software tools are vendor-independent and can be used to analyze any LCD scanner performance. Furthermore, the method employed can be used in conjunction with the relationships established in this study between LCD and tissue type to extend these concepts to patients’ clinical CT

  6. A NEW CONTROL CIRCUIT AND COMPUTER SOFTWARE FOR CONTROLING PHOTOVOLTAIC SYSTEMS

    Directory of Open Access Journals (Sweden)

    Mustafa Berkant SELEK

    2008-02-01

    Full Text Available In this study, a new microcontroller circuit was designed and new computer software was implemented to control power flow currents of renewable energy system, which is established in Solar Energy Institute, Ege University, Bornova, Izmir, Turkey. PIC18F452 microcontroller based electronic circuit was designed to control another electronic circuit that includes power electronic switching components. Readily available standard control circuits are designed for switching single level inverters. In contrary, implemented circuit allows to switch multilevel inverters. In addition, because the efficiency of solar energy panels is considerably low, solar panels should be operated under the maximum power point (MPP. Therefore, MPP algorithm is included in the designed control circuit. Next, the control circuit also includes a serial communication interface based on RS232 standard. Using this interface enables the user to choose all functions available in the control circuit and take status report via computer software. Last, a general purpose command set was designed to establish communication between the computer software and the microcontroller-based control circuit. As a result, it is aimed that this study supply a basis for the researchers who want to develop own control circuits or more visual software.

  7. Advances in Single-Photon Emission Computed Tomography Hardware and Software.

    Science.gov (United States)

    Piccinelli, Marina; Garcia, Ernest V

    2016-02-01

    Nuclear imaging techniques remain today's most reliable modality for the assessment and quantification of myocardial perfusion. In recent years, the field has experienced tremendous progress both in terms of dedicated cameras for cardiac applications and software techniques for image reconstruction. The most recent advances in single-photon emission computed tomography hardware and software are reviewed, focusing on how these improvements have resulted in an even more powerful diagnostic tool with reduced injected radiation dose and acquisition time. Copyright © 2016 Elsevier Inc. All rights reserved.

  8. Interactive computer graphics system for structural sizing and analysis of aircraft structures

    Science.gov (United States)

    Bendavid, D.; Pipano, A.; Raibstein, A.; Somekh, E.

    1975-01-01

    A computerized system for preliminary sizing and analysis of aircraft wing and fuselage structures was described. The system is based upon repeated application of analytical program modules, which are interactively interfaced and sequence-controlled during the iterative design process with the aid of design-oriented graphics software modules. The entire process is initiated and controlled via low-cost interactive graphics terminals driven by a remote computer in a time-sharing mode.

  9. Hardware and software and machine-tool simulation with parallel structures mechanisms

    Directory of Open Access Journals (Sweden)

    Keba P.V.

    2016-12-01

    Full Text Available The usage spectrum of mechanisms with parallel structure is spreading all the time. The mechanisms of machine-tools and manipulators become more complicated and it is necessary to improve the program-controlled modules. Closed circuit mechanisms are mostly spread in robotic complexes, where manipulator performs complicated spatial movements by the given trajectory. The usage spectrum is very wide and the most popular are sorting, welding, assembling and others. However, the problem of designing the operating programs is still present even today. It is just because the developed post-processors are created for the equipment that we have for now. But new machine tool constructions appear every day and there is a necessity to control them. The problems associated with using of hardware and software of mechanisms with parallel structure in computer-aided simulation are considered. The program for inverse problem kinematics solving is designed. New method of designing the control programs is found. The kinematic analysis methods options and calculated data obtained by computer mathematics systems are shown with «Tools Glide» software taken as an example.

  10. A NEW DIGITAL IMAGE CORRELATION SOFTWARE FOR DISPLACEMENTS FIELD MEASUREMENT IN STRUCTURAL APPLICATIONS

    Directory of Open Access Journals (Sweden)

    R. Ravanelli

    2017-07-01

    Full Text Available Recently, there has been a growing interest in studying non-contact techniques for strain and displacement measurement. Within photogrammetry, Digital Image Correlation (DIC has received particular attention thanks to the recent advances in the field of lowcost, high resolution digital cameras, computer power and memory storage. DIC is indeed an optical technique able to measure full field displacements and strain by comparing digital images of the surface of a material sample at different stages of deformation and thus can play a major role in structural monitoring applications. For all these reasons, a free and open source 2D DIC software, named py2DIC, was developed at the Geodesy and Geomatics Division of DICEA, University of Rome La Sapienza. Completely written in python, the software is based on the template matching method and computes the displacement and strain fields. The potentialities of Py2DIC were evaluated by processing the images captured during a tensile test performed in the Lab of Structural Engineering, where three different Glass Fiber Reinforced Polymer samples were subjected to a controlled tension by means of a universal testing machine. The results, compared with the values independently measured by several strain gauges fixed on the samples, demonstrate the possibility to successfully characterize the deformation mechanism of the investigated material. Py2DIC is indeed able to highlight displacements at few microns level, in reasonable agreement with the reference, both in terms of displacements (again, at few microns in the average and Poisson’s module.

  11. THE TECHNIQUE OF ANALYSIS OF SOFTWARE OF ON-BOARD COMPUTERS OF AIR VESSEL TO ABSENCE OF UNDECLARED CAPABILITIES BY SIGNATURE-HEURISTIC WAY

    Directory of Open Access Journals (Sweden)

    Viktor Ivanovich Petrov

    2017-01-01

    Full Text Available The article considers the issues of civil aviation aircraft onboard computers data safety. Infor- mation security undeclared capabilities stand for technical equipment or software possibilities, which are not mentioned in the documentation. Documentation and tests content requirements are imposed during the software certification. Documentation requirements include documents composition and content of control (specification, description and program code, the source code. Test requirements include: static analysis of program codes (including the compliance of the sources with their loading modules monitoring; dynamic analysis of source code (including implementation of routes monitor- ing. Currently, there are no complex measures for checking onboard computer software. There are no rules and regulations that can allow controlling foreign production aircraft software, and the actual receiving of software is difficult. Consequently, the author suggests developing the basics of aviation rules and regulations, which allow to analyze the programs of CA aircraft onboard computers. If there are no software source codes the two approaches of code analysis are used: a structural static and dy- namic analysis of the source code; signature-heuristic analysis of potentially dangerous operations. Static analysis determines the behavior of the program by reading the program code (without running the program which is represented in the assembler language - disassembly listing. Program tracing is performed by the dynamic analysis. The analysis of aircraft software ability to detect undeclared capa- bilities using the interactive disassembler was considered in this article.

  12. Soil structure characterized using computed tomographic images

    Science.gov (United States)

    Zhanqi Cheng; Stephen H. Anderson; Clark J. Gantzer; J. W. Van Sambeek

    2003-01-01

    Fractal analysis of soil structure is a relatively new method for quantifying the effects of management systems on soil properties and quality. The objective of this work was to explore several methods of studying images to describe and quantify structure of soils under forest management. This research uses computed tomography and a topological method called Multiple...

  13. Data Structures in Classical and Quantum Computing

    NARCIS (Netherlands)

    M.J. Fillinger (Max)

    2013-01-01

    textabstractThis survey summarizes several results about quantum computing related to (mostly static) data structures. First, we describe classical data structures for the set membership and the predecessor search problems: Perfect Hash tables for set membership by Fredman, Koml\\'{o}s and

  14. The Integration of Computer Software with Printed Materials To Enhance the Reading Skills of Middle School Students.

    Science.gov (United States)

    Dixon, Gaynelle O.

    A practicum described a strategy involving the integration of computer software reading games with traditional printed texts to improve student reading scores and attitudes toward reading. A target group of five seventh-grade remedial reading students participated in the study. The two computer software programs, "Super Solvers Midnight…

  15. [Hardware-software system for monitoring parameters and characteristics of X-ray computer tomographs under operation conditions].

    Science.gov (United States)

    Blinov, N N; Zelikman, M I; Kruchinin, S A

    2007-01-01

    The results of testing of hardware and software for monitoring parameters (mean number of CT units, noise, field uniformity, high-contrast spatial resolution, layer width, dose) and characteristics (modulation transfer function) of X-ray computer tomographs are presented. The developed hardware and software are used to monitor the stability of X-ray computer tomograph parameters under operation conditions.

  16. 77 FR 50720 - Test Documentation for Digital Computer Software Used in Safety Systems of Nuclear Power Plants

    Science.gov (United States)

    2012-08-22

    ... COMMISSION Test Documentation for Digital Computer Software Used in Safety Systems of Nuclear Power Plants... Nuclear Power Plants.'' The DG-1207 is proposed Revision 1 of RG 1.170, dated September 1997. This... for Digital Computer Software Used in Safety Systems of Nuclear Power Plants'' is temporarily...

  17. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1 : ASC software quality engineering practices version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Minana, Molly A.; Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in DOE/AL Quality Criteria (QC-1) as conformance to customer requirements and expectations. This quality plan defines the ASC program software quality practices and provides mappings of these practices to the SNL Corporate Process Requirements (CPR 1.3.2 and CPR 1.3.6) and the Department of Energy (DOE) document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines (GP&G). This quality plan identifies ASC management and software project teams' responsibilities for cost-effective software engineering quality practices. The SNL ASC Software Quality Plan establishes the signatories commitment to improving software products by applying cost-effective software engineering quality practices. This document explains the project teams opportunities for tailoring and implementing the practices; enumerates the practices that compose the development of SNL ASC's software products; and includes a sample assessment checklist that was developed based upon the practices in this document.

  18. Review Essay: Guidance in the World of Computer-Assisted Qualitative Data Analysis Software (CAQDAS Programs

    Directory of Open Access Journals (Sweden)

    Áine Humble

    2015-03-01

    Full Text Available This review discusses Christina SILVER and Ann LEWINS' book, "Using Software in Qualitative Research: A Step-by-Step Guide" (2nd ed.. This book is an impressive undertaking, with online supplemental material in the form of three data sets consisting of many different types of data, detailed instructions for seven CAQDAS (Computer-Assisted Qualitative Data Analysis Software programs, and full-color reproductions of illustrations from the book. The 14 chapters in the book cover a wide range of analysis issues when working with software programs, and the authors encourage critical use of such tools. Readers will benefit from engaging with the online supplemental tools. URN: http://nbn-resolving.de/urn:nbn:de:0114-fqs1502223

  19. Usefulness of Cone-Beam Computed Tomography and Automatic Vessel Detection Software in Emergency Transarterial Embolization

    Energy Technology Data Exchange (ETDEWEB)

    Carrafiello, Gianpaolo, E-mail: gcarraf@gmail.com; Ierardi, Anna Maria, E-mail: amierardi@yahoo.it; Duka, Ejona, E-mail: ejonaduka@hotmail.com [Insubria University, Department of Radiology, Interventional Radiology (Italy); Radaelli, Alessandro, E-mail: alessandro.radaelli@philips.com [Philips Healthcare (Netherlands); Floridi, Chiara, E-mail: chiara.floridi@gmail.com [Insubria University, Department of Radiology, Interventional Radiology (Italy); Bacuzzi, Alessandro, E-mail: alessandro.bacuzzi@ospedale.varese.it [University of Insubria, Anaesthesia and Palliative Care (Italy); Bucourt, Maximilian de, E-mail: maximilian.de-bucourt@charite.de [Charité - University Medicine Berlin, Department of Radiology (Germany); Marchi, Giuseppe De, E-mail: giuseppedemarchi@email.it [Insubria University, Department of Radiology, Interventional Radiology (Italy)

    2016-04-15

    BackgroundThis study was designed to evaluate the utility of dual phase cone beam computed tomography (DP-CBCT) and automatic vessel detection (AVD) software to guide transarterial embolization (TAE) of angiographically challenging arterial bleedings in emergency settings.MethodsTwenty patients with an arterial bleeding at computed tomography angiography and an inconclusive identification of the bleeding vessel at the initial 2D angiographic series were included. Accuracy of DP-CBCT and AVD software were defined as the ability to detect the bleeding site and the culprit arterial bleeder, respectively. Technical success was defined as the correct positioning of the microcatheter using AVD software. Clinical success was defined as the successful embolization. Total volume of iodinated contrast medium and overall procedure time were registered.ResultsThe bleeding site was not detected by initial angiogram in 20 % of cases, while impossibility to identify the bleeding vessel was the reason for inclusion in the remaining cases. The bleeding site was detected by DP-CBCT in 19 of 20 (95 %) patients; in one case CBCT-CT fusion was required. AVD software identified the culprit arterial branch in 18 of 20 (90 %) cases. In two cases, vessel tracking required manual marking of the candidate arterial bleeder. Technical success was 95 %. Successful embolization was achieved in all patients. Mean contrast volume injected for each patient was 77.5 ml, and mean overall procedural time was 50 min.ConclusionsC-arm CBCT and AVD software during TAE of angiographically challenging arterial bleedings is feasible and may facilitate successful embolization. Staff training in CBCT imaging and software manipulation is necessary.

  20. Use of Monte Carlo simulation software for calculating effective dose in cone beam computed tomography

    Energy Technology Data Exchange (ETDEWEB)

    Gomes B, W. O., E-mail: wilsonottobatista@gmail.com [Instituto Federal da Bahia, Rua Emidio dos Santos s/n, Barbalho 40301-015, Salvador de Bahia (Brazil)

    2016-10-15

    This study aimed to develop a geometry of irradiation applicable to the software PCXMC and the consequent calculation of effective dose in applications of the Computed Tomography Cone Beam (CBCT). We evaluated two different CBCT equipment s for dental applications: Care stream Cs 9000 3-dimensional tomograph; i-CAT and GENDEX GXCB-500. Initially characterize each protocol measuring the surface kerma input and the product kerma air-area, P{sub KA}, with solid state detectors RADCAL and PTW transmission chamber. Then we introduce the technical parameters of each preset protocols and geometric conditions in the PCXMC software to obtain the values of effective dose. The calculated effective dose is within the range of 9.0 to 15.7 μSv for 3-dimensional computer 9000 Cs; within the range 44.5 to 89 μSv for GXCB-500 equipment and in the range of 62-111 μSv for equipment Classical i-CAT. These values were compared with results obtained dosimetry using TLD implanted in anthropomorphic phantom and are considered consistent. Os effective dose results are very sensitive to the geometry of radiation (beam position in mathematical phantom). This factor translates to a factor of fragility software usage. But it is very useful to get quick answers to regarding process optimization tool conclusions protocols. We conclude that use software PCXMC Monte Carlo simulation is useful assessment protocols for CBCT tests in dental applications. (Author)

  1. Madagascar: open-source software project for multidimensional data analysis and reproducible computational experiments

    Directory of Open Access Journals (Sweden)

    Sergey Fomel

    2013-11-01

    Full Text Available The Madagascar software package is designed for analysis of large-scale multidimensional data, such as those occurring in exploration geophysics. Madagascar provides a framework for reproducible research. By “reproducible research” we refer to the discipline of attaching software codes and data to computational results reported in publications. The package contains a collection of (a computational modules, (b data-processing scripts, and (c research papers. Madagascar is distributed on SourceForge under a GPL v2 license https://sourceforge.net/projects/rsf/. By October 2013, more than 70 people from different organizations around the world have contributed to the project, with increasing year-to-year activity. The Madagascar website is http://www.ahay.org/.

  2. Software engineer's pocket book

    CERN Document Server

    Tooley, Michael

    2013-01-01

    Software Engineer's Pocket Book provides a concise discussion on various aspects of software engineering. The book is comprised of six chapters that tackle various areas of concerns in software engineering. Chapter 1 discusses software development, and Chapter 2 covers programming languages. Chapter 3 deals with operating systems. The book also tackles discrete mathematics and numerical computation. Data structures and algorithms are also explained. The text will be of great use to individuals involved in the specification, design, development, implementation, testing, maintenance, and qualit

  3. Identifying Coordination Problems in Software Development: Finding Mismatches between Software and Project Team Structures

    NARCIS (Netherlands)

    Amrit, Chintan Amrit; van Hillegersberg, Jos; Kumar, Kuldeep

    2012-01-01

    Today’s dynamic and iterative development environment brings significant challenges for software project management. In distributed project settings, “management by walking around” is no longer an option and project managers may miss out on key project insights. The TESNA (TEchnical Social Network

  4. Identifying Coordination Problems in Software Development : Finding Mismatches between Software and Project Team Structures

    NARCIS (Netherlands)

    Amrit, Chintan Amrit; van Hillegersberg, Jos; Kumar, Kuldeep

    2012-01-01

    Today’s dynamic and iterative development environment brings significant challenges for software project management. In distributed project settings, “management by walking around” is no longer an option and project managers may miss out on key project insights. The TESNA (TEchnical Social Network

  5. Conceptual Considerations for Reducing the Computational Complexity in Software Defined Radio using Cooperative Wireless Networks

    DEFF Research Database (Denmark)

    Kristensen, Jesper Michael; Fitzek, Frank H. P.; Koch, Peter

    2005-01-01

    This paper motivates the application of Software defined radio as the enabling technology in the implementation of future wireless terminals for 4G. It outlines the advantages and disadvantages of SDR in terms of Flexibility and reconfigurability versus computational complexity. To mitigate...... the expected increase in complexity leading to a decrease in energy efficiency, cooperative wireless networks are introduced. Cooperative wireless networks enables the concept of resource sharing. Resource sharing is interpreted as collaborative signal processing. This interpretation leads to the concept...

  6. Computational Intelligence for User and Data Classification in Hospital Software Development

    OpenAIRE

    Mohammadian, Masoud; Hatzinakos, Dimitrios; Spachos, Petros

    2012-01-01

    Part 3: Second International Workshop on Computational Intelligence in Software Engineering (CISE 2012); International audience; Lives are saved by utilization and application of the latest technologies in hospitals around the world to improve patient treatments and well being. Secure, accurate, near real time data acquisition and analysis of patient data and the ability to update such data will reduce cost and improve the quality of patient’s care. This paper considers a wireless framework b...

  7. Software to provide security for Web Browser Cookies and Passwords using Trusted Computing Technology

    OpenAIRE

    Kannamanani, R.

    2008-01-01

    For further development and enhancement of browser security, there is need to develop something new which provides the better security for the user credential information which will be stored in user PC through the web browser. Although we have many softwares which provide the security functionalities for system, there are many vulnerabilities, so for this there is need to use some functions and specifications mentioned by Trusted Computing group. Therefore, the goal of this semester thesis w...

  8. Hardware and Software Implementations of Prim’s Algorithm for Efficient Minimum Spanning Tree Computation

    OpenAIRE

    Mariano, Artur; Lee, Dongwook; Gerstlauer, Andreas; Chiou, Derek

    2013-01-01

    Part 4: Performance Analysis; International audience; Minimum spanning tree (MST) problems play an important role in many networking applications, such as routing and network planning. In many cases, such as wireless ad-hoc networks, this requires efficient high-performance and low-power implementations that can run at regular intervals in real time on embedded platforms. In this paper, we study custom software and hardware realizations of one common algorithm for MST computations, Prim’s alg...

  9. The concept of computer software designed to identify and analyse logistics costs in agricultural enterprises

    Directory of Open Access Journals (Sweden)

    Karol Wajszczyk

    2009-01-01

    Full Text Available The study comprised research, development and computer programming works concerning the development of a concept for the IT tool to be used in the identification and analysis of logistics costs in agricultural enterprises in terms of the process-based approach. As a result of research and programming work an overall functional and IT concept of software was developed for the identification and analysis of logistics costs for agricultural enterprises.

  10. Development of software for computing forming information using a component based approach

    Directory of Open Access Journals (Sweden)

    Kwang Hee Ko

    2009-12-01

    Full Text Available In shipbuilding industry, the manufacturing technology has advanced at an unprecedented pace for the last decade. As a result, many automatic systems for cutting, welding, etc. have been developed and employed in the manufacturing process and accordingly the productivity has been increased drastically. Despite such improvement in the manufacturing technology, however, development of an automatic system for fabricating a curved hull plate remains at the beginning stage since hardware and software for the automation of the curved hull fabrication process should be developed differently depending on the dimensions of plates, forming methods and manufacturing processes of each shipyard. To deal with this problem, it is necessary to create a “plug-in” framework, which can adopt various kinds of hardware and software to construct a full automatic fabrication system. In this paper, a framework for automatic fabrication of curved hull plates is proposed, which consists of four components and related software. In particular the software module for computing fabrication information is developed by using the ooCBD development methodology, which can interface with other hardware and software with minimum effort. Examples of the proposed framework applied to medium and large shipyards are presented.

  11. Current trends in hardware and software for brain-computer interfaces (BCIs).

    Science.gov (United States)

    Brunner, P; Bianchi, L; Guger, C; Cincotti, F; Schalk, G

    2011-04-01

    A brain-computer interface (BCI) provides a non-muscular communication channel to people with and without disabilities. BCI devices consist of hardware and software. BCI hardware records signals from the brain, either invasively or non-invasively, using a series of device components. BCI software then translates these signals into device output commands and provides feedback. One may categorize different types of BCI applications into the following four categories: basic research, clinical/translational research, consumer products, and emerging applications. These four categories use BCI hardware and software, but have different sets of requirements. For example, while basic research needs to explore a wide range of system configurations, and thus requires a wide range of hardware and software capabilities, applications in the other three categories may be designed for relatively narrow purposes and thus may only need a very limited subset of capabilities. This paper summarizes technical aspects for each of these four categories of BCI applications. The results indicate that BCI technology is in transition from isolated demonstrations to systematic research and commercial development. This process requires several multidisciplinary efforts, including the development of better integrated and more robust BCI hardware and software, the definition of standardized interfaces, and the development of certification, dissemination and reimbursement procedures.

  12. Computer-aided diagnosis software for vulvovaginal candidiasis detection from Pap smear images.

    Science.gov (United States)

    Momenzadeh, Mohammadreza; Vard, Alireza; Talebi, Ardeshir; Mehri Dehnavi, Alireza; Rabbani, Hossein

    2018-01-01

    Vulvovaginal candidiasis (VVC) is a common gynecologic infection and it occurs when there is overgrowth of the yeast called Candida. VVC diagnosis is usually done by observing a Pap smear sample under a microscope and searching for the conidium and mycelium components of Candida. This manual method is time consuming, subjective and tedious. Any diagnosis tools that detect VVC, semi- or full-automatically, can be very helpful to pathologists. This article presents a computer aided diagnosis (CAD) software to improve human diagnosis of VVC from Pap smear samples. The proposed software is designed based on phenotypic and morphology features of the Candida in Pap smear sample images. This software provide a user-friendly interface which consists of a set of image processing tools and analytical results that helps to detect Candida and determine severity of illness. The software was evaluated on 200 Pap smear sample images and obtained specificity of 91.04% and sensitivity of 92.48% to detect VVC. As a result, the use of the proposed software reduces diagnostic time and can be employed as a second objective opinion for pathologists. © 2017 Wiley Periodicals, Inc.

  13. Development of software for computing forming information using a component based approach

    Science.gov (United States)

    Ko, Kwang Hee; Park, Jiing Seo; Kim, Jung; Kim, Young Bum; Shin, Jong Gye

    2009-12-01

    In shipbuilding industry, the manufacturing technology> has advanced at an unprecedented pace for the last decade. As a result, many automatic systems for cutting, welding, etc. have been developed and employed in the manufacturing process and accordingly the productivity has been increased drastically. Despite such improvement in the manufacturing technology', however, development of an automatic system for fabricating a curved hull plate remains at the beginning stage since hardware and software for the automation of the curved hull fabrication process should be developed differently depending on the dimensions of plates, forming methods and manufacturing processes of each shipyard. To deal with this problem, it is necessary> to create a "plug-in ''framework, which can adopt various kinds of hardware and software to construct a full automatic fabrication system. In this paper, a frame-work for automatic fabrication of curved hull plates is proposed, which consists of four components and related software. In particular the software module for computing fabrication information is developed by using the ooCBD development methodology; which can interface with other hardware and software with minimum effort. Examples of the proposed framework applied to medium and large shipyards are presented.

  14. Vertical bone measurements from cone beam computed tomography images using different software packages

    Energy Technology Data Exchange (ETDEWEB)

    Vasconcelos, Taruska Ventorini; Neves, Frederico Sampaio; Moraes, Livia Almeida Bueno; Freitas, Deborah Queiroz, E-mail: tataventorini@hotmail.com [Universidade Estadual de Campinas (UNICAMP), Piracicaba, SP (Brazil). Faculdade de Odontologia

    2015-03-01

    This article aimed at comparing the accuracy of linear measurement tools of different commercial software packages. Eight fully edentulous dry mandibles were selected for this study. Incisor, canine, premolar, first molar and second molar regions were selected. Cone beam computed tomography (CBCT) images were obtained with i-CAT Next Generation. Linear bone measurements were performed by one observer on the cross-sectional images using three different software packages: XoranCat®, OnDemand3D® and KDIS3D®, all able to assess DICOM images. In addition, 25% of the sample was reevaluated for the purpose of reproducibility. The mandibles were sectioned to obtain the gold standard for each region. Intraclass coefficients (ICC) were calculated to examine the agreement between the two periods of evaluation; the one-way analysis of variance performed with the post-hoc Dunnett test was used to compare each of the software-derived measurements with the gold standard. The ICC values were excellent for all software packages. The least difference between the software-derived measurements and the gold standard was obtained with the OnDemand3D and KDIS3D (‑0.11 and ‑0.14 mm, respectively), and the greatest, with the XoranCAT (+0.25 mm). However, there was no statistical significant difference between the measurements obtained with the different software packages and the gold standard (p > 0.05). In conclusion, linear bone measurements were not influenced by the software package used to reconstruct the image from CBCT DICOM data. (author)

  15. TongueWise: Tongue-computer interface software for people with tetraplegia.

    Science.gov (United States)

    Caltenco, Hector A; Andreasen Struijk, Lotte N S; Breidegard, Bjorn

    2010-01-01

    Many computer interfaces and assistive devices for people with motor disabilities limit the input dimensionality from user to system, in many cases leading to single switch interfaces where the user can only press one button. This can, either limit the level of direct access to the functionalities of the operating system, or slow down speed of interaction. In this paper we present TongueWise: a software developed for a tongue computer interface that can be activated with the tip of the tongue and that provides direct input that covers most of the standard keyboard and mouse commands.

  16. Survey of ANL organization plans for word processors, personal computers, workstations, and associated software. Revision 3

    Energy Technology Data Exchange (ETDEWEB)

    Fenske, K.R.

    1991-11-01

    The Computing and Telecommunications Division (CTD) has compiled this Survey of ANL Organization Plans for Word Processors, Personal Computers, Workstations, and Associated Software to provide DOE and Argonne with a record of recent growth in the acquisition and use of personal computers, microcomputers, and word processors at ANL. Laboratory planners, service providers, and people involved in office automation may find the Survey useful. It is for internal use only, and any unauthorized use is prohibited. Readers of the Survey should use it as a reference that documents the plans of each organization for office automation, identifies appropriate planners and other contact people in those organizations, and encourages the sharing of this information among those people making plans for organizations and decisions about office automation. The Survey supplements information in both the ANL Statement of Site Strategy for Computing Workstations and the ANL Site Response for the DOE Information Technology Resources Long-Range Plan.

  17. Survey of ANL organization plans for word processors, personal computers, workstations, and associated software

    Energy Technology Data Exchange (ETDEWEB)

    Fenske, K.R.

    1991-11-01

    The Computing and Telecommunications Division (CTD) has compiled this Survey of ANL Organization Plans for Word Processors, Personal Computers, Workstations, and Associated Software to provide DOE and Argonne with a record of recent growth in the acquisition and use of personal computers, microcomputers, and word processors at ANL. Laboratory planners, service providers, and people involved in office automation may find the Survey useful. It is for internal use only, and any unauthorized use is prohibited. Readers of the Survey should use it as a reference that documents the plans of each organization for office automation, identifies appropriate planners and other contact people in those organizations, and encourages the sharing of this information among those people making plans for organizations and decisions about office automation. The Survey supplements information in both the ANL Statement of Site Strategy for Computing Workstations and the ANL Site Response for the DOE Information Technology Resources Long-Range Plan.

  18. The Impact Of Using Computer Software On Vocabulary Learning Of Iranian EFL University Students

    Directory of Open Access Journals (Sweden)

    Samira Pahlavanpoorfard

    2014-07-01

    Full Text Available Today, using computer is common in all fields. Education is not an exception. In fact, this approach of technology has been using increasingly in language classrooms. We have witnessed there are more and more language teachers are using computers in their classrooms. This research study investigates the impact of using computer   on vocabulary learning of Iranian EFL university students. To this end, a sample of 40 university students in Islamic Azad University, Larestan branch were randomly assigned into the experimental and control groups. Prior the treatment and to catch the initial deferences between the participants, all the students sat for a pre-test that was an Oxford Placement Test. Then the students were received the treatment for 10 weeks. The students in the experimental group were taught by computer software for vocabulary learning while the students in the control group were taught through traditional method for vocabulary learning. After the treatment, all the students sat for a post-test. The statistical analysis through running Independent-Sample T-tests revealed thatthe students in the experimental group who used the computer software for vocabulary learning performed better than the students in the control group were taught through traditional method for vocabulary learning.

  19. A REVIEW ON USER INTERFACE DESIGN PRINCIPLES TO INCREASE SOFTWARE USABILITY FOR USERS WITH LESS COMPUTER LITERACY

    OpenAIRE

    Ali Darejeh; Dalbir Singh

    2013-01-01

    This article presents a review on how software usability could be increased for users with less computer literacy. The literature was reviewed to extract user interface design principles by identifying the similar problems of this group of users. There are different groups of users with less computer literacy. However, based on the literature three groups of them need special attention from software designers. The first group is elderly users, as users with lack of computer background. The se...

  20. A Structured Hardware/Software Architecture for Embedded Sensor Nodes

    OpenAIRE

    Merrett, Geoff V; Weddell, Alex S.; Harris, Nick R; Al-Hashimi, Bashir M; White, Neil M

    2008-01-01

    Owing to the limited requirement for sensor processing in early networked sensor nodes, embedded software was generally built around the communication stack. Modern sensor nodes have evolved to contain significant on-board functionality in addition to communications, including sensor processing, energy management, actuation and locationing. The embedded software for this functionality, however, is often implemented in the application layer of the communications stack, resulting in an unstruct...

  1. TriBITS lifecycle model. Version 1.0, a lean/agile software lifecycle model for research-based computational science and engineering and applied mathematical software.

    Energy Technology Data Exchange (ETDEWEB)

    Willenbring, James M.; Bartlett, Roscoe Ainsworth (Oak Ridge National Laboratory, Oak Ridge, TN); Heroux, Michael Allen

    2012-01-01

    Software lifecycles are becoming an increasingly important issue for computational science and engineering (CSE) software. The process by which a piece of CSE software begins life as a set of research requirements and then matures into a trusted high-quality capability is both commonplace and extremely challenging. Although an implicit lifecycle is obviously being used in any effort, the challenges of this process - respecting the competing needs of research vs. production - cannot be overstated. Here we describe a proposal for a well-defined software lifecycle process based on modern Lean/Agile software engineering principles. What we propose is appropriate for many CSE software projects that are initially heavily focused on research but also are expected to eventually produce usable high-quality capabilities. The model is related to TriBITS, a build, integration and testing system, which serves as a strong foundation for this lifecycle model, and aspects of this lifecycle model are ingrained in the TriBITS system. Here, we advocate three to four phases or maturity levels that address the appropriate handling of many issues associated with the transition from research to production software. The goals of this lifecycle model are to better communicate maturity levels with customers and to help to identify and promote Software Engineering (SE) practices that will help to improve productivity and produce better software. An important collection of software in this domain is Trilinos, which is used as the motivation and the initial target for this lifecycle model. However, many other related and similar CSE (and non-CSE) software projects can also make good use of this lifecycle model, especially those that use the TriBITS system. Indeed this lifecycle process, if followed, will enable large-scale sustainable integration of many complex CSE software efforts across several institutions.

  2. Computer assistance for the structural chemist

    Science.gov (United States)

    Carhart, R. E.; Varkony, T. H.; Smith, D. H.

    1977-01-01

    A description is presented of the approaches used to modify the molecular structure generator program, CONGEN. The CONGEN program for constructing structures under constraints has been discussed by Carhart et al. (1975). The modifications reported are to lead to a more efficient structure generation on the basis of a translation of structural data input to the program. From an algorithmic standpoint, CONGEN is successful if it can, in a reasonable amount of time and without exhausting storage resources, produce a list of candidate structures satisfying the chemist's constraints. However, this list is often quite large, and it remains for the chemist to discriminate among the candidates, eventually reducing the possibilities to just one structure. Ways are studied for providing computer assistance in examining and further constraining lists of structural candidates.

  3. Evaluation of Available Software for Reconstruction of a Structure from its Imagery

    Science.gov (United States)

    2017-04-01

    UNCLASSIFIED Evaluation of Available Software for Reconstruction of a Structure from its Imagery Leonid K Antanovskii Weapons and Combat Systems...Structure from Motion software are evaluated on three datasets of airborne imagery, provided by Defence Research and Development Canada, and on a dataset...of synthetic imagery generated by the VIRSuite software developed in the Defence Science and Technology Group. The user interface for a developed

  4. Corroded Anchor Structure Stability/Reliability (CAS_Stab-R) Software for Hydraulic Structures

    Science.gov (United States)

    2017-12-01

    R) Software for Hydraulic Structures In fo rm at io n Te ch no lo gy L ab or at or y Terry W. Warren, Barry C. White, and Robert M. Ebeling...Engineering Geology 7:287–332. Barton, N., S. Bandis, and K. Bakhtar. 1985. Strength, deformation, and conductivity coupling of rock joints...256. Washington, DC: Headquarters, U.S. Army Corps of Engineers. Murphy, W. L., R. M. Ebeling, and J. M. Andersen. 2002. Assessment of geology as it

  5. Computational Methods for Structural Mechanics and Dynamics

    Science.gov (United States)

    Stroud, W. Jefferson (Editor); Housner, Jerrold M. (Editor); Tanner, John A. (Editor); Hayduk, Robert J. (Editor)

    1989-01-01

    Topics addressed include: transient dynamics; transient finite element method; transient analysis in impact and crash dynamic studies; multibody computer codes; dynamic analysis of space structures; multibody mechanics and manipulators; spatial and coplanar linkage systems; flexible body simulation; multibody dynamics; dynamical systems; and nonlinear characteristics of joints.

  6. FORMATION OF SENIOR PUPILS ALGORITHMIC CULTURE IN THE PROCESS OF SOLVING COMPUTATIONAL PROBLEMS USING SOFTWARE TOOLS: RESULTS OF THE STUDY

    Directory of Open Access Journals (Sweden)

    Liudmyla V. Osipа

    2013-06-01

    Full Text Available Introduced a new practical solution to the urgent problem of the formation of algorithmic culture of senior pupils in the process of solving computational problems using software tools. Identified and theoretically grounded teaching conditions for the formation of algorithmic culture of high school students in the process of solving computational problems with the use of software tools and developed a training program as an elective course "The decision of computational problems using software tools", the introduction of which is necessary for the implementation of the teaching conditions for the formation of algorithmic culture.

  7. 48 CFR 252.227-7018 - Rights in noncommercial technical data and computer software-Small Business Innovation Research...

    Science.gov (United States)

    2010-10-01

    ... technical data and computer software-Small Business Innovation Research (SBIR) Program. 252.227-7018 Section... Noncommercial Technical Data and Computer Software—Small Business Innovation Research (SBIR) Program (JUN 1995... the Rights in Noncommercial Technical Data and Computer Software—Small Business Innovative Research...

  8. RNA secondary structure prediction using soft computing.

    Science.gov (United States)

    Ray, Shubhra Sankar; Pal, Sankar K

    2013-01-01

    Prediction of RNA structure is invaluable in creating new drugs and understanding genetic diseases. Several deterministic algorithms and soft computing-based techniques have been developed for more than a decade to determine the structure from a known RNA sequence. Soft computing gained importance with the need to get approximate solutions for RNA sequences by considering the issues related with kinetic effects, cotranscriptional folding, and estimation of certain energy parameters. A brief description of some of the soft computing-based techniques, developed for RNA secondary structure prediction, is presented along with their relevance. The basic concepts of RNA and its different structural elements like helix, bulge, hairpin loop, internal loop, and multiloop are described. These are followed by different methodologies, employing genetic algorithms, artificial neural networks, and fuzzy logic. The role of various metaheuristics, like simulated annealing, particle swarm optimization, ant colony optimization, and tabu search is also discussed. A relative comparison among different techniques, in predicting 12 known RNA secondary structures, is presented, as an example. Future challenging issues are then mentioned.

  9. Software For Genetic Algorithms

    Science.gov (United States)

    Wang, Lui; Bayer, Steve E.

    1992-01-01

    SPLICER computer program is genetic-algorithm software tool used to solve search and optimization problems. Provides underlying framework and structure for building genetic-algorithm application program. Written in Think C.

  10. i3Drefine software for protein 3D structure refinement and its assessment in CASP10.

    Science.gov (United States)

    Bhattacharya, Debswapna; Cheng, Jianlin

    2013-01-01

    Protein structure refinement refers to the process of improving the qualities of protein structures during structure modeling processes to bring them closer to their native states. Structure refinement has been drawing increasing attention in the community-wide Critical Assessment of techniques for Protein Structure prediction (CASP) experiments since its addition in 8(th) CASP experiment. During the 9(th) and recently concluded 10(th) CASP experiments, a consistent growth in number of refinement targets and participating groups has been witnessed. Yet, protein structure refinement still remains a largely unsolved problem with majority of participating groups in CASP refinement category failed to consistently improve the quality of structures issued for refinement. In order to alleviate this need, we developed a completely automated and computationally efficient protein 3D structure refinement method, i3Drefine, based on an iterative and highly convergent energy minimization algorithm with a powerful all-atom composite physics and knowledge-based force fields and hydrogen bonding (HB) network optimization technique. In the recent community-wide blind experiment, CASP10, i3Drefine (as 'MULTICOM-CONSTRUCT') was ranked as the best method in the server section as per the official assessment of CASP10 experiment. Here we provide the community with free access to i3Drefine software and systematically analyse the performance of i3Drefine in strict blind mode on the refinement targets issued in CASP10 refinement category and compare with other state-of-the-art refinement methods participating in CASP10. Our analysis demonstrates that i3Drefine is only fully-automated server participating in CASP10 exhibiting consistent improvement over the initial structures in both global and local structural quality metrics. Executable version of i3Drefine is freely available at http://protein.rnet.missouri.edu/i3drefine/.

  11. i3Drefine software for protein 3D structure refinement and its assessment in CASP10.

    Directory of Open Access Journals (Sweden)

    Debswapna Bhattacharya

    Full Text Available Protein structure refinement refers to the process of improving the qualities of protein structures during structure modeling processes to bring them closer to their native states. Structure refinement has been drawing increasing attention in the community-wide Critical Assessment of techniques for Protein Structure prediction (CASP experiments since its addition in 8(th CASP experiment. During the 9(th and recently concluded 10(th CASP experiments, a consistent growth in number of refinement targets and participating groups has been witnessed. Yet, protein structure refinement still remains a largely unsolved problem with majority of participating groups in CASP refinement category failed to consistently improve the quality of structures issued for refinement. In order to alleviate this need, we developed a completely automated and computationally efficient protein 3D structure refinement method, i3Drefine, based on an iterative and highly convergent energy minimization algorithm with a powerful all-atom composite physics and knowledge-based force fields and hydrogen bonding (HB network optimization technique. In the recent community-wide blind experiment, CASP10, i3Drefine (as 'MULTICOM-CONSTRUCT' was ranked as the best method in the server section as per the official assessment of CASP10 experiment. Here we provide the community with free access to i3Drefine software and systematically analyse the performance of i3Drefine in strict blind mode on the refinement targets issued in CASP10 refinement category and compare with other state-of-the-art refinement methods participating in CASP10. Our analysis demonstrates that i3Drefine is only fully-automated server participating in CASP10 exhibiting consistent improvement over the initial structures in both global and local structural quality metrics. Executable version of i3Drefine is freely available at http://protein.rnet.missouri.edu/i3drefine/.

  12. Application of a newly developed software program for image quality assessment in cone-beam computed tomography

    Energy Technology Data Exchange (ETDEWEB)

    De Oliveira, Marcus Vinicius Linhares; Campos, Paulo Sergio Flores [Federal Institute of Bahia, Salvador (Brazil); Paulo, Graciano; Santos, Antonio Carvalho; Santos, Joana [Coimbra Health School, Polytechnic Institute of Coimbra, Coimbra (Portugal)

    2017-06-15

    The purpose of this study was to apply a newly developed free software program, at low cost and with minimal time, to evaluate the quality of dental and maxillofacial cone-beam computed tomography (CBCT) images. A polymethyl methacrylate (PMMA) phantom, CQP-IFBA, was scanned in 3 CBCT units with 7 protocols. A macro program was developed, using the free software ImageJ, to automatically evaluate the image quality parameters. The image quality evaluation was based on 8 parameters: uniformity, the signal-to-noise ratio (SNR), noise, the contrast-to-noise ratio (CNR), spatial resolution, the artifact index, geometric accuracy, and low-contrast resolution. The image uniformity and noise depended on the protocol that was applied. Regarding the CNR, high-density structures were more sensitive to the effect of scanning parameters. There were no significant differences between SNR and CNR in centered and peripheral objects. The geometric accuracy assessment showed that all the distance measurements were lower than the real values. Low-contrast resolution was influenced by the scanning parameters, and the 1-mm rod present in the phantom was not depicted in any of the 3 CBCT units. Smaller voxel sizes presented higher spatial resolution. There were no significant differences among the protocols regarding artifact presence. This software package provided a fast, low-cost, and feasible method for the evaluation of image quality parameters in CBCT.

  13. Computer-assisted teaching of skin flap surgery: validation of a mobile platform software for medical students.

    Science.gov (United States)

    de Sena, David P; Fabricio, Daniela D; Lopes, Maria Helena I; da Silva, Vinicius D

    2013-01-01

    The purpose of this study was to develop and validate a multimedia software application for mobile platforms to assist in the teaching and learning process of design and construction of a skin flap. Traditional training in surgery is based on learning by doing. Initially, the use of cadavers and animal models appeared to be a valid alternative for training. However, many conflicts with these training models prompted progression to synthetic and virtual reality models. Fifty volunteer fifth- and sixth-year medical students completed a pretest and were randomly allocated into two groups of 25 students each. The control group was exposed for 5 minutes to a standard text-based print article, while the test group used multimedia software describing how to fashion a rhomboid flap. Each group then performed a cutaneous flap on a training bench model while being evaluated by three blinded BSPS (Brazilian Society of Plastic Surgery) board-certified surgeons using the OSATS (Objective Structured Assessment of Technical Skill) protocol and answered a post-test. The text-based group was then tested again using the software. The computer-assisted learning (CAL) group had superior performance as confirmed by checklist scores (pfashioning rhomboid flaps as compared to those taught with standard print material. These findings indicate that students preferred to learn using the multimedia method.

  14. Application of a newly developed software program for image quality assessment in cone-beam computed tomography.

    Science.gov (United States)

    de Oliveira, Marcus Vinicius Linhares; Santos, António Carvalho; Paulo, Graciano; Campos, Paulo Sergio Flores; Santos, Joana

    2017-06-01

    The purpose of this study was to apply a newly developed free software program, at low cost and with minimal time, to evaluate the quality of dental and maxillofacial cone-beam computed tomography (CBCT) images. A polymethyl methacrylate (PMMA) phantom, CQP-IFBA, was scanned in 3 CBCT units with 7 protocols. A macro program was developed, using the free software ImageJ, to automatically evaluate the image quality parameters. The image quality evaluation was based on 8 parameters: uniformity, the signal-to-noise ratio (SNR), noise, the contrast-to-noise ratio (CNR), spatial resolution, the artifact index, geometric accuracy, and low-contrast resolution. The image uniformity and noise depended on the protocol that was applied. Regarding the CNR, high-density structures were more sensitive to the effect of scanning parameters. There were no significant differences between SNR and CNR in centered and peripheral objects. The geometric accuracy assessment showed that all the distance measurements were lower than the real values. Low-contrast resolution was influenced by the scanning parameters, and the 1-mm rod present in the phantom was not depicted in any of the 3 CBCT units. Smaller voxel sizes presented higher spatial resolution. There were no significant differences among the protocols regarding artifact presence. This software package provided a fast, low-cost, and feasible method for the evaluation of image quality parameters in CBCT.

  15. Infrastructure for Multiphysics Software Integration in High Performance Computing-Aided Science and Engineering

    Energy Technology Data Exchange (ETDEWEB)

    Campbell, Michael T. [Illinois Rocstar LLC, Champaign, IL (United States); Safdari, Masoud [Illinois Rocstar LLC, Champaign, IL (United States); Kress, Jessica E. [Illinois Rocstar LLC, Champaign, IL (United States); Anderson, Michael J. [Illinois Rocstar LLC, Champaign, IL (United States); Horvath, Samantha [Illinois Rocstar LLC, Champaign, IL (United States); Brandyberry, Mark D. [Illinois Rocstar LLC, Champaign, IL (United States); Kim, Woohyun [Illinois Rocstar LLC, Champaign, IL (United States); Sarwal, Neil [Illinois Rocstar LLC, Champaign, IL (United States); Weisberg, Brian [Illinois Rocstar LLC, Champaign, IL (United States)

    2016-10-15

    . There are over 100 unit tests provided that run through the Illinois Rocstar Application Development (IRAD) lightweight testing infrastructure that is also supplied along with IMPACT. The package as a whole provides an excellent base for developing high-quality multiphysics applications using modern software development practices. To facilitate understanding how to utilize IMPACT effectively, two multiphysics systems have been developed and are available open-source through gitHUB. The simpler of the two systems, named ElmerFoamFSI in the repository, is a multiphysics, fluid-structure-interaction (FSI) coupling of the solid mechanics package Elmer with a fluid dynamics module from OpenFOAM. This coupling illustrates how to combine software packages that are unrelated by either author or architecture and combine them into a robust, parallel multiphysics system. A more complex multiphysics tool is the Illinois Rocstar Rocstar Multiphysics code that was rebuilt during the project around IMPACT. Rocstar Multiphysics was already an HPC multiphysics tool, but now that it has been rearchitected around IMPACT, it can be readily expanded to capture new and different physics in the future. In fact, during this project, the Elmer and OpenFOAM tools were also coupled into Rocstar Multiphysics and demonstrated. The full Rocstar Multiphysics codebase is also available on gitHUB, and licensed for any organization to use as they wish. Finally, the new IMPACT product is already being used in several multiphysics code coupling projects for the Air Force, NASA and the Missile Defense Agency, and initial work on expansion of the IMPACT-enabled Rocstar Multiphysics has begun in support of a commercial company. These initiatives promise to expand the interest and reach of IMPACT and Rocstar Multiphysics, ultimately leading to the envisioned standardization and consortium of users that was one of the goals of this project.

  16. 2nd International Workshop on Eigenvalue Problems : Algorithms, Software and Applications in Petascale Computing

    CERN Document Server

    Zhang, Shao-Liang; Imamura, Toshiyuki; Yamamoto, Yusaku; Kuramashi, Yoshinobu; Hoshi, Takeo

    2017-01-01

    This book provides state-of-the-art and interdisciplinary topics on solving matrix eigenvalue problems, particularly by using recent petascale and upcoming post-petascale supercomputers. It gathers selected topics presented at the International Workshops on Eigenvalue Problems: Algorithms; Software and Applications, in Petascale Computing (EPASA2014 and EPASA2015), which brought together leading researchers working on the numerical solution of matrix eigenvalue problems to discuss and exchange ideas – and in so doing helped to create a community for researchers in eigenvalue problems. The topics presented in the book, including novel numerical algorithms, high-performance implementation techniques, software developments and sample applications, will contribute to various fields that involve solving large-scale eigenvalue problems.

  17. The impact of organizational structure on flight software cost risk

    Science.gov (United States)

    Hihn, Jairus; Lum, Karen; Monson, Erik

    2004-01-01

    This paper summarizes the final results of the follow-up study updating the estimated software effort growth for those projects that were still under development and including an evaluation of the roles versus observed cost risk for the missions included in the original study which expands the data set to thirteen missions.

  18. Exploring Coordination Structures in Open Source Software Development

    NARCIS (Netherlands)

    van Hillegersberg, Jos; Harmsen, Frank; Hegeman, J.H.; Amrit, Chintan Amrit; Geisberger, Eva; Keil, Patrick; Kuhrmann, Marco

    2007-01-01

    Coordination is difficult to achieve in a large globally distributed project setting. The problem is multiplied in open source software development projects, where most of the traditional means of coordination such as plans, system-level designs, schedules and defined process are not used. In order

  19. SMOG 2: A Versatile Software Package for Generating Structure-Based Models.

    Directory of Open Access Journals (Sweden)

    Jeffrey K Noel

    2016-03-01

    Full Text Available Molecular dynamics simulations with coarse-grained or simplified Hamiltonians have proven to be an effective means of capturing the functionally important long-time and large-length scale motions of proteins and RNAs. Originally developed in the context of protein folding, structure-based models (SBMs have since been extended to probe a diverse range of biomolecular processes, spanning from protein and RNA folding to functional transitions in molecular machines. The hallmark feature of a structure-based model is that part, or all, of the potential energy function is defined by a known structure. Within this general class of models, there exist many possible variations in resolution and energetic composition. SMOG 2 is a downloadable software package that reads user-designated structural information and user-defined energy definitions, in order to produce the files necessary to use SBMs with high performance molecular dynamics packages: GROMACS and NAMD. SMOG 2 is bundled with XML-formatted template files that define commonly used SBMs, and it can process template files that are altered according to the needs of each user. This computational infrastructure also allows for experimental or bioinformatics-derived restraints or novel structural features to be included, e.g. novel ligands, prosthetic groups and post-translational/transcriptional modifications. The code and user guide can be downloaded at http://smog-server.org/smog2.

  20. The VENUS/NWChem Software Package. Tight Coupling Between Chemical Dynamics Simulations and Electronic Structure Theory

    Energy Technology Data Exchange (ETDEWEB)

    Lourderaj, Upakarasamy; Sun, Rui; De Jong, Wibe A.; Windus, Theresa L.; Hase, William L.

    2014-03-01

    The interface for VENUS and NWChem, and the resulting software package for direct dynamics simulations are described. The coupling of the two codes is considered to be a tight coupling. The two codes are compiled and linked together and act as one executable with data being passed between the two codes through routine calls. The advantages of this type of coupling are discussed. The interface has been designed to have as little interference as possible with the core codes of both VENUS and NWChem. VENUS is the code that propagates the direct dynamics trajectories and, therefore, is the program that drives the overall execution of VENUS/NWChem. VENUS has remained an essentially sequential code, which uses the highly parallel structure of NWChem. Subroutines of the interface which accomplish the data transmission and communication between the two computer programs are described. Recent examples of the use of VENUS/NWChem for direct dynamics simulations are summarized.

  1. CORBA-Based Distributed Software Framework for the NIF Integrated Computer Control System

    Energy Technology Data Exchange (ETDEWEB)

    Stout, E A; Carey, R W; Estes, C M; Fisher, J M; Lagin, L J; Mathisen, D G; Reynolds, C A; Sanchez, R J

    2007-11-20

    The National Ignition Facility (NIF), currently under construction at the Lawrence Livermore National Laboratory, is a stadium-sized facility containing a 192-beam, 1.8 Megajoule, 500-Terawatt, ultra-violet laser system together with a 10-meter diameter target chamber with room for nearly 100 experimental diagnostics. The NIF is operated by the Integrated Computer Control System (ICCS) which is a scalable, framework-based control system distributed over 800 computers throughout the NIF. The framework provides templates and services at multiple levels of abstraction for the construction of software applications that communicate via CORBA (Common Object Request Broker Architecture). Object-oriented software design patterns are implemented as templates and extended by application software. Developers extend the framework base classes to model the numerous physical control points and implement specializations of common application behaviors. An estimated 140 thousand software objects, each individually addressable through CORBA, will be active at full scale. Many of these objects have persistent configuration information stored in a database. The configuration data is used to initialize the objects at system start-up. Centralized server programs that implement events, alerts, reservations, data archival, name service, data access, and process management provide common system wide services. At the highest level, a model-driven, distributed shot automation system provides a flexible and scalable framework for automatic sequencing of work-flow for control and monitoring of NIF shots. The shot model, in conjunction with data defining the parameters and goals of an experiment, describes the steps to be performed by each subsystem in order to prepare for and fire a NIF shot. Status and usage of this distributed framework are described.

  2. CORBA-based distributed software framework for the NIF integrated computer control system

    Energy Technology Data Exchange (ETDEWEB)

    Stout, E.A. [Lawrence Livermore National Laboratory, P.O. Box 808, Livermore, CA 94550 (United States)], E-mail: stout6@llnl.gov; Carey, R.W.; Estes, C.M.; Fisher, J.M.; Lagin, L.J.; Mathisen, D.G.; Reynolds, C.A.; Sanchez, R.J. [Lawrence Livermore National Laboratory, P.O. Box 808, Livermore, CA 94550 (United States)

    2008-04-15

    The National Ignition Facility (NIF), currently under construction at the Lawrence Livermore National Laboratory, is a stadium-sized facility containing a 192-beam, 1.8 MJ, 500 TW, ultra-violet laser system together with a 10-meter diameter target chamber with room for nearly 100 experimental diagnostics. The NIF is operated by the Integrated Computer Control System (ICCS) which is a scalable, framework-based control system distributed over 800 computers throughout the NIF. The framework provides templates and services at multiple levels of abstraction for the construction of software applications that communicate via CORBA (Common Object Request Broker Architecture). Object-oriented software design patterns are implemented as templates and extended by application software. Developers extend the framework base classes to model the numerous physical control points and implement specializations of common application behaviors. An estimated 140,000 software objects, each individually addressable through CORBA, will be active at full scale. Many of these objects have persistent configuration information stored in a database. The configuration data is used to initialize the objects at system start-up. Centralized server programs that implement events, alerts, reservations, data archival, name service, data access, and process management provide common system wide services. At the highest level, a model-driven, distributed shot automation system provides a flexible and scalable framework for automatic sequencing of workflow for control and monitoring of NIF shots. The shot model, in conjunction with data defining the parameters and goals of an experiment, describes the steps to be performed by each subsystem in order to prepare for and fire a NIF shot. Status and usage of this distributed framework are described.

  3. Scilab software as an alternative low-cost computing in solving the linear equations problem

    Science.gov (United States)

    Agus, Fahrul; Haviluddin

    2017-02-01

    Numerical computation packages are widely used both in teaching and research. These packages consist of license (proprietary) and open source software (non-proprietary). One of the reasons to use the package is a complexity of mathematics function (i.e., linear problems). Also, number of variables in a linear or non-linear function has been increased. The aim of this paper was to reflect on key aspects related to the method, didactics and creative praxis in the teaching of linear equations in higher education. If implemented, it could be contribute to a better learning in mathematics area (i.e., solving simultaneous linear equations) that essential for future engineers. The focus of this study was to introduce an additional numerical computation package of Scilab as an alternative low-cost computing programming. In this paper, Scilab software was proposed some activities that related to the mathematical models. In this experiment, four numerical methods such as Gaussian Elimination, Gauss-Jordan, Inverse Matrix, and Lower-Upper Decomposition (LU) have been implemented. The results of this study showed that a routine or procedure in numerical methods have been created and explored by using Scilab procedures. Then, the routine of numerical method that could be as a teaching material course has exploited.

  4. Software usage in unsupervised digital doorway computing environments in disadvantaged South African communities: Focusing on youthful users

    CSIR Research Space (South Africa)

    Gush, K

    2011-01-01

    Full Text Available Digital Doorways provide computing infrastructure in low-income communities in South Africa. The unsupervised DD terminals offer various software applications, from entertainment through educational resources to research material, encouraging...

  5. Parallel structures in human and computer memory

    Science.gov (United States)

    Kanerva, P.

    1986-01-01

    If one thinks of our experiences as being recorded continuously on film, then human memory can be compared to a film library that is indexed by the contents of the film strips stored in it. Moreover, approximate retrieval cues suffice to retrieve information stored in this library. One recognizes a familiar person in a fuzzy photograph or a familiar tune played on a strange instrument. A computer memory that would allow a computer to recognize patterns and to recall sequences the way humans do is constructed. Such a memory is remarkably similiar in structure to a conventional computer memory and also to the neural circuits in the cortex of the cerebellum of the human brain. It is concluded that the frame problem of artificial intelligence could be solved by the use of such a memory if one were able to encode information about the world properly.

  6. Assess/Mitigate Risk through the Use of Computer-Aided Software Engineering (CASE) Tools

    Science.gov (United States)

    Aguilar, Michael L.

    2013-01-01

    The NASA Engineering and Safety Center (NESC) was requested to perform an independent assessment of the mitigation of the Constellation Program (CxP) Risk 4421 through the use of computer-aided software engineering (CASE) tools. With the cancellation of the CxP, the assessment goals were modified to capture lessons learned and best practices in the use of CASE tools. The assessment goal was to prepare the next program for the use of these CASE tools. The outcome of the assessment is contained in this document.

  7. NEXUS/NASCAD- NASA ENGINEERING EXTENDIBLE UNIFIED SOFTWARE SYSTEM WITH NASA COMPUTER AIDED DESIGN

    Science.gov (United States)

    Purves, L. R.

    1994-01-01

    NEXUS, the NASA Engineering Extendible Unified Software system, is a research set of computer programs designed to support the full sequence of activities encountered in NASA engineering projects. This sequence spans preliminary design, design analysis, detailed design, manufacturing, assembly, and testing. NEXUS primarily addresses the process of prototype engineering, the task of getting a single or small number of copies of a product to work. Prototype engineering is a critical element of large scale industrial production. The time and cost needed to introduce a new product are heavily dependent on two factors: 1) how efficiently required product prototypes can be developed, and 2) how efficiently required production facilities, also a prototype engineering development, can be completed. NEXUS extendibility and unification are achieved by organizing the system as an arbitrarily large set of computer programs accessed in a common manner through a standard user interface. The NEXUS interface is a multipurpose interactive graphics interface called NASCAD (NASA Computer Aided Design). NASCAD can be used to build and display two and three-dimensional geometries, to annotate models with dimension lines, text strings, etc., and to store and retrieve design related information such as names, masses, and power requirements of components used in the design. From the user's standpoint, NASCAD allows the construction, viewing, modification, and other processing of data structures that represent the design. Four basic types of data structures are supported by NASCAD: 1) three-dimensional geometric models of the object being designed, 2) alphanumeric arrays to hold data ranging from numeric scalars to multidimensional arrays of numbers or characters, 3) tabular data sets that provide a relational data base capability, and 4) procedure definitions to combine groups of system commands or other user procedures to create more powerful functions. NASCAD has extensive abilities to

  8. Structured Parallel Programming Patterns for Efficient Computation

    CERN Document Server

    McCool, Michael; Robison, Arch

    2012-01-01

    Programming is now parallel programming. Much as structured programming revolutionized traditional serial programming decades ago, a new kind of structured programming, based on patterns, is relevant to parallel programming today. Parallel computing experts and industry insiders Michael McCool, Arch Robison, and James Reinders describe how to design and implement maintainable and efficient parallel algorithms using a pattern-based approach. They present both theory and practice, and give detailed concrete examples using multiple programming models. Examples are primarily given using two of th

  9. HPC Institutional Computing Project: W15_lesreactiveflow KIVA-hpFE Development: A Robust and Accurate Engine Modeling Software

    Energy Technology Data Exchange (ETDEWEB)

    Carrington, David Bradley [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Waters, Jiajia [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-01-05

    KIVA-hpFE is a high performance computer software for solving the physics of multi-species and multiphase turbulent reactive flow in complex geometries having immersed moving parts. The code is written in Fortran 90/95 and can be used on any computer platform with any popular complier. The code is in two versions, a serial version and a parallel version utilizing MPICH2 type Message Passing Interface (MPI or Intel MPI) for solving distributed domains. The parallel version is at least 30x faster than the serial version and much faster than our previous generation of parallel engine modeling software, by many factors. The 5th generation algorithm construction is a Galerkin type Finite Element Method (FEM) solving conservative momentum, species, and energy transport equations along with two-equation turbulent model k-ω Reynolds Averaged Navier-Stokes (RANS) model and a Vreman type dynamic Large Eddy Simulation (LES) method. The LES method is capable modeling transitional flow from laminar to fully turbulent; therefore, this LES method does not require special hybrid or blending to walls. The FEM projection method also uses a Petrov-Galerkin (P-G) stabilization along with pressure stabilization. We employ hierarchical basis sets, constructed on the fly with enrichment in areas associated with relatively larger error as determined by error estimation methods. In addition, when not using the hp-adaptive module, the code employs Lagrangian basis or shape functions. The shape functions are constructed for hexahedral, prismatic and tetrahedral elements. The software is designed to solve many types of reactive flow problems, from burners to internal combustion engines and turbines. In addition, the formulation allows for direct integration of solid bodies (conjugate heat transfer), as in heat transfer through housings, parts, cylinders. It can also easily be extended to stress modeling of solids, used in fluid structure interactions problems, solidification, porous media

  10. A Visual Basic Software for Computing Fisher\\'s Exact Probability

    Directory of Open Access Journals (Sweden)

    Haseeb Ahmad Khan

    2003-12-01

    Full Text Available Fisher's exact test (FET is an important statistical method for testing association between two groups. However, the computations involved in FET are extremely tedious and time consuming due to multi-step factorial calculations after the construction of numerous 2x2 tables depending on the smallest cell value. A Visual-Basic computer program, CalcFisher, has been developed to handle the complexities of FET resorting the techniques of looping subroutines and logarithmic conversions. The software automatically calculates the P-value after entering the respective cell values and has been validated for proper functioning with a wide range of frequencies (tens to several thousands. The important features of the program include, easy data entry, tail-selection, comprehensive report format and the facility of printing and saving of results.

  11. Prepare for X-Win32 - the new X11 server software for Windows computers

    CERN Multimedia

    IT Department

    2011-01-01

    Starnet X-Win32 will replace Exceed as the X11 Server software on Windows computers by February 2012. X11 Server software allows a Windows user to have a graphical user interface on a remote Linux server. This change, initially motivated by a significant change of license conditions for Exceed, brings an easier integration of Windows and Linux logon mechanisms. At the same time, X-Win32 addresses the common use cases while providing a more intuitive configuration interface. CERN Predefined Connections will be available as before. They offer an easy way of starting applications on LXPLUS using PuTTY or starting the KDE, GNOME or ICE window managers. Since X-Win32 is better integrated with SSH and CERN Kerberos compared to Exceed, it is much simpler to set up secure access to Linux services. The decision to choose X-Win32 as the new X11 software resulted from an evaluation that involved various user communities and support teams. More information, including the documented use cases, is available at https://...

  12. Development and application of computer software for cell culture laboratory management.

    Science.gov (United States)

    Selznick, S H; Thatcher, M L; Brown, K S; Haussler, C A

    2001-01-01

    Prototype computer software for a Cell Culture Laboratory Management System (CCLMS) has been developed to relieve cell culture specialists of the burden of manual recordkeeping. Conventional data archives in cell culture laboratories are prone to error and expensive to maintain. The reliance upon cell culture to provide models for biochemical and molecular biological research serves to magnify errors at great expense. The CCLMS prototype encapsulates a modular software application that manages the many aspects of cell culture laboratory recordkeeping. A transaction-based database stores detailed information on subcultures, freezes and thaws, prints waterproof labels for culture vessels, and provides for immediate historical trace-back of any cultured cell line. Linked database files store information specific to an individual culture flask while removing redundancy between similar groups of flasks. A frozen cell log maintains locations of all vials within any type of cryogenic storage unit, locates spaces for newly frozen cell lines, and generates alphabetical or numerical reports. Finally, modules for maintaining cell counts, user records, and culture vessel specifications to support a comprehensive automation process are incorporated within this software. The developed CCLMS prototype has been demonstrated to be an adaptable, reliable tool for improving training, efficiency, and historical rigor for two independent cell culture facilities.

  13. High Resolution Topography of Polar Regions from Commercial Satellite Imagery, Petascale Computing and Open Source Software

    Science.gov (United States)

    Morin, Paul; Porter, Claire; Cloutier, Michael; Howat, Ian; Noh, Myoung-Jong; Willis, Michael; Kramer, WIlliam; Bauer, Greg; Bates, Brian; Williamson, Cathleen

    2017-04-01

    Surface topography is among the most fundamental data sets for geosciences, essential for disciplines ranging from glaciology to geodynamics. Two new projects are using sub-meter, commercial imagery licensed by the National Geospatial-Intelligence Agency and open source photogrammetry software to produce a time-tagged 2m posting elevation model of the Arctic and an 8m posting reference elevation model for the Antarctic. When complete, this publically available data will be at higher resolution than any elevation models that cover the entirety of the Western United States. These two polar projects are made possible due to three equally important factors: 1) open-source photogrammetry software, 2) petascale computing, and 3) sub-meter imagery licensed to the United States Government. Our talk will detail the technical challenges of using automated photogrammetry software; the rapid workflow evolution to allow DEM production; the task of deploying the workflow on one of the world's largest supercomputers; the trials of moving massive amounts of data, and the management strategies the team needed to solve in order to meet deadlines. Finally, we will discuss the implications of this type of collaboration for future multi-team use of leadership-class systems such as Blue Waters, and for further elevation mapping.

  14. Combined Visualization of Structural and Metric Information for Software Evolution Analysis

    NARCIS (Netherlands)

    Gonzalez, Antonio; Theron, Roberto; Telea, Alexandru; Garcia, Francisco J.

    2009-01-01

    This paper discusses a proposal for the visualization of software evolution, with a focus on combining insight on changes that affect software metrics at project and class level, the project structure, the class hierarchy and the viewing of source code correlated to indirect class coupling. The

  15. An Object-Oriented Network-Centric Software Architecture for Physical Computing

    Science.gov (United States)

    Palmer, Richard

    1997-08-01

    Recent developments in object-oriented computer languages and infrastructure such as the Internet, Web browsers, and the like provide an opportunity to define a more productive computational environment for scientific programming that is based more closely on the underlying mathematics describing physics than traditional programming languages such as FORTRAN or C++. In this talk I describe an object-oriented software architecture for representing physical problems that includes classes for such common mathematical objects as geometry, boundary conditions, partial differential and integral equations, discretization and numerical solution methods, etc. In practice, a scientific program written using this architecture looks remarkably like the mathematics used to understand the problem, is typically an order of magnitude smaller than traditional FORTRAN or C++ codes, and hence easier to understand, debug, describe, etc. All objects in this architecture are ``network-enabled,'' which means that components of a software solution to a physical problem can be transparently loaded from anywhere on the Internet or other global network. The architecture is expressed as an ``API,'' or application programmers interface specification, with reference embeddings in Java, Python, and C++. A C++ class library for an early version of this API has been implemented for machines ranging from PC's to the IBM SP2, meaning that phidentical codes run on all architectures.

  16. The changing industry structure of software development for consumer electronics and its consequences for software architectures

    NARCIS (Netherlands)

    Hartmann, Herman; Trew, Tim; Bosch, Jan

    During the last decade the structure of the consumer electronics industry has been changing profoundly. Current consumer electronics products are built using components from a large variety of specialized firms, whereas previously each product was developed by a single, vertically integrated

  17. Legato: Personal Computer Software for Analyzing Pressure-Sensitive Paint Data

    Science.gov (United States)

    Schairer, Edward T.

    2001-01-01

    'Legato' is personal computer software for analyzing radiometric pressure-sensitive paint (PSP) data. The software is written in the C programming language and executes under Windows 95/98/NT operating systems. It includes all operations normally required to convert pressure-paint image intensities to normalized pressure distributions mapped to physical coordinates of the test article. The program can analyze data from both single- and bi-luminophore paints and provides for both in situ and a priori paint calibration. In addition, there are functions for determining paint calibration coefficients from calibration-chamber data. The software is designed as a self-contained, interactive research tool that requires as input only the bare minimum of information needed to accomplish each function, e.g., images, model geometry, and paint calibration coefficients (for a priori calibration) or pressure-tap data (for in situ calibration). The program includes functions that can be used to generate needed model geometry files for simple model geometries (e.g., airfoils, trapezoidal wings, rotor blades) based on the model planform and airfoil section. All data files except images are in ASCII format and thus are easily created, read, and edited. The program does not use database files. This simplifies setup but makes the program inappropriate for analyzing massive amounts of data from production wind tunnels. Program output consists of Cartesian plots, false-colored real and virtual images, pressure distributions mapped to the surface of the model, assorted ASCII data files, and a text file of tabulated results. Graphical output is displayed on the computer screen and can be saved as publication-quality (PostScript) files.

  18. Detection of Steganography-Producing Software Artifacts on Crime-Related Seized Computers

    Directory of Open Access Journals (Sweden)

    Asawaree Kulkarni

    2009-06-01

    Full Text Available Steganography is the art and science of hiding information within information so that an observer does not know that communication is taking place. Bad actors passing information using steganography are of concern to the national security establishment and law enforcement. An attempt was made to determine if steganography was being used by criminals to communicate information. Web crawling technology was used and images were downloaded from Web sites that were considered as likely candidates for containing information hidden using steganographic techniques. A detection tool was used to analyze these images. The research failed to demonstrate that steganography was prevalent on the public Internet. The probable reasons included the growth and availability of large number of steganography-producing tools and the limited capacity of the detection tools to cope with them. Thus, a redirection was introduced in the methodology and the detection focus was shifted from the analysis of the ‘product’ of the steganography-producing software; viz. the images, to the 'artifacts’ left by the steganography-producing software while it is being used to generate steganographic images. This approach was based on the concept of ‘Stego-Usage Timeline’. As a proof of concept, a sample set of criminal computers was scanned for the remnants of steganography-producing software. The results demonstrated that the problem of ‘the detection of the usage of steganography’ could be addressed by the approach adopted after the research redirection and that certain steganographic software was popular among the criminals. Thus, the contribution of the research was in demonstrating that the limitations of the tools based on the signature detection of steganographically altered images can be overcome by focusing the detection effort on detecting the artifacts of the steganography-producing tools.

  19. Computational Finite Element Software Assisted Development of a 3D Inductively Coupled Power Transfer System

    Directory of Open Access Journals (Sweden)

    P. Raval

    2014-02-01

    Full Text Available To date inductively coupled power transfer (ICPT systems have already found many practical applications including battery charging pads. In fact, current charging platforms tend to largely support only one- or two-dimensional planar movement in load. This paper proposes a new concept of extending the aspect ratios of the operating power transfer volume of ICPT systems to support arbitrary three dimensional load movements with respect to the primary coils. This is done by use of modern finite element method analysis software to propose the primary and secondary magnetic structures of such an ICPT system. Firstly, two primary magnetic structures are proposed based on contrasting modes of operation and different field directions. This includes a single-phase and multi-phase current model. Next, a secondary magnetic structure is customized to be compatible with both primary structures. The resulting system is shown to produce a 3D power transfer volume for battery cell charging applications.

  20. JADAMILU: a software code for computing selected eigenvalues of large sparse symmetric matrices

    Science.gov (United States)

    Bollhöfer, Matthias; Notay, Yvan

    2007-12-01

    A new software code for computing selected eigenvalues and associated eigenvectors of a real symmetric matrix is described. The eigenvalues are either the smallest or those closest to some specified target, which may be in the interior of the spectrum. The underlying algorithm combines the Jacobi-Davidson method with efficient multilevel incomplete LU (ILU) preconditioning. Key features are modest memory requirements and robust convergence to accurate solutions. Parameters needed for incomplete LU preconditioning are automatically computed and may be updated at run time depending on the convergence pattern. The software is easy to use by non-experts and its top level routines are written in FORTRAN 77. Its potentialities are demonstrated on a few applications taken from computational physics. Program summaryProgram title: JADAMILU Catalogue identifier: ADZT_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADZT_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 101 359 No. of bytes in distributed program, including test data, etc.: 7 493 144 Distribution format: tar.gz Programming language: Fortran 77 Computer: Intel or AMD with g77 and pgf; Intel EM64T or Itanium with ifort; AMD Opteron with g77, pgf and ifort; Power (IBM) with xlf90. Operating system: Linux, AIX RAM: problem dependent Word size: real:8; integer: 4 or 8, according to user's choice Classification: 4.8 Nature of problem: Any physical problem requiring the computation of a few eigenvalues of a symmetric matrix. Solution method: Jacobi-Davidson combined with multilevel ILU preconditioning. Additional comments: We supply binaries rather than source code because JADAMILU uses the following external packages: MC64. This software is copyrighted software and not freely available. COPYRIGHT (c) 1999

  1. Software Defined Resource Orchestration System for Multitask Application in Heterogeneous Mobile Cloud Computing

    Directory of Open Access Journals (Sweden)

    Qi Qi

    2016-01-01

    Full Text Available The mobile cloud computing (MCC that combines mobile computing and cloud concept takes wireless access network as the transmission medium and uses mobile devices as the client. When offloading the complicated multitask application to the MCC environment, each task executes individually in terms of its own computation, storage, and bandwidth requirement. Due to user’s mobility, the provided resources contain different performance metrics that may affect the destination choice. Nevertheless, these heterogeneous MCC resources lack integrated management and can hardly cooperate with each other. Thus, how to choose the appropriate offload destination and orchestrate the resources for multitask is a challenge problem. This paper realizes a programming resource provision for heterogeneous energy-constrained computing environments, where a software defined controller is responsible for resource orchestration, offload, and migration. The resource orchestration is formulated as multiobjective optimal problem that contains the metrics of energy consumption, cost, and availability. Finally, a particle swarm algorithm is used to obtain the approximate optimal solutions. Simulation results show that the solutions for all of our studied cases almost can hit Pareto optimum and surpass the comparative algorithm in approximation, coverage, and execution time.

  2. Survey of ANL organization plans for word processors, personal computers, workstations, and associated software

    Energy Technology Data Exchange (ETDEWEB)

    Fenske, K.R.; Rockwell, V.S.

    1992-08-01

    The Computing and Telecommunications Division (CTD) has compiled this Survey of ANL Organization plans for Word Processors, Personal Computers, Workstations, and Associated Software (ANL/TM, Revision 4) to provide DOE and Argonne with a record of recent growth in the acquisition and use of personal computers, microcomputers, and word processors at ANL. Laboratory planners, service providers, and people involved in office automation may find the Survey useful. It is for internal use only, and any unauthorized use is prohibited. Readers of the Survey should use it as a reference document that (1) documents the plans of each organization for office automation, (2) identifies appropriate planners and other contact people in those organizations and (3) encourages the sharing of this information among those people making plans for organizations and decisions about office automation. The Survey supplements information in both the ANL Statement of Site Strategy for Computing Workstations (ANL/TM 458) and the ANL Site Response for the DOE Information Technology Resources Long-Range Plan (ANL/TM 466).

  3. Survey of ANL organization plans for word processors, personal computers, workstations, and associated software. Revision 4

    Energy Technology Data Exchange (ETDEWEB)

    Fenske, K.R.; Rockwell, V.S.

    1992-08-01

    The Computing and Telecommunications Division (CTD) has compiled this Survey of ANL Organization plans for Word Processors, Personal Computers, Workstations, and Associated Software (ANL/TM, Revision 4) to provide DOE and Argonne with a record of recent growth in the acquisition and use of personal computers, microcomputers, and word processors at ANL. Laboratory planners, service providers, and people involved in office automation may find the Survey useful. It is for internal use only, and any unauthorized use is prohibited. Readers of the Survey should use it as a reference document that (1) documents the plans of each organization for office automation, (2) identifies appropriate planners and other contact people in those organizations and (3) encourages the sharing of this information among those people making plans for organizations and decisions about office automation. The Survey supplements information in both the ANL Statement of Site Strategy for Computing Workstations (ANL/TM 458) and the ANL Site Response for the DOE Information Technology Resources Long-Range Plan (ANL/TM 466).

  4. Exploratory research for the development of a computer aided software design environment with the software technology program

    Science.gov (United States)

    Hardwick, Charles

    1991-01-01

    Field studies were conducted by MCC to determine areas of research of mutual interest to MCC and JSC. NASA personnel from the Information Systems Directorate and research faculty from UHCL/RICIS visited MCC in Austin, Texas to examine tools and applications under development in the MCC Software Technology Program. MCC personnel presented workshops in hypermedia, design knowledge capture, and design recovery on site at JSC for ISD personnel. The following programs were installed on workstations in the Software Technology Lab, NASA/JSC: (1) GERM (Graphic Entity Relations Modeler); (2) gIBIS (Graphic Issues Based Information System); and (3) DESIRE (Design Recovery tool). These applications were made available to NASA for inspection and evaluation. Programs developed in the MCC Software Technology Program run on the SUN workstation. The programs do not require special configuration, but they will require larger than usual amounts of disk space and RAM to operate properly.

  5. Development and testing of computer software for nursing assessment and care planning at a spinal cord injury center.

    Science.gov (United States)

    Lloyd, E E; Toth, L L; Rogers, S

    1994-08-01

    This paper describes a pilot project using a Macintosh personal computer and customized software to computerize nursing admission assessment and care planning data. The project setting is a 47-bed Spinal Cord Injury Center with two inpatient units and an outpatient department serving approximately 1,000 patients with spinal cord injury at a Department of Veterans Affairs Medical Center in northern California. The computer software development, implementation, and evaluation are described. This software (MacNursing) was found to be a low cost, customized approach to computerizing spinal cord injury admission assessment data and care planning which reduces repetitive writing and facilitates continuity of care. Personal computers and this software have provided the mechanism for establishing a spinal cord injury patient database.

  6. Software-Design-Analyzer System

    Science.gov (United States)

    Tausworthe, Robert C.

    1991-01-01

    CRISP-90 software-design-analyzer system, update of CRISP-80, is set of computer programs constituting software tool for design and documentation of other software and supporting top-down, hierarchical, modular, structured methodologies for design and programming. Written in Microsoft QuickBasic.

  7. al3c: high-performance software for parameter inference using Approximate Bayesian Computation.

    Science.gov (United States)

    Stram, Alexander H; Marjoram, Paul; Chen, Gary K

    2015-11-01

    The development of Approximate Bayesian Computation (ABC) algorithms for parameter inference which are both computationally efficient and scalable in parallel computing environments is an important area of research. Monte Carlo rejection sampling, a fundamental component of ABC algorithms, is trivial to distribute over multiple processors but is inherently inefficient. While development of algorithms such as ABC Sequential Monte Carlo (ABC-SMC) help address the inherent inefficiencies of rejection sampling, such approaches are not as easily scaled on multiple processors. As a result, current Bayesian inference software offerings that use ABC-SMC lack the ability to scale in parallel computing environments. We present al3c, a C++ framework for implementing ABC-SMC in parallel. By requiring only that users define essential functions such as the simulation model and prior distribution function, al3c abstracts the user from both the complexities of parallel programming and the details of the ABC-SMC algorithm. By using the al3c framework, the user is able to scale the ABC-SMC algorithm in parallel computing environments for his or her specific application, with minimal programming overhead. al3c is offered as a static binary for Linux and OS-X computing environments. The user completes an XML configuration file and C++ plug-in template for the specific application, which are used by al3c to obtain the desired results. Users can download the static binaries, source code, reference documentation and examples (including those in this article) by visiting https://github.com/ahstram/al3c. astram@usc.edu Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  8. A Survey of Structured and Object-Oriented Software Specification Methods and Techniques

    NARCIS (Netherlands)

    Wieringa, Roelf J.

    1998-01-01

    This article surveys techniques used in structured and object-oriented software specification methods. The techniques are classified as techniques for the specification of external interaction and internal decomposition. The external specification techniques are further subdivided into techniques

  9. Software Architecture Evolution

    Science.gov (United States)

    2013-12-01

    Software Architecture Evolution Jeffrey M. Barnes December 2013 CMU-ISR-13-118 Institute for Software Research School of Computer Science Carnegie...DEC 2013 2. REPORT TYPE 3. DATES COVERED 00-00-2013 to 00-00-2013 4. TITLE AND SUBTITLE Software Architecture Evolution 5a. CONTRACT NUMBER 5b...systems eventually undergo changes to their basic architectural structure. Such changes may be prompted by new feature requests, new quality attribute

  10. STRUCTURE AND INTERFACE OF PROGRAM FACILITIES FOR RESEARCH OF PHYSICAL PROCESSES ON COMPUTER MODELS

    Directory of Open Access Journals (Sweden)

    Vitalii M. Bazurin

    2014-11-01

    Full Text Available Research of physical processes on computer models is the one of ways of research approach in the study of general physics in pedagogical universities. The basic elements of software for research of physical processes on computer models are certain in the article: structure of program mean and feature of interface. The author offers his vision of structure of program facilities by means which the computer models of the physical phenomena and processes are realized: block of registration, block of entrance background check, block of design of physical process and block of results verification. A structure of this kinds of software is given, in opinion of author, is universal. The program facilities developed by an author are described in the article.

  11. Computational Methods for Protein Structure Prediction and Modeling Volume 1: Basic Characterization

    CERN Document Server

    Xu, Ying; Liang, Jie

    2007-01-01

    Volume one of this two volume sequence focuses on the basic characterization of known protein structures as well as structure prediction from protein sequence information. The 11 chapters provide an overview of the field, covering key topics in modeling, force fields, classification, computational methods, and struture prediction. Each chapter is a self contained review designed to cover (1) definition of the problem and an historical perspective, (2) mathematical or computational formulation of the problem, (3) computational methods and algorithms, (4) performance results, (5) existing software packages, and (6) strengths, pitfalls, challenges, and future research directions.

  12. Modeling and performance analysis for composite network–compute service provisioning in software-defined cloud environments

    Directory of Open Access Journals (Sweden)

    Qiang Duan

    2015-08-01

    Full Text Available The crucial role of networking in Cloud computing calls for a holistic vision of both networking and computing systems that leads to composite network–compute service provisioning. Software-Defined Network (SDN is a fundamental advancement in networking that enables network programmability. SDN and software-defined compute/storage systems form a Software-Defined Cloud Environment (SDCE that may greatly facilitate composite network–compute service provisioning to Cloud users. Therefore, networking and computing systems need to be modeled and analyzed as composite service provisioning systems in order to obtain thorough understanding about service performance in SDCEs. In this paper, a novel approach for modeling composite network–compute service capabilities and a technique for evaluating composite network–compute service performance are developed. The analytic method proposed in this paper is general and agnostic to service implementation technologies; thus is applicable to a wide variety of network–compute services in SDCEs. The results obtained in this paper provide useful guidelines for federated control and management of networking and computing resources to achieve Cloud service performance guarantees.

  13. PhasePlot: A Software Program for Visualizing Phase Relations Computed Using Thermochemical Models and Databases

    Science.gov (United States)

    Ghiorso, M. S.

    2011-12-01

    A new software program has been developed for Macintosh computers that permits the visualization of phase relations calculated from thermodynamic data-model collections. The data-model collections of MELTS (Ghiorso and Sack, 1995, CMP 119, 197-212), pMELTS (Ghiorso et al., 2002, G-cubed 3, 10.1029/2001GC000217) and the deep mantle database of Stixrude and Lithgow-Bertelloni (2011, GJI 184, 1180-1213) are currently implemented. The software allows users to enter a system bulk composition and a range of reference conditions and then calculate a grid of phase relations. These relations may be visualized in a variety of ways including phase diagrams, phase proportion plots, and contour diagrams of phase compositions and abundances. Results may be exported into Excel or similar spreadsheet applications. Flexibility in stipulating reference conditions permit the construction of temperature-pressure, temperature-volume, entropy-pressure, or entropy-volume display grids. Calculations on the grid are performed for fixed bulk composition or in open systems governed by user specified constraints on component chemical potentials (e.g., specified oxygen fugacity buffers). The calculation engine for the software is optimized for multi-core compute architectures and is very fast, allowing a typical grid of 64 points to be calculated in under 10 seconds on a dual-core laptop/iMac. The underlying computational thermodynamic algorithms have been optimized for speed and robust behavior. Taken together, both of these advances facilitate in classroom demonstrations and permit novice users to work with the program effectively, focusing on problem specification and interpretation of results rather than on manipulation and mechanics of computation - a key feature of an effective instructional tool. The emphasis in this software package is graphical visualization, which aids in better comprehension of complex phase relations in multicomponent systems. Anecdotal experience in using Phase

  14. Implementation of Grid-computing Framework for Simulation in Multi-scale Structural Analysis

    Directory of Open Access Journals (Sweden)

    Data Iranata

    2010-05-01

    Full Text Available A new grid-computing framework for simulation in multi-scale structural analysis is presented. Two levels of parallel processing will be involved in this framework: multiple local distributed computing environments connected by local network to form a grid-based cluster-to-cluster distributed computing environment. To successfully perform the simulation, a large-scale structural system task is decomposed into the simulations of a simplified global model and several detailed component models using various scales. These correlated multi-scale structural system tasks are distributed among clusters and connected together in a multi-level hierarchy and then coordinated over the internet. The software framework for supporting the multi-scale structural simulation approach is also presented. The program architecture design allows the integration of several multi-scale models as clients and servers under a single platform. To check its feasibility, a prototype software system has been designed and implemented to perform the proposed concept. The simulation results show that the software framework can increase the speedup performance of the structural analysis. Based on this result, the proposed grid-computing framework is suitable to perform the simulation of the multi-scale structural analysis.

  15. Personal computer security: part 1. Firewalls, antivirus software, and Internet security suites.

    Science.gov (United States)

    Caruso, Ronald D

    2003-01-01

    Personal computer (PC) security in the era of the Health Insurance Portability and Accountability Act of 1996 (HIPAA) involves two interrelated elements: safeguarding the basic computer system itself and protecting the information it contains and transmits, including personal files. HIPAA regulations have toughened the requirements for securing patient information, requiring every radiologist with such data to take further precautions. Security starts with physically securing the computer. Account passwords and a password-protected screen saver should also be set up. A modern antivirus program can easily be installed and configured. File scanning and updating of virus definitions are simple processes that can largely be automated and should be performed at least weekly. A software firewall is also essential for protection from outside intrusion, and an inexpensive hardware firewall can provide yet another layer of protection. An Internet security suite yields additional safety. Regular updating of the security features of installed programs is important. Obtaining a moderate degree of PC safety and security is somewhat inconvenient but is necessary and well worth the effort. Copyright RSNA, 2003

  16. Combining Topological Hardware and Topological Software: Color-Code Quantum Computing with Topological Superconductor Networks

    Directory of Open Access Journals (Sweden)

    Daniel Litinski

    2017-09-01

    Full Text Available We present a scalable architecture for fault-tolerant topological quantum computation using networks of voltage-controlled Majorana Cooper pair boxes and topological color codes for error correction. Color codes have a set of transversal gates which coincides with the set of topologically protected gates in Majorana-based systems, namely, the Clifford gates. In this way, we establish color codes as providing a natural setting in which advantages offered by topological hardware can be combined with those arising from topological error-correcting software for full-fledged fault-tolerant quantum computing. We provide a complete description of our architecture, including the underlying physical ingredients. We start by showing that in topological superconductor networks, hexagonal cells can be employed to serve as physical qubits for universal quantum computation, and we present protocols for realizing topologically protected Clifford gates. These hexagonal-cell qubits allow for a direct implementation of open-boundary color codes with ancilla-free syndrome read-out and logical T gates via magic-state distillation. For concreteness, we describe how the necessary operations can be implemented using networks of Majorana Cooper pair boxes, and we give a feasibility estimate for error correction in this architecture. Our approach is motivated by nanowire-based networks of topological superconductors, but it could also be realized in alternative settings such as quantum-Hall–superconductor hybrids.

  17. Quantum computation with indefinite causal structures

    Science.gov (United States)

    Araújo, Mateus; Guérin, Philippe Allard; Baumeler, ńmin

    2017-11-01

    One way to study the physical plausibility of closed timelike curves (CTCs) is to examine their computational power. This has been done for Deutschian CTCs (D-CTCs) and postselection CTCs (P-CTCs), with the result that they allow for the efficient solution of problems in PSPACE and PP, respectively. Since these are extremely powerful complexity classes, which are not expected to be solvable in reality, this can be taken as evidence that these models for CTCs are pathological. This problem is closely related to the nonlinearity of this models, which also allows, for example, cloning quantum states, in the case of D-CTCs, or distinguishing nonorthogonal quantum states, in the case of P-CTCs. In contrast, the process matrix formalism allows one to model indefinite causal structures in a linear way, getting rid of these effects and raising the possibility that its computational power is rather tame. In this paper, we show that process matrices correspond to a linear particular case of P-CTCs, and therefore that its computational power is upperbounded by that of PP. We show, furthermore, a family of processes that can violate causal inequalities but nevertheless can be simulated by a causally ordered quantum circuit with only a constant overhead, showing that indefinite causality is not necessarily hard to simulate.

  18. Wing-Body Aeroelasticity Using Finite-Difference Fluid/Finite-Element Structural Equations on Parallel Computers

    Science.gov (United States)

    Byun, Chansup; Guruswamy, Guru P.; Kutler, Paul (Technical Monitor)

    1994-01-01

    In recent years significant advances have been made for parallel computers in both hardware and software. Now parallel computers have become viable tools in computational mechanics. Many application codes developed on conventional computers have been modified to benefit from parallel computers. Significant speedups in some areas have been achieved by parallel computations. For single-discipline use of both fluid dynamics and structural dynamics, computations have been made on wing-body configurations using parallel computers. However, only a limited amount of work has been completed in combining these two disciplines for multidisciplinary applications. The prime reason is the increased level of complication associated with a multidisciplinary approach. In this work, procedures to compute aeroelasticity on parallel computers using direct coupling of fluid and structural equations will be investigated for wing-body configurations. The parallel computer selected for computations is an Intel iPSC/860 computer which is a distributed-memory, multiple-instruction, multiple data (MIMD) computer with 128 processors. In this study, the computational efficiency issues of parallel integration of both fluid and structural equations will be investigated in detail. The fluid and structural domains will be modeled using finite-difference and finite-element approaches, respectively. Results from the parallel computer will be compared with those from the conventional computers using a single processor. This study will provide an efficient computational tool for the aeroelastic analysis of wing-body structures on MIMD type parallel computers.

  19. Joint Logistics Commanders’ Workshop on Post Deployment Software Support (PDSS) for Mission-Critical Computer Software. Volume 2. Workshop Proceedings.

    Science.gov (United States)

    1984-06-01

    fit into the software life cycle? At what specific points is it employed? What is the relationship of IV&V to test, integracion , QA, system...computer. It is recognized that organizational and social factors are an important part of the work environment. These factors must be considered in the...the system and the organiza- tionai and social setting (ELZER, P.F., 1979). In industrial engineering there has been considerable work on facilities

  20. Development of a computer-aided design software for dental splint in orthognathic surgery

    Science.gov (United States)

    Chen, Xiaojun; Li, Xing; Xu, Lu; Sun, Yi; Politis, Constantinus; Egger, Jan

    2016-12-01

    In the orthognathic surgery, dental splints are important and necessary to help the surgeon reposition the maxilla or mandible. However, the traditional methods of manual design of dental splints are difficult and time-consuming. The research on computer-aided design software for dental splints is rarely reported. Our purpose is to develop a novel special software named EasySplint to design the dental splints conveniently and efficiently. The design can be divided into two steps, which are the generation of initial splint base and the Boolean operation between it and the maxilla-mandibular model. The initial splint base is formed by ruled surfaces reconstructed using the manually picked points. Then, a method to accomplish Boolean operation based on the distance filed of two meshes is proposed. The interference elimination can be conducted on the basis of marching cubes algorithm and Boolean operation. The accuracy of the dental splint can be guaranteed since the original mesh is utilized to form the result surface. Using EasySplint, the dental splints can be designed in about 10 minutes and saved as a stereo lithography (STL) file for 3D printing in clinical applications. Three phantom experiments were conducted and the efficiency of our method was demonstrated.

  1. Hand-held computers in healthcare: what software programs are available?

    Science.gov (United States)

    Gillingham, Wayne; Holt, Alec; Gillies, John

    2002-09-27

    The technology sector of healthcare is entering a new evolutionary phase. The medical community has an obligation to the public to provide the safest, most effective healthcare possible. This is more achievable with the use of computer technology at the point of care, and small, portable devices could fulfil this role. A PriceWaterhouse Coopers 2001 survey on information technology in physician practices found that 60% of respondents say that physicians in their organisation use personal digital assistants (PDAs), compared with 26% in the 2000 technology survey. This trend is expected to continue to the point where these devices will have their position on a physician s desk next to the stethoscope. Once this electronic evolution occurs, doctors will be able to practice medicine with greater ease and safety. In our opinion, the new generation of PDA mobile devices will be the tools to enable a transformation of healthcare to a paperless, wireless world. This article focuses on uses of PDAs in healthcare, whether by the registrar, consultant, nurse, student, teacher, patient, medical or surgical director. Current PDA healthcare software is categorised and discussed in the following five groups: 1) reference/text book; 2) calculator; 3) patient management/logbook; 4) personal clinical/study notebook; 5) utility software.

  2. Performance Assessment of Three Rendering Engines in 3D Computer Graphics Software

    Directory of Open Access Journals (Sweden)

    Žan Vidmar

    2015-03-01

    Full Text Available The aim of the research was the determination of testing conditions and visual and numerical evaluation of renderings made with three different rendering engines in Maya software, which is widely used for educational and computer art purposes. In the theoretical part the overview of light phenomena and their simulation in virtual space is presented. This is followed by a detailed presentation of the main rendering methods and the results and limitations of their applications to 3D objects. At the end of the theoretical part the importance of a proper testing scene and especially the role of Cornell box are explained. In the experimental part the terms and conditions as well as hardware and software used for the research are presented. This is followed by a description of the procedures, where we focused on the rendering quality and time, which enabled the comparison of settings of different render engines and determination of conditions for further rendering of testing scenes. The experimental part continued with rendering a variety of simple virtual scenes including Cornell box and virtual object with different materials and colours. Apart from visual evaluation, which was the starting point for comparison of renderings, a procedure for numerical estimation and colour deviations of renderings using the selected regions of interest in the final images is presented.

  3. Development of a computer-aided design software for dental splint in orthognathic surgery

    Science.gov (United States)

    Chen, Xiaojun; Li, Xing; Xu, Lu; Sun, Yi; Politis, Constantinus; Egger, Jan

    2016-01-01

    In the orthognathic surgery, dental splints are important and necessary to help the surgeon reposition the maxilla or mandible. However, the traditional methods of manual design of dental splints are difficult and time-consuming. The research on computer-aided design software for dental splints is rarely reported. Our purpose is to develop a novel special software named EasySplint to design the dental splints conveniently and efficiently. The design can be divided into two steps, which are the generation of initial splint base and the Boolean operation between it and the maxilla-mandibular model. The initial splint base is formed by ruled surfaces reconstructed using the manually picked points. Then, a method to accomplish Boolean operation based on the distance filed of two meshes is proposed. The interference elimination can be conducted on the basis of marching cubes algorithm and Boolean operation. The accuracy of the dental splint can be guaranteed since the original mesh is utilized to form the result surface. Using EasySplint, the dental splints can be designed in about 10 minutes and saved as a stereo lithography (STL) file for 3D printing in clinical applications. Three phantom experiments were conducted and the efficiency of our method was demonstrated. PMID:27966601

  4. Computer-assisted teaching of skin flap surgery: validation of a mobile platform software for medical students.

    Directory of Open Access Journals (Sweden)

    David P de Sena

    Full Text Available The purpose of this study was to develop and validate a multimedia software application for mobile platforms to assist in the teaching and learning process of design and construction of a skin flap. Traditional training in surgery is based on learning by doing. Initially, the use of cadavers and animal models appeared to be a valid alternative for training. However, many conflicts with these training models prompted progression to synthetic and virtual reality models. Fifty volunteer fifth- and sixth-year medical students completed a pretest and were randomly allocated into two groups of 25 students each. The control group was exposed for 5 minutes to a standard text-based print article, while the test group used multimedia software describing how to fashion a rhomboid flap. Each group then performed a cutaneous flap on a training bench model while being evaluated by three blinded BSPS (Brazilian Society of Plastic Surgery board-certified surgeons using the OSATS (Objective Structured Assessment of Technical Skill protocol and answered a post-test. The text-based group was then tested again using the software. The computer-assisted learning (CAL group had superior performance as confirmed by checklist scores (p<0.002, overall global assessment (p = 0.017 and post-test results (p<0.001. All participants ranked the multimedia method as the best study tool. CAL learners exhibited better subjective and objective performance when fashioning rhomboid flaps as compared to those taught with standard print material. These findings indicate that students preferred to learn using the multimedia method.

  5. PUMPSIM: a software package for simulating computer-controlled drug infusion pumps.

    Science.gov (United States)

    Hamann, G E; Doyle, D J

    1992-01-01

    Advances in microcomputer technology have lead to the development of volumetric infusion pumps which can be controlled by a remote computer. Such pumps offer powerful capabilities for the administration of medications and fluids. Complex control strategies can be preprogrammed and the pump allowed to run with a minimum of supervision. The safety of patients is of the utmost importance. Therefore, great care must be taken to implement robust error avoidance, detection and correction protocols in the host control program. This paper describes a software simulator of a commercial infusion pump (Abbott Lifecare Model 4). Using the simulator in the development and debugging of control programs has several advantages over using the real pump: it provides detailed pump status information and it can stimulate various error and alarm conditions to comprehensively test the error recovery procedures of the control program.

  6. New Method and Software for Computer-Aided Flowsheet Design and Analysis

    DEFF Research Database (Denmark)

    K.Tula, Anjan; Gani, Rafiqul; Eden, Mario R.

    2017-01-01

    Process synthesis-design involves identification of the best processing route to produce a desired product from a specified set of raw materials. In this work a systematic, multistage method and corresponding software (ProCAFD) is developed for sustainable process synthesis-design and analysis...... in the same way as atoms or groups of atoms are synthesized to form molecules in computer aided molecular design (CAMD) techniques. Another important aspect of this method is the integration of economic, sustainability and LCA analyses in the early stages of process synthesis to identify process hotspots...... and further improve the process using innovative design strategies. An overview of the key concepts and the method is presented along with a case study to generate sustainable designs for production of benzene from toluene and hydrogen....

  7. MIAWARE Software

    DEFF Research Database (Denmark)

    Wilkowski, Bartlomiej; Pereira, Oscar N. M.; Dias, Paulo

    2008-01-01

    This article presents MIAWARE, a software for Medical Image Analysis With Automated Reporting Engine, which was designed and developed for doctor/radiologist assistance. It allows to analyze an image stack from computed axial tomography scan of lungs (thorax) and, at the same time, to mark all...... pathologies on images and report their characteristics. The reporting process is normalized - radiologists cannot describe pathological changes with their own words, but can only use some terms from a specific vocabulary set provided by the software. Consequently, a normalized radiological report...... is automatically generated. Furthermore, MIAWARE software is accompanied with an intelligent search engine for medical reports, based on the relations between parts of the lungs. A logical structure of the lungs is introduced to the search algorithm through the specially developed ontology. As a result...

  8. Computation of Estonian CORS data using Bernese 5.2 and Gipsy 6.4 softwares

    Science.gov (United States)

    Kollo, Karin; Kall, Tarmo; Liibusk, Aive

    2017-04-01

    GNSS permanent station network in Estonia (ESTREF) was established already in 2007. In 2014-15 extensive reconstruction of ESTREF was carried out, including the establishment of 18 new stations, change of the hardware in CORS stations as well as establishing GNSS-RTK service for the whole Estonia. For GNSS-RTK service one needs precise coordinates in well-defined reference frame, i.e., ETRS89. For long time stability of stations and time-series analysis the re-processing of Estonian CORS data is ongoing. We re-process data from 2007 until 2015 with program Bernese GNSS 5.2 (Dach, 2015). For the set of ESTREF stations established in 2007, we perform as well computations with GIPSY 6.4 software (Ries et al., 2015). In the computations daily GPS-only solution was used. For precise orbits, final products from CODE (CODE analysis centre at the Astronomical Institute of the University of Bern) and JPL (Jet Propulsion Laboratory) for Bernese and GIPSY solutions were used, respectively. The cut-off angle was set to 10 degrees in order to avoid near-field multipath influence. In GIPSY, precise point positioning method with fixing ambiguities was used. Bernese calculations were performed based on double difference processing. Antenna phase centers were modelled based on igs08.atx and epnc_08.atx files. Vienna mapping function was used for mapping tropospheric delays. For the GIPSY solution, the higher order ionospheric term was modelled based on IRI-2012b model. For the Bernese solution higher order ionospheric term was neglected. FES2004 ocean tide loading model was used for the both computation strategies. As a result, two solutions using different scientific GNSS computation programs were obtained. The results from Bernese and GIPSY solutions were compared, using station repeatability values, RMS and coordinate differences. KEYWORDS: GNSS reference station network, Bernese GNSS 5.2, Gipsy 6.4, Estonia. References: Dach, R., S. Lutz, P. Walser, P. Fridez (Eds); 2015

  9. Homogenization and structural topology optimization theory, practice and software

    CERN Document Server

    Hassani, Behrooz

    1999-01-01

    Structural topology optimization is a fast growing field that is finding numerous applications in automotive, aerospace and mechanical design processes. Homogenization is a mathematical theory with applications in several engineering problems that are governed by partial differential equations with rapidly oscillating coefficients Homogenization and Structural Topology Optimization brings the two concepts together and successfully bridges the previously overlooked gap between the mathematical theory and the practical implementation of the homogenization method. The book is presented in a unique self-teaching style that includes numerous illustrative examples, figures and detailed explanations of concepts. The text is divided into three parts which maintains the book's reader-friendly appeal.

  10. A comparison of conventional and computer-assisted semen analysis (CRISMAS software) using samples from 166 young Danish men

    DEFF Research Database (Denmark)

    Vested, Anne; Ramlau-Hansen, Cecilia H; Bonde, Jens P

    2011-01-01

    The aim of the present study was to compare assessments of sperm concentration and sperm motility analysed by conventional semen analysis with those obtained by computer-assisted semen analysis (CASA) (Copenhagen Rigshospitalet Image House Sperm Motility Analysis System (CRISMAS) 4.6 software) us...... and motility analysis. This needs to be accounted for in clinics using this software and in studies of determinants of these semen characteristics.......The aim of the present study was to compare assessments of sperm concentration and sperm motility analysed by conventional semen analysis with those obtained by computer-assisted semen analysis (CASA) (Copenhagen Rigshospitalet Image House Sperm Motility Analysis System (CRISMAS) 4.6 software......) using semen samples from 166 young Danish men. The CRISMAS software identifies sperm concentration and classifies spermatozoa into three motility categories. To enable comparison of the two methods, the four motility stages obtained by conventional semen analysis were, based on their velocity...

  11. Multi-physics fluid-structure interaction modelling software

    CSIR Research Space (South Africa)

    Malan, AG

    2008-11-01

    Full Text Available The CSIR reseachers developed a new ground-breaking sofware modelling technlogies to be used in the design of safe and efficient next-generation aircraft. The field of Fluid-structure interaction (FIS) covers a massive range of engineering problems...

  12. The Metric Lens : Visualizing Metrics and Structure on Software Diagrams

    NARCIS (Netherlands)

    Byelas, Heorhiy; Telea, Alexandru; Hassan, AE; Zaidman, A; DiPenta, M

    2008-01-01

    We present the metric lens, a new visualization of method-level code metrics atop UML class diagrams, which allows performing metric-metric and metric-structure correlations on large diagrams. Me demonstrate air interactive visualization tool in which users can quickly specify a wide palette of

  13. 78 FR 47804 - Verification, Validation, Reviews, and Audits for Digital Computer Software Used in Safety...

    Science.gov (United States)

    2013-08-06

    ... Software Verification and Validation,'' and IEEE Std. 1028-2008, ``IEEE Standard for Software Reviews and... Institute of Electrical and Electronic Engineers (IEEE) Standard 1012-2004, ``IEEE ] Standard for Software Verification and Validation,'' and IEEE Std. 1028-2008, ``IEEE Standard for Software Reviews and Audits...

  14. Variation of densitometry on computed tomography in COPD--influence of different software tools.

    Directory of Open Access Journals (Sweden)

    Mark O Wielpütz

    Full Text Available Quantitative multidetector computed tomography (MDCT as a potential biomarker is increasingly used for severity assessment of emphysema in chronic obstructive pulmonary disease (COPD. Aim of this study was to evaluate the user-independent measurement variability between five different fully-automatic densitometry software tools.MDCT and full-body plethysmography incl. forced expiratory volume in 1s and total lung capacity were available for 49 patients with advanced COPD (age = 64±9 years, forced expiratory volume in 1 s = 31±6% predicted. Measurement variation regarding lung volume, emphysema volume, emphysema index, and mean lung density was evaluated for two scientific and three commercially available lung densitometry software tools designed to analyze MDCT from different scanner types.One scientific tool and one commercial tool failed to process most or all datasets, respectively, and were excluded. One scientific and another commercial tool analyzed 49, the remaining commercial tool 30 datasets. Lung volume, emphysema volume, emphysema index and mean lung density were significantly different amongst these three tools (p<0.001. Limits of agreement for lung volume were [-0.195, -0.052 l], [-0.305, -0.131 l], and [-0.123, -0.052 l] with correlation coefficients of r = 1.00 each. Limits of agreement for emphysema index were [-6.2, 2.9%], [-27.0, 16.9%], and [-25.5, 18.8%], with r = 0.79 to 0.98. Correlation of lung volume with total lung capacity was good to excellent (r = 0.77 to 0.91, p<0.001, but segmented lung volume (6.7±1.3-6.8±1.3 l were significantly lower than total lung capacity (7.7±1.7 l, p<0.001.Technical incompatibilities hindered evaluation of two of five tools. The remaining three showed significant measurement variation for emphysema, hampering quantitative MDCT as a biomarker in COPD. Follow-up studies should currently use identical software, and standardization efforts should encompass software as

  15. Why are people downloading the freeware AIDA diabetes computing software program: a pilot study.

    Science.gov (United States)

    Lehmann, Eldon D

    2002-01-01

    The purpose of this paper is to report a pilot survey about why people are downloading the AIDA interactive educational diabetes simulator. AIDA is a diabetes computer program that permits the interactive simulation of plasma insulin and blood glucose profiles for teaching, demonstration, and self-learning purposes. It has been made freely available, without charge, on the Internet as a noncommercial contribution to continuing diabetes education. Since its launch in 1996 well over 200,000 visits have been logged at the main AIDA Website--www.2aida.org--and over 40,000 copies of the AIDA program have been downloaded free-of-charge. This article documents a pilot survey of comments left by Website visitors while they were downloading the AIDA software, before they had a chance to actually use the program. The overall paradigm adopted for this study has endeavored to establish why people are resorting to the Internet to obtain diabetes information. Specific intended goals of the study were: (1) to demonstrate ongoing use of the World Wide Web for surveying diabetes software users by obtaining their free-text comments; (2) to identify what sort of things people were planning to do with the AIDA software simulator; and (3) to more generally gain some insight into why people are turning to the Web for healthcare-related information. The Internet-based survey methodology was found to be robust and reliable. Over an 8-month period (from February 2, 2001 to October 1, 2001) 642 responses were received. During the corresponding period 2,248 actual visits were made to the Website survey page--giving a response rate to this pilot study of 28.6%. Responses were received from participants in over 56 countries--although over half of these (n = 343; 53.4%) originated from the United States and United Kingdom. Two hundred forty-four responses (38.0%) were received from patients with diabetes, and 73 (11.4%) from relatives of patients, with fewer responses from doctors, students

  16. High performance computing and communications Grand Challenges program: Computational structural biology. Final report, August 15, 1992--January 14, 1997

    Energy Technology Data Exchange (ETDEWEB)

    Solomon, J.E.

    1997-10-02

    The Grand Challenge project consists of two elements: (1) a hierarchical methodology for 3D protein structure prediction; and (2) development of a parallel computing environment, the Protein Folding Workbench, for carrying out a variety of protein structure prediction/modeling computations. During the first three years of this project the author focused on the use of selected proteins from the Brookhaven Protein Data Base (PDB) of known structures to provide validation of the prediction algorithms and their software implementation, both serial and parallel. Two proteins in particular have been selected to provide the project with direct interaction with experimental molecular biology. A variety of site-specific mutagenesis experiments are performed on these two proteins to explore the many-to-one mapping characteristics of sequence to structure.

  17. Multi-atlas segmentation of subcortical brain structures via the AutoSeg software pipeline

    Science.gov (United States)

    Wang, Jiahui; Vachet, Clement; Rumple, Ashley; Gouttard, Sylvain; Ouziel, Clémentine; Perrot, Emilie; Du, Guangwei; Huang, Xuemei; Gerig, Guido; Styner, Martin

    2014-01-01

    Automated segmenting and labeling of individual brain anatomical regions, in MRI are challenging, due to the issue of individual structural variability. Although atlas-based segmentation has shown its potential for both tissue and structure segmentation, due to the inherent natural variability as well as disease-related changes in MR appearance, a single atlas image is often inappropriate to represent the full population of datasets processed in a given neuroimaging study. As an alternative for the case of single atlas segmentation, the use of multiple atlases alongside label fusion techniques has been introduced using a set of individual “atlases” that encompasses the expected variability in the studied population. In our study, we proposed a multi-atlas segmentation scheme with a novel graph-based atlas selection technique. We first paired and co-registered all atlases and the subject MR scans. A directed graph with edge weights based on intensity and shape similarity between all MR scans is then computed. The set of neighboring templates is selected via clustering of the graph. Finally, weighted majority voting is employed to create the final segmentation over the selected atlases. This multi-atlas segmentation scheme is used to extend a single-atlas-based segmentation toolkit entitled AutoSeg, which is an open-source, extensible C++ based software pipeline employing BatchMake for its pipeline scripting, developed at the Neuro Image Research and Analysis Laboratories of the University of North Carolina at Chapel Hill. AutoSeg performs N4 intensity inhomogeneity correction, rigid registration to a common template space, automated brain tissue classification based skull-stripping, and the multi-atlas segmentation. The multi-atlas-based AutoSeg has been evaluated on subcortical structure segmentation with a testing dataset of 20 adult brain MRI scans and 15 atlas MRI scans. The AutoSeg achieved mean Dice coefficients of 81.73% for the subcortical structures

  18. Exercise manual for the Augmented Computer Exercise for Inspection Training (ACE-IT) software

    Energy Technology Data Exchange (ETDEWEB)

    Dobranich, P.R.; Widney, T.W.; Goolsby, P.T. [Sandia National Labs., Albuquerque, NM (United States). Cooperative Monitoring Center and Regional Security; Nelson, J.D.; Evanko, D.A. [Ogden Environmental and Energy Services, Inc., Albuquerque, NM (United States)

    1997-09-01

    The on-site inspection provisions in many current and proposed arms control agreements require extensive preparation and training on the part of both the Inspected Party and the Inspection Team. Current training techniques include table-top inspections and practice inspections. The Augmented Computer Exercise for Inspection Training (ACE-IT), an interactive computer training tool, increases the utility of table-top inspections. ACE-IT has been designed to provide training for a hypothetical challenge inspection under the Chemical Weapons Convention (CWC); however, this training tool can be modified for other inspection regimes. Although ACE-IT provides training from notification of an inspection through post-inspection activities, the primary emphasis of ACE-IT is in the inspection itself--particularly with the concept of managed access. ACE-IT also demonstrates how inspection provisions impact compliance determination and the protection of sensitive information. The Exercise Manual supplements the ACE-IT software by providing general information on on-site inspections and detailed information for the CWC challenge inspection exercise. The detailed information includes the pre-inspection briefing, maps, list of sensitive items, medical records, and shipping records.

  19. Ermod: fast and versatile computation software for solvation free energy with approximate theory of solutions.

    Science.gov (United States)

    Sakuraba, Shun; Matubayasi, Nobuyuki

    2014-08-05

    ERmod is a software package to efficiently and approximately compute the solvation free energy using the method of energy representation. Molecular simulation is to be conducted at two condensed-phase systems of the solution of interest and the reference solvent with test-particle insertion of the solute. The subprogram ermod in ERmod then provides a set of energy distribution functions from the simulation trajectories, and another subprogram slvfe determines the solvation free energy from the distribution functions through an approximate functional. This article describes the design and implementation of ERmod, and illustrates its performance in solvent water for two organic solutes and two protein solutes. Actually, the free-energy computation with ERmod is not restricted to the solvation in homogeneous medium such as fluid and polymer and can treat the binding into weakly ordered system with nano-inhomogeneity such as micelle and lipid membrane. ERmod is available on web at http://sourceforge.net/projects/ermod. Copyright © 2014 Wiley Periodicals, Inc.

  20. Principles of design and database structure of "Resort Manager" sanatorium software

    OpenAIRE

    Моторний, Анатолій Павлович; Злепко, Сергій Макарович; Коваль, Леонід Григорович; Костішин, Сергій Володимирович

    2013-01-01

    The present article deals with problems of the spa facilities management, and the database structure of modern sanatorium complex software, which automates the functions and activities and also improves the quality of institutions streamline information flows and brings customer’s service to a new higher level. The conclusions about the feasibility of using this system in the workflow institution.The “Resort Manager" sanatorium software consists of 3 units: “The Settle”,”The Setting Procedure...

  1. Logic and algebraic structures in quantum computing

    CERN Document Server

    Eskandarian, Ali; Harizanov, Valentina S

    2016-01-01

    Arising from a special session held at the 2010 North American Annual Meeting of the Association for Symbolic Logic, this volume is an international cross-disciplinary collaboration with contributions from leading experts exploring connections across their respective fields. Themes range from philosophical examination of the foundations of physics and quantum logic, to exploitations of the methods and structures of operator theory, category theory, and knot theory in an effort to gain insight into the fundamental questions in quantum theory and logic. The book will appeal to researchers and students working in related fields, including logicians, mathematicians, computer scientists, and physicists. A brief introduction provides essential background on quantum mechanics and category theory, which, together with a thematic selection of articles, may also serve as the basic material for a graduate course or seminar.

  2. Understanding Structures and Affordances of Extended Teams in Global Software Development

    DEFF Research Database (Denmark)

    Ali Babar, Muhammad; Zahedi, Mansooreh

    2013-01-01

    Growing popularity of Global Software Development (GSD) has resulted in an increasing number of cross-organizational teams that are formed according to Extended Team Model (ETM). There is little known about the structures (work, social, and communication) that may exist in ETM and what affordances...... into the types of structures and associated challenges for implementing ETM in GSD....

  3. Exploring the impact of socio-technical core-periphery structures in open source software development

    NARCIS (Netherlands)

    Amrit, Chintan Amrit; van Hillegersberg, Jos

    2010-01-01

    In this paper we apply the social network concept of core-periphery structure to the socio-technical structure of a software development team. We propose a socio-technical pattern that can be used to locate emerging coordination problems in Open Source projects. With the help of our tool and method

  4. A Quantitative Analysis of the Effect of Organizational Structure on Software Engineering Management.

    Science.gov (United States)

    1982-05-01

    The effect of organizational structure on software engineering management has been quantified using graph models of the organizational flow of...maximum number of interfaces that an individual can handle effectively. Optimization of the organizational structure is achieved by minimizing the result of trade-offs between the L’s and the B’s.

  5. Angpow: a software for the fast computation of accurate tomographic power spectra

    Science.gov (United States)

    Campagne, J.-E.; Neveu, J.; Plaszczynski, S.

    2017-06-01

    Aims: The statistical distribution of galaxies is a powerful probe to constrain cosmological models and gravity. In particular, the matter power spectrum P(k) provides information about the cosmological distance evolution and the galaxy clustering. However the building of P(k) from galaxy catalogs requires a cosmological model to convert angles on the sky and redshifts into distances, which leads to difficulties when comparing data with predicted P(k) from other cosmological models, and for photometric surveys like the Large Synoptic Survey Telescope (LSST). The angular power spectrum Cℓ(z1,z2) between two bins located at redshift z1 and z2 contains the same information as the matter power spectrum, and is free from any cosmological assumption, but the prediction of Cℓ(z1,z2) from P(k) is a costly computation when performed precisely. Methods: The Angpow software aims at quickly and accurately computing the auto (z1 = z2) and cross (z1 ≠ z2) angular power spectra between redshift bins. We describe the developed algorithm based on developments on the Chebyshev polynomial basis and on the Clenshaw-Curtis quadrature method. We validate the results with other codes, and benchmark the performance. Results: Angpow is flexible and can handle any user-defined power spectra, transfer functions, and redshift selection windows. The code is fast enough to be embedded inside programs exploring large cosmological parameter spaces through the Cℓ(z1,z2) comparison with data. We emphasize that the Limber's approximation, often used to speed up the computation, gives incorrect Cℓ values for cross-correlations. The C++ code is available from http://https://gitlab.in2p3.fr/campagne/AngPow

  6. Monitoring of computing resource use of active software releases at ATLAS

    Science.gov (United States)

    Limosani, Antonio; ATLAS Collaboration

    2017-10-01

    The LHC is the world’s most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the TierO at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end-user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as “MemoryMonitor”, to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and displayed in plots generated using Python visualization libraries and collected into pre-formatted auto-generated Web pages, which allow the ATLAS developer community to track the performance of their algorithms. This information is however preferentially filtered to domain leaders and developers through the use of JIRA and via reports given at ATLAS software meetings. Finally, we take a glimpse of the future by reporting on the expected CPU and RAM usage in benchmark workflows associated with the High Luminosity LHC and anticipate the ways performance monitoring will evolve to understand and benchmark future workflows.

  7. Gmat. A software tool for the computation of the rovibrational G matrix

    Science.gov (United States)

    Castro, M. E.; Niño, A.; Muñoz-Caro, C.

    2009-07-01

    Gmat is a C++ program able to compute the rovibrational G matrix in molecules of arbitrary size. This allows the building of arbitrary rovibrational Hamiltonians. In particular, the program is designed to work with the structural results of potential energy hypersurface mappings computed in computer clusters or computational Grid environments. In the present version, 1.0, the program uses internal coordinates as vibrational coordinates, with the principal axes of inertia as body-fixed system. The main design implements a complete separation of the interface and functional parts of the program. The interface part permits the automatic reading of the molecular structures from the output files of different electronic structure codes. At present, Gamess and Gaussian output files are allowed. To such an end, use is made of the object orientation polymorphism characteristic. The functional part computes numerically the derivatives of the nuclear positions respect to the vibrational coordinates. Very accurate derivatives are obtained by using central differences embedded in a nine levels Richardson extrapolation procedure. Program summaryProgram title: Gmat Catalogue identifier: AECZ_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AECZ_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 17 023 No. of bytes in distributed program, including test data, etc.: 274 714 Distribution format: tar.gz Programming language: Standard C++ Computer: All running Linux/Windows Operating system: Linux, Windows Classification: 16.2 Nature of problem: Computation of the rovibrational G matrix in molecules of any size. This allows the building of arbitrary rovibrational Hamiltonians. It must be possible to obtain the input data from the output files of standard electronic structure codes

  8. High-Quality Random Number Generation Software for High-Performance Computing Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Random number (RN) generation is the key software component that permits random sampling. Software for parallel RN generation (RNG) should be based on RNGs that are...

  9. Some key considerations in evolving a computer system and software engineering support environment for the space station program

    Science.gov (United States)

    Mckay, C. W.; Bown, R. L.

    1985-01-01

    The space station data management system involves networks of computing resources that must work cooperatively and reliably over an indefinite life span. This program requires a long schedule of modular growth and an even longer period of maintenance and operation. The development and operation of space station computing resources will involve a spectrum of systems and software life cycle activities distributed across a variety of hosts, an integration, verification, and validation host with test bed, and distributed targets. The requirement for the early establishment and use of an apporopriate Computer Systems and Software Engineering Support Environment is identified. This environment will support the Research and Development Productivity challenges presented by the space station computing system.

  10. Design, Implementation and Optimization of Innovative Internet Access Networks, based on Fog Computing and Software Defined Networking

    OpenAIRE

    Iotti, Nicola

    2017-01-01

    1. DESIGN In this dissertation we introduce a new approach to Internet access networks in public spaces, such as Wi-Fi network commonly known as Hotspot, based on Fog Computing (or Edge Computing), Software Defined Networking (SDN) and the deployment of Virtual Machines (VM) and Linux containers, on the edge of the network. In this vision we deploy specialized network elements, called Fog Nodes, on the edge of the network, able to virtualize the physical infrastructure and expose APIs to e...

  11. 77 FR 50723 - Verification, Validation, Reviews, and Audits for Digital Computer Software Used in Safety...

    Science.gov (United States)

    2012-08-22

    ... Institute of Electrical and Electronics Engineers (IEEE) Standard 1012- 2004, ``IEEE Standard for Software Verification and Validation'' and IEEE Standard 1028-2008, ``IEEE Standard for Software Reviews and Audits... hardware while IEEE Std. 1012- 2004 only provides software V&V task guidance. DATES: Submit comments by...

  12. Incorporating a Human-Computer Interaction Course into Software Development Curriculums

    Science.gov (United States)

    Janicki, Thomas N.; Cummings, Jeffrey; Healy, R. Joseph

    2015-01-01

    Individuals have increasing options on retrieving information related to hardware and software. Specific hardware devices include desktops, tablets and smart devices. Also, the number of software applications has significantly increased the user's capability to access data. Software applications include the traditional web site, smart device…

  13. Benchmarking Close-range Structure from Motion 3D Reconstruction Software under Varying Capturing Conditions

    DEFF Research Database (Denmark)

    Nikolov, Ivan Adriyanov; Madsen, Claus B.

    2016-01-01

    Structure from Motion 3D reconstruction has become widely used in recent years in a number of fields such as industrial surface in- inspection, archeology, cultural heritage preservation and geomapping. A number of software solutions have been released using variations of this technique. In this ......Structure from Motion 3D reconstruction has become widely used in recent years in a number of fields such as industrial surface in- inspection, archeology, cultural heritage preservation and geomapping. A number of software solutions have been released using variations of this technique...

  14. STRUCTURAL COMPONENTS OF THE PROFESSIONAL FOREIGN LANGUAGE COMMUNICATIVE COMPETENCE OF THE FUTURE SOFTWARE ENGINEERS

    Directory of Open Access Journals (Sweden)

    Alina Mykytyshyn

    2016-09-01

    Full Text Available The concepts of the professional foreign language communicative training and professional foreign language communicative competence of the future software engineers have been analyzed in the article. Different approaches to the interpretation of these concepts in the contemporary scientific literature have been described. The essence and the structural components of the professional foreign language communicative competence of software engineers in accordance with the objectives and tasks of the program of study. The expediency of the chosen approach to the structuring of the professional foreign language communicative competence has been provided. The recommendations, regarding their practical implementation in the educational process of students of the technical universities.

  15. Methods for design and evaluation of integrated hardware/software systems for concurrent computation

    Science.gov (United States)

    Pratt, Terrence W.

    1987-01-01

    Two testbed programming environments to support the evaluation of a large range of parallel architectures have been implemented under the program Parallel Implementation of Scientific Computing Environments (PISCES). The PISCES 1 environment was applied to two areas of aerospace interest: a sparse matrix iterative equation solver and a dynamic scene analysis system. Currently, the NICE/SPAR testbed system for structural analysis is being modified for parallel operation under PISCES 2; the PISCES 1 applications are also being adapted for PISCES 2. A new formal model of concurrent computation has been developed, based on the mathematical system known as H graph semantics together with a timed Petri net model of the parallel aspects of a system.

  16. BioModels.net Web Services, a free and integrated toolkit for computational modelling software.

    Science.gov (United States)

    Li, Chen; Courtot, Mélanie; Le Novère, Nicolas; Laibe, Camille

    2010-05-01

    Exchanging and sharing scientific results are essential for researchers in the field of computational modelling. BioModels.net defines agreed-upon standards for model curation. A fundamental one, MIRIAM (Minimum Information Requested in the Annotation of Models), standardises the annotation and curation process of quantitative models in biology. To support this standard, MIRIAM Resources maintains a set of standard data types for annotating models, and provides services for manipulating these annotations. Furthermore, BioModels.net creates controlled vocabularies, such as SBO (Systems Biology Ontology) which strictly indexes, defines and links terms used in Systems Biology. Finally, BioModels Database provides a free, centralised, publicly accessible database for storing, searching and retrieving curated and annotated computational models. Each resource provides a web interface to submit, search, retrieve and display its data. In addition, the BioModels.net team provides a set of Web Services which allows the community to programmatically access the resources. A user is then able to perform remote queries, such as retrieving a model and resolving all its MIRIAM Annotations, as well as getting the details about the associated SBO terms. These web services use established standards. Communications rely on SOAP (Simple Object Access Protocol) messages and the available queries are described in a WSDL (Web Services Description Language) file. Several libraries are provided in order to simplify the development of client software. BioModels.net Web Services make one step further for the researchers to simulate and understand the entirety of a biological system, by allowing them to retrieve biological models in their own tool, combine queries in workflows and efficiently analyse models.

  17. Development and Use of a Computer Software for Learning by Observation and Appreciation: A New Way of Planetary Science Education

    Science.gov (United States)

    Mikouchi, A. K.; Mikouchi, T.

    2000-01-01

    We developed a computer software to make users learn about the Moon through their observation and appreciation. We performed a usability test at school, and knew that 7th grade students enjoyed it, making them more interested in the Moon than before.

  18. Forecast and restoration of geomagnetic activity indices by using the software-computational neural network complex

    Science.gov (United States)

    Barkhatov, Nikolay; Revunov, Sergey

    2010-05-01

    It is known that currently used indices of geomagnetic activity to some extent reflect the physical processes occurring in the interaction of the perturbed solar wind with Earth's magnetosphere. Therefore, they are connected to each other and with the parameters of near-Earth space. The establishment of such nonlinear connections is interest. For such purposes when the physical problem is complex or has many parameters the technology of artificial neural networks is applied. Such approach for development of the automated forecast and restoration method of geomagnetic activity indices with the establishment of creative software-computational neural network complex is used. Each neural network experiments were carried out at this complex aims to search for a specific nonlinear relation between the analyzed indices and parameters. At the core of the algorithm work program a complex scheme of the functioning of artificial neural networks (ANN) of different types is contained: back propagation Elman network, feed forward network, fuzzy logic network and Kohonen layer classification network. Tools of the main window of the complex (the application) the settings used by neural networks allow you to change: the number of hidden layers, the number of neurons in the layer, the input and target data, the number of cycles of training. Process and the quality of training the ANN is a dynamic plot of changing training error. Plot of comparison of network response with the test sequence is result of the network training. The last-trained neural network with established nonlinear connection for repeated numerical experiments can be run. At the same time additional training is not executed and the previously trained network as a filter input parameters get through and output parameters with the test event are compared. At statement of the large number of different experiments provided the ability to run the program in a "batch" mode is stipulated. For this purpose the user a

  19. Using RNA Sequence and Structure for the Prediction of Riboswitch Aptamer: A Comprehensive Review of Available Software and Tools

    Directory of Open Access Journals (Sweden)

    Deborah Antunes

    2018-01-01

    Full Text Available RNA molecules are essential players in many fundamental biological processes. Prokaryotes and eukaryotes have distinct RNA classes with specific structural features and functional roles. Computational prediction of protein structures is a research field in which high confidence three-dimensional protein models can be proposed based on the sequence alignment between target and templates. However, to date, only a few approaches have been developed for the computational prediction of RNA structures. Similar to proteins, RNA structures may be altered due to the interaction with various ligands, including proteins, other RNAs, and metabolites. A riboswitch is a molecular mechanism, found in the three kingdoms of life, in which the RNA structure is modified by the binding of a metabolite. It can regulate multiple gene expression mechanisms, such as transcription, translation initiation, and mRNA splicing and processing. Due to their nature, these entities also act on the regulation of gene expression and detection of small metabolites and have the potential to helping in the discovery of new classes of antimicrobial agents. In this review, we describe software and web servers currently available for riboswitch aptamer identification and secondary and tertiary structure prediction, including applications.

  20. Advancing global marine biogeography research with open-source GIS software and cloud-computing

    Science.gov (United States)

    Fujioka, Ei; Vanden Berghe, Edward; Donnelly, Ben; Castillo, Julio; Cleary, Jesse; Holmes, Chris; McKnight, Sean; Halpin, patrick

    2012-01-01

    Across many scientific domains, the ability to aggregate disparate datasets enables more meaningful global analyses. Within marine biology, the Census of Marine Life served as the catalyst for such a global data aggregation effort. Under the Census framework, the Ocean Biogeographic Information System was established to coordinate an unprecedented aggregation of global marine biogeography data. The OBIS data system now contains 31.3 million observations, freely accessible through a geospatial portal. The challenges of storing, querying, disseminating, and mapping a global data collection of this complexity and magnitude are significant. In the face of declining performance and expanding feature requests, a redevelopment of the OBIS data system was undertaken. Following an Open Source philosophy, the OBIS technology stack was rebuilt using PostgreSQL, PostGIS, GeoServer and OpenLayers. This approach has markedly improved the performance and online user experience while maintaining a standards-compliant and interoperable framework. Due to the distributed nature of the project and increasing needs for storage, scalability and deployment flexibility, the entire hardware and software stack was built on a Cloud Computing environment. The flexibility of the platform, combined with the power of the application stack, enabled rapid re-development of the OBIS infrastructure, and ensured complete standards-compliance.

  1. Computer-aided disease prediction system: development of application software with SAS component language.

    Science.gov (United States)

    Chang, Chi-Ming; Kuo, Hsu-Sung; Chang, Shu-Hui; Chang, Hong-Jen; Liou, Der-Ming; Laszlo, Tabar; Chen, Tony Hsiu-Hsi

    2005-04-01

    The intricacy of predictive models associated with prognosis and risk classification of disease often discourages medical personnel who are interested in this field. The aim of this study was therefore to develop a computer-aided disease prediction model underpinning a step-by-step statistics-guided approach including five components: (1) data management; (2) exploratory analysis; (3) type of predictive model; (4) model verification; (5) interactive mode of disease prediction using SAS 8.02 Windows 2000 as a platform. The application of this system was illustrated by using data from the Swedish Two-County Trial on breast cancer screening. The effects of tumour size, node status, and histological grade on breast cancer death using logistic regression model or survival models were predicted. A total of 20 questions were designed to exemplify the usefulness of each component. We also evaluated the system using a controlled randomized trial. Times to finish the above 20 questions were used as endpoint to evaluate the performance of the current system. User satisfaction with the current system such as easy to use, the efficiency of risk prediction, and the reduction of barrier to predictive model was also evaluated. The intervention group not only performed more efficiently than the control group but also satisfied with this application software. The MD-DP-SOS system characterized by menu-driven style, comprehensiveness, accuracy and adequacy assessment, and interactive mode of disease prediction is helpful for medical personnel who are involved in disease prediction.

  2. Monitoring of Computing Resource Use of Active Software Releases in ATLAS

    CERN Document Server

    Limosani, Antonio; The ATLAS collaboration

    2016-01-01

    The LHC is the world's most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the Tier0 at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as "MemoryMonitor", to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and displayed...

  3. Reorganization of water distribution system of Greater Jammu using computer software OPT DIS.

    Science.gov (United States)

    Gupta, Indrani; Bassin, Jagdish; Kelkar, Prakash; Gupta, Apurba

    2003-01-01

    Water supply in old Jammu city is not evenly distributed due to undulating terrain and increase in population density. A study was undertaken to suggest measures for improvement to the distribution system. The paper presents results of analysis carried out using computer package OPT DIS developed at NEERI for optimal design of distribution sytem including system augmentation and reorganization measures needed for the design year 2021. The analysis of zone 1 of the distribution network indicates that seven pipes need alteation to cater to the demand for the year 2006 and fifteen pipes need replacement for the year 2021. New pipelines (13 nos.) are to be provided in parallel to the existing pipes in Upper Reach and 28 pipes in Lower Reach to ensure the service standards of 225 lpcd with minimum residual head of 17 m. Efficient algorithm for augmentations of water supply systems comprising combination of dead-end and loop network are having much wider application in developing countries. The software OPT DIS developed at NEERI was found to be extremely powerful and user friendly for addressing such systems with varying demand pattern, peaking factor, pipe condition, water supply modes and design horizones.

  4. Monitoring of computing resource use of active software releases at ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00219183; The ATLAS collaboration

    2017-01-01

    The LHC is the world’s most powerful particle accelerator, colliding protons at centre of mass energy of 13 TeV. As the energy and frequency of collisions has grown in the search for new physics, so too has demand for computing resources needed for event reconstruction. We will report on the evolution of resource usage in terms of CPU and RAM in key ATLAS offline reconstruction workflows at the TierO at CERN and on the WLCG. Monitoring of workflows is achieved using the ATLAS PerfMon package, which is the standard ATLAS performance monitoring system running inside Athena jobs. Systematic daily monitoring has recently been expanded to include all workflows beginning at Monte Carlo generation through to end-user physics analysis, beyond that of event reconstruction. Moreover, the move to a multiprocessor mode in production jobs has facilitated the use of tools, such as “MemoryMonitor”, to measure the memory shared across processors in jobs. Resource consumption is broken down into software domains and dis...

  5. [Establishment and evaluation of a computer-based software system for detection of initial approximal caries].

    Science.gov (United States)

    Li, Yun; Ye, Wei-ping; Li, Yu-jing; Yu, Yang

    2006-03-01

    To establish and evaluate a computer-based software system (caries diagnosing system, CDS): CDS for better detection of initial approximal caries based on the performance of radiography. A total of 190 approximal surfaces from 95 extracted posterior teeth were examined by bite-wing radiography. After analysing and extracting the approximal lesion's characteristics of X-ray performance, using Matlab and VC language, the CDS was established. Then the proximal surfaces of these teeth were detected automatically by CDS, and scored by naked eyes of 8 experienced dentists for the extent of lesions. The histological appearance of sectioned teeth was used as the gold standard to assess sensitivity, specificity, accuracy, and reproducibility. The sensitivity of CDS was 0.728, Pearson's correlation was 0.722, better than that of optic estimation. The reproducibility of CDS was good too. The CDS can report the lesion site and extension, its diagnostic level is better than that of optic estimation, it is an reasonably sensitive and accurate method for the detection of initial approximal caries in present clinic practice, and a tool of providing objective data.

  6. Flexibility of Bricard's linkages and other structures via resultants and computer algebra.

    Science.gov (United States)

    Lewis, Robert H; Coutsias, Evangelos A

    2016-07-01

    Flexibility of structures is extremely important for chemistry and robotics. Following our earlier work, we study flexibility using polynomial equations, resultants, and a symbolic algorithm of our creation that analyzes the resultant. We show that the software solves a classic arrangement of quadrilaterals in the plane due to Bricard. We fill in several gaps in Bricard's work and discover new flexible arrangements that he was apparently unaware of. This provides strong evidence for the maturity of the software, and is a wonderful example of mathematical discovery via computer assisted experiment.

  7. Benefits of Matching Domain Structure for Planning Software: The Right Stuff

    Science.gov (United States)

    Billman, Dorrit Owen; Arsintescu, Lucica; Feary, Michael S.; Lee, Jessica Chia-Rong; Smith, Asha Halima; Tiwary, Rachna

    2011-01-01

    We investigated the role of domain structure in software design. We compared 2 planning applications, for a Mission Control group (International Space Station), and measured users speed and accuracy. Based on our needs analysis, we identified domain structure and used this to develop new prototype software that matched domain structure better than the legacy system. We took a high-fidelity analog of the natural task into the laboratory and found (large) periformance differences, favoring the system that matched domain structure. Our task design enabled us to attribute better periormance to better match of domain structure. We ran through the whole development cycle, in miniature, from needs analysis through design, development, and evaluation. Doing so enabled inferences not just about the particular systems compared, but also provided evidence for the viability of the design process (particularly needs analysis) that we are exploring.

  8. Coupling Sensing Hardware with Data Interrogation Software for Structural Health Monitoring

    Directory of Open Access Journals (Sweden)

    Charles R. Farrar

    2006-01-01

    Full Text Available The process of implementing a damage detection strategy for aerospace, civil and mechanical engineering infrastructure is referred to as structural health monitoring (SHM. The authors' approach is to address the SHM problem in the context of a statistical pattern recognition paradigm. In this paradigm, the process can be broken down into four parts: (1 Operational Evaluation, (2 Data Acquisition and Cleansing, (3 Feature Extraction and Data Compression, and (4 Statistical Model Development for Feature Discrimination. These processes must be implemented through hardware or software and, in general, some combination of these two approaches will be used. This paper will discuss each portion of the SHM process with particular emphasis on the coupling of a general purpose data interrogation software package for structural health monitoring with a modular wireless sensing and processing platform. More specifically, this paper will address the need to take an integrated hardware/software approach to developing SHM solutions.

  9. Locking Editor A Utility For Protecting Software Exercises In The Computer Laboratory Of AMA University

    Directory of Open Access Journals (Sweden)

    Paul M. Grafilon

    2017-07-01

    Full Text Available The student of AMA University persistence in computing which has the keys to providing their talent needed to fill the computer laboratory in the computing professions. A range of factors can affect a students decision to remain in a computing major or change to another major if ever they feel that computing education is difficult. This has to describe the activities in computer laboratory specifically exercises machine problems and computing case studies interacting different application programs as the basis of their skills and knowledge in programming capability. The nature of those activities addresses by using of IDE as open source in all programming applications which may result of specific intervention such as using the editor to create a source file the code blocks comments and program statements are entered and the file saved. In case there are no corrective actions taken as the editor does not know this is supposed to be a source file as opposed to notes for class. If working in a position-dependent language like Java the developer would have to be very careful about indenting. The file has to be saved with the correct file extension and in a directory where the compiler can find it. Each source file has to be compiled separately if the program has a few source files they all have to be named separately in the compiler. When invoking the compiler it has to be directed to look in the correct directory for the source files and where the output files should be stored. If there is an error in the source file the compiler will output messages and fail to complete. For any errors the developer goes back and edits the source file working from line numbers and compiler messages to fix the problems and these steps continue until all the source files compile without errors. When linking each object file is specified as being part of the build. Again the locations for the object files and executable are given. There may be errors at this point

  10. 15th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    2015-01-01

    This edited book presents scientific results of 15th IEEE/ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2014) held on June 30 – July 2, 2014 in Las Vegas Nevada, USA. The aim of this conference was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas, research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected the 13 outstanding papers from those papers accepted for presentation at the conference.

  11. 14th ACIS/IEEE International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing

    CERN Document Server

    Studies in Computational Intelligence : Volume 492

    2013-01-01

    This edited book presents scientific results of the 14th ACIS/IEEE International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD 2013), held in Honolulu, Hawaii, USA on July 1-3, 2013. The aim of this conference was to bring together scientists, engineers, computer users, and students to share their experiences and exchange new ideas, research results about all aspects (theory, applications and tools) of computer and information science, and to discuss the practical challenges encountered along the way and the solutions adopted to solve them. The conference organizers selected the 17 outstanding papers from those papers accepted for presentation at the conference.  

  12. Development and application of a complex numerical model and software for the computation of dose conversion factors for radon progenies.

    Science.gov (United States)

    Farkas, Árpád; Balásházy, Imre

    2015-04-01

    A more exact determination of dose conversion factors associated with radon progeny inhalation was possible due to the advancements in epidemiological health risk estimates in the last years. The enhancement of computational power and the development of numerical techniques allow computing dose conversion factors with increasing reliability. The objective of this study was to develop an integrated model and software based on a self-developed airway deposition code, an own bronchial dosimetry model and the computational methods accepted by International Commission on Radiological Protection (ICRP) to calculate dose conversion coefficients for different exposure conditions. The model was tested by its application for exposure and breathing conditions characteristic of mines and homes. The dose conversion factors were 8 and 16 mSv WLM(-1) for homes and mines when applying a stochastic deposition model combined with the ICRP dosimetry model (named PM-A model), and 9 and 17 mSv WLM(-1) when applying the same deposition model combined with authors' bronchial dosimetry model and the ICRP bronchiolar and alveolar-interstitial dosimetry model (called PM-B model). User friendly software for the computation of dose conversion factors has also been developed. The software allows one to compute conversion factors for a large range of exposure and breathing parameters and to perform sensitivity analyses. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  13. Pivotal role of computers and software in mass spectrometry - SEQUEST and 20 years of tandem MS database searching.

    Science.gov (United States)

    Yates, John R

    2015-11-01

    Advances in computer technology and software have driven developments in mass spectrometry over the last 50 years. Computers and software have been impactful in three areas: the automation of difficult calculations to aid interpretation, the collection of data and control of instruments, and data interpretation. As the power of computers has grown, so too has the utility and impact on mass spectrometers and their capabilities. This has been particularly evident in the use of tandem mass spectrometry data to search protein and nucleotide sequence databases to identify peptide and protein sequences. This capability has driven the development of many new approaches to study biological systems, including the use of "bottom-up shotgun proteomics" to directly analyze protein mixtures. Graphical Abstract ᅟ.

  14. Theory and Computation

    Data.gov (United States)

    Federal Laboratory Consortium — Flexible computational infrastructure, software tools and theoretical consultation are provided to support modeling and understanding of the structure and properties...

  15. GSEVM v.2: MCMC software to analyse genetically structured environmental variance models

    DEFF Research Database (Denmark)

    Ibáñez-Escriche, N; Garcia, M; Sorensen, D

    2010-01-01

    This note provides a description of software that allows to fit Bayesian genetically structured variance models using Markov chain Monte Carlo (MCMC). The gsevm v.2 program was written in Fortran 90. The DOS and Unix executable programs, the user's guide, and some example files are freely availab...

  16. Using the PSCPCSP computer software for optimization of the composition of industrial alloys and development of new high-temperature nickel-base alloys

    Science.gov (United States)

    Rtishchev, V. V.

    1995-11-01

    Using computer programs some foreign firms have developed new deformable and castable high-temperature nickel-base alloys such as IN, Rene, Mar-M, Udimet, TRW, TM, TMS, TUT, with equiaxial, columnar, and single-crystal structures for manufacturing functional and nozzle blades and other parts of the hot duct of transport and stationary gas-turbine installations (GTI). Similar investigations have been carried out in Russia. This paper presents examples of the use of the PSCPCSP computer software for a quantitative analysis of structural und phase characteristics and properties of industrial alloys with change (within the grade range) in the concentrations of the alloying elements for optimizing the composition of the alloys and regimes of their heat treatment.

  17. Recent developments in the PHENIX software for automated crystallographic structure determination.

    Science.gov (United States)

    Adams, Paul D; Gopal, Kreshna; Grosse-Kunstleve, Ralf W; Hung, Li-Wei; Ioerger, Thomas R; McCoy, Airlie J; Moriarty, Nigel W; Pai, Reetal K; Read, Randy J; Romo, Tod D; Sacchettini, James C; Sauter, Nicholas K; Storoni, Laurent C; Terwilliger, Thomas C

    2004-01-01

    A new software system called PHENIX (Python-based Hierarchical ENvironment for Integrated Xtallography) is being developed for the automation of crystallographic structure solution. This will provide the necessary algorithms to proceed from reduced intensity data to a refined molecular model, and facilitate structure solution for both the novice and expert crystallographer. Here, the features of PHENIXare reviewed and the recent advances in infrastructure and algorithms are briefly described.

  18. A new cone-beam computed tomography system for dental applications with innovative 3D software

    Energy Technology Data Exchange (ETDEWEB)

    Pasini, Alessandro; Bianconi, D.; Rossi, A. [University of Bologna, Department of Physics, Bologna (Italy); NECTAR Imaging srl Imola (Italy); Casali, F. [University of Bologna, Department of Physics, Bologna (Italy); Bontempi, M. [CEFLA Dental Group Imola (Italy)

    2007-02-15

    Objective Cone beam computed tomography (CBCT) is an important image technique for oral surgery (dentoalveolar surgery and dental implantology) and maxillofacial applications. This technique requires compact sized scanners with a relatively low radiation dosage, which makes them suitable for imaging of the craniofacial region. This article aims to present the concept and the preliminary findings obtained with the prototype of a new CBCT scanner with dedicated 3D software, specifically designed for dental imaging. Methods The prototype implements an X-ray tube with a nominal focal spot of 0.5 mm operating at 70-100 kVp and 1-4 mA. The detector is a 6 in. image intensifier coupled with a digital CCD camera. Dosimetry was performed on a RANDO anthropomorphic phantom using Beryllium Oxide thermo-luminescent dosimeters positioned in the phantom in the following site: eyes, thyroid, skin (lips, cheeks, back of the neck), brain, mandible, maxilla and parotid glands. Doses were measured using four configurations, changing the field-of-view (4'' and 6'') and acquisition time (10 and 20 s) of the CBCT. Acquisitions were performed with different parameters regarding the x-ray tube, pixel size and acquisition geometries to evaluate image quality in relation to modulation transfer function (MTF), noise and geometric accuracy. Results The prototype was able to acquire a complete maxillofacial scan in 10-15 s. The CT reconstruction algorithm delivered images that were judged to have high quality, allowing for precise volume rendering. The radiation dose was determined to be 1-1.5 times that of the dose applied during conventional dental panoramic studies. Conclusion Preliminary studies using the CBCT prototype indicate that this device provides images with acceptable diagnostic content at a relatively low radiation dosage, if compared to systems currently available on the market. (orig.)

  19. Computer software as an aid tool for the development of new asphalt products; Programa computacional como ferramenta de auxilio no desenvolvimento de novos produtos asfalticos

    Energy Technology Data Exchange (ETDEWEB)

    Souza, Flavio Vasconcelos de [University of Nebraska (United States). Coll. of Engineering; Soares, Jorge Barbosa [Universidade Federal do Ceara (UFC), Fortaleza, CE (Brazil). Dept. de Engenharia de Transportes. Lab. de Mecanica dos Pavimentos

    2008-04-15

    The use of composite materials in structural applications have been significantly increasing in the last few years, with emphasis to the use in aerospace, biomedical, civil, mechanical and oil engineering. This is due to the fact that the composite materials may offer optimal structural features for specific applications that their components are unable to offer separately. Modified binders and asphalt mixtures are examples of composite materials. Therefore, in order to optimize the performance of these asphalt products, it is necessary to use methodologies capable of retaining the maximum information about its microstructures so as the designer may establish the optimal fractions and distributions of their components for each type of application. A methodology that has been widely used in the scientific community for composite analysis is the so called multi scale modeling. The purpose of this paper is to model the structural behavior of asphalt mixtures using a multistage computer software and to demonstrate how this software can be used to improve quality an to the development of new asphalt products. The software was used in the diametral compression assessment simulation in an asphalt mix and its results are deemed satisfactory when compared to the results obtained experimentally. (author)

  20. Complementing anatomy education using three-dimensional anatomy mobile software applications on tablet computers.

    Science.gov (United States)

    Lewis, T L; Burnett, B; Tunstall, R G; Abrahams, P H

    2014-04-01

    Anatomy has traditionally been a cornerstone of medical education, which has been taught via dissection and didactic lectures. The rising prevalence of mobile tablet technology means medical software applications ("apps") play an increasingly important role in medical education. The applications highlighted in this article will aid anatomical educators to identify which are the most useful in clinical, academic, and educational environments. These have been systematically identified by downloading all applications with keywords related to anatomy and then carrying out qualitative assessment. Novel anatomy applications from developers such as Visible Body, 3D4Medical, and Pocket Anatomy allow students to visualize and manipulate complex anatomical structures using detailed 3D models. They often contain additional content including clinical correlations and a range of media from instructional videos to interactive quiz functions. The strength of tablet technology lies in its ability to consolidate and present anatomical information to the user in the most appropriate manner for their learning style. The only question mark remains over the level of detail and accuracy of these applications. Innovative medical educators who embrace tablet technology will find that anatomy applications serve as a useful learning tool when used in conjunction with existing teaching setups. Copyright © 2013 Wiley Periodicals, Inc.

  1. An Analysis of the RCA Price-S Cost Estimation Model as it Relates to Current Air Force Computer Software Acquisition and Management.

    Science.gov (United States)

    1979-12-01

    that their hardware counter- parts. Recent advances in the state-of-the-art in computer software development techniques hold some promise for reduc... computer software development . The model provides procedures to enable project description with a small set of cost drivers, and to permit calibration

  2. Seeing Red? The Agency of Computer Software in the Production and Management of Students' School Absences

    Science.gov (United States)

    Bodén, Linnea

    2013-01-01

    An increasing number of Swedish municipalities use digital software to manage the registration of students' school absences. The software is regarded as a problem-solving tool to make registration more efficient, but its effects on the educational setting have been largely neglected. Focusing on an event with two students from a class of…

  3. Media Richness, Knowledge sharing and computer progamming by virtual Software teams

    DEFF Research Database (Denmark)

    Williams, Idongesit; Gyamfi, Albert

    2017-01-01

    and the task characteristics involved in software development, the question is if the media richness of the current media platform is effective in enabling knowledge sharing among these virtual teams? An exploratory research was conducted on a software company in Denmark. The data was gathered was analyzed...

  4. The feasibility of using pattern recognition software to measure the influence of computer use on the consultation.

    Science.gov (United States)

    de Lusignan, Simon; Wilson, Eben; Dyble, Andy; Grant, Tristan; Theadom, Alice; Chan, Tom

    2003-11-26

    A key feature of a good general practice consultation is that it is patient-centred. A number of verbal and non-verbal behaviours have been identified as important to establish a good relationship with the patient. However, the use of the computer detracts the doctor's attention away from the patient, compromising these essential elements of the consultation. Current methods to assess the consultation and the influence of the computer on them are time consuming and subjective. If it were possible to measure these quantitatively, it could provide the basis for the first truly objective way of studying the influence of the computer on the consultation. The aim was to assess whether pattern recognition software could be used to measure the influence and pattern of computer use in the consultation. If this proved possible it would provide, for the first time, an objective quantitative measure of computer use and a measure of the attention and responsiveness of the general practitioner towards the patient. A feasibility study using pattern recognition software to analyse a consultation was conducted. A web camera, linked to a data-gathering node was used to film a simulated consultation in a standard office. Members of the research team enacted the role of the doctor and the patient, using pattern recognition software to try and capture patient-centred, non-verbal behaviour. As this was a feasibility study detailed results of the analysis are not presented. It was revealed that pattern recognition software could be used to analyse certain aspects of a simulated consultation. For example, trigger lines enabled the number of times the clinician's hand covered the keyboard to be counted and wrapping recorded the number of times the clinician nodded his head. It was also possible to measure time sequences and whether the movement was brief or lingering. Pattern recognition software enables movements associated with patient-centredness to be recorded. Pattern recognition

  5. The feasibility of using pattern recognition software to measure the influence of computer use on the consultation

    Directory of Open Access Journals (Sweden)

    Theadom Alice

    2003-11-01

    Full Text Available Abstract Background A key feature of a good general practice consultation is that it is patient-centred. A number of verbal and non-verbal behaviours have been identified as important to establish a good relationship with the patient. However, the use of the computer detracts the doctor's attention away from the patient, compromising these essential elements of the consultation. Current methods to assess the consultation and the influence of the computer on them are time consuming and subjective. If it were possible to measure these quantitatively, it could provide the basis for the first truly objective way of studying the influence of the computer on the consultation. The aim was to assess whether pattern recognition software could be used to measure the influence and pattern of computer use in the consultation. If this proved possible it would provide, for the first time, an objective quantitative measure of computer use and a measure of the attention and responsiveness of the general practitioner towards the patient. Methods A feasibility study using pattern recognition software to analyse a consultation was conducted. A web camera, linked to a data-gathering node was used to film a simulated consultation in a standard office. Members of the research team enacted the role of the doctor and the patient, using pattern recognition software to try and capture patient-centred, non-verbal behaviour. As this was a feasibility study detailed results of the analysis are not presented. Results It was revealed that pattern recognition software could be used to analyse certain aspects of a simulated consultation. For example, trigger lines enabled the number of times the clinician's hand covered the keyboard to be counted and wrapping recorded the number of times the clinician nodded his head. It was also possible to measure time sequences and whether the movement was brief or lingering. Conclusion Pattern recognition software enables movements

  6. Validity of two methods to assess computer use: Self-report by questionnaire and computer use software

    NARCIS (Netherlands)

    Douwes, M.; Kraker, H.de; Blatter, B.M.

    2007-01-01

    A long duration of computer use is known to be positively associated with Work Related Upper Extremity Disorders (WRUED). Self-report by questionnaire is commonly used to assess a worker's duration of computer use. The aim of the present study was to assess the validity of self-report and computer

  7. Optimization of Time Structures in Manufacturing Management by using Scheduling Software Lekin

    Directory of Open Access Journals (Sweden)

    Michal Balog

    2016-08-01

    Full Text Available In each manufacturing plant it is one of the basic requirements to produce the largest quantity of products in the shortest time and at the lowest price. In performance of these requirements used are diverse modern methods, technologies and software which ensure the efficiency improvement of manufacturing, costs minimization, production time minimization etc. Presented article is focused on time structures optimization of real manufacturing process of engineering component by using scheduling software Lekin. It is based on theoretical scientific knowledge on which is afterwards found the optimal layout of the manufacturing process in practice. The optimal layout it created by construction and analysis of Gantt charts in scheduling software Lekin with minimum production time condition.

  8. Development of paint area estimation software for ship compartments and structures

    Directory of Open Access Journals (Sweden)

    Doo-Yeoun Cho

    2016-03-01

    Full Text Available The painting process of large ships is an intense manual operation that typically comprises 9–12% of the total shipbuilding cost. Accordingly, shipbuilders need to estimate the required amount of anti-corrosive coatings and painting resources for inventory and cost control. This study aims to develop a software system which enables the shipbuilders to estimate paint area using existing 3D CAD ship structural models. The geometric information of the ships structure are extracted from the existing shipbuilding CAD/CAM system and used to create painting zones. After specifying the painting zones, users can generate the paint faces by clipping structural parts inside each zone. Finally, the paint resources may be obtained from the product of the paint areas and required paint thickness. Implementing the developed software system to real shipbuilders' operations has contributed to improved productivity, faster resource estimation, better accuracy, and fewer coating defects over their conventional manual calculation methods for painting resource estimation.

  9. Introducing the RadBioStat Educational Software: Computer-Assisted Teaching of the Random Nature of Cell Killing

    Directory of Open Access Journals (Sweden)

    Safari A

    2014-06-01

    Full Text Available The interaction of radiation with cells and tissues has a random nature. Therefore, understanding the random nature of cell killing that is determined by Poisson distribution statistics is an essential point in education of radiation biology. RadBioStat is a newly developed educational MATLAB-based software designed for computer-assisted learning of the target theory in radiation biology. Although its potential applications is developing rapidly, currently RadBioStat software can be a useful tool in computerassisted education of radiobiological models such as single target single hit, multiple target single hit and multiple target multiple hit. Scholars’ feedback is valuable to the producers of this software and help them continuously improve this product, add new features and increase its desirability and functionality.

  10. C-CAT: a computer software used to analyze and select Chinese characters and character components for psychological research.

    Science.gov (United States)

    Lo, Ming; Hue, Chih-Wei

    2008-11-01

    The Character-Component Analysis Toolkit (C-CAT) software was designed to assist researchers in constructing experimental materials using traditional Chinese characters. The software package contains two sets of character stocks: one suitable for research using literate adults as subjects and one suitable for research using schoolchildren as subjects. The software can identify linguistic properties, such as the number of strokes contained, the character-component pronunciation regularity, and the arrangement of character components within a character. Moreover, it can compute a character's linguistic frequency, neighborhood size, and phonetic validity with respect to a user-selected character stock. It can also search the selected character stock for similar characters or for character components with user-specified linguistic properties.

  11. Assisting people with multiple disabilities to improve computer typing efficiency through a mouse wheel and on-screen keyboard software.

    Science.gov (United States)

    Shih, Ching-Hsiang

    2014-09-01

    The main purpose of this study was to find out whether three students with multiple disabilities could increase their keyboard typing performance by poking the standard mouse scroll wheel with the newly developed Dynamic Typing Assistive Program (DTAP) and the built-in On-Screen Keyboard (OSK) computer software. The DTAP is a software solution that allows users to complete typing tasks with OSK software easily, quickly, and accurately by poking the mouse wheel. This study was performed according to a multiple baseline design across participants, and the experimental data showed that all of the participants significantly increased their typing efficiency in the intervention phase. Moreover, this improved performance was maintained during the maintenance phase. Practical and developmental implications of the findings were discussed. Copyright © 2014 Elsevier Ltd. All rights reserved.

  12. The Structure of the Computer Industry

    Science.gov (United States)

    1992-03-01

    1980s focuses on the circuitry of silicon chips. Recognizing the economical and technical significance of microchip designs in the expanding computer...technologies of television, compact discs, and personal computers into a product that "can do everything but walk the dog ." (Brandt, August 12, 1991, p... microchip has on hardware. These simple, self-contained, reliable components increase the understanding 124 of designs, allow greater reuse, and simplify

  13. Program design by a multidisciplinary team. [for structural finite element analysis on STAR-100 computer

    Science.gov (United States)

    Voigt, S.

    1975-01-01

    The use of software engineering aids in the design of a structural finite-element analysis computer program for the STAR-100 computer is described. Nested functional diagrams to aid in communication among design team members were used, and a standardized specification format to describe modules designed by various members was adopted. This is a report of current work in which use of the functional diagrams provided continuity and helped resolve some of the problems arising in this long-running part-time project.

  14. Development of computer tablet software for clinical quantification of lateral knee compartment translation during the pivot shift test.

    Science.gov (United States)

    Muller, Bart; Hofbauer, Marcus; Rahnemai-Azar, Amir Ata; Wolf, Megan; Araki, Daisuke; Hoshino, Yuichi; Araujo, Paulo; Debski, Richard E; Irrgang, James J; Fu, Freddie H; Musahl, Volker

    2016-01-01

    The pivot shift test is a commonly used clinical examination by orthopedic surgeons to evaluate knee function following injury. However, the test can only be graded subjectively by the examiner. Therefore, the purpose of this study is to develop software for a computer tablet to quantify anterior translation of the lateral knee compartment during the pivot shift test. Based on the simple image analysis method, software for a computer tablet was developed with the following primary design constraint - the software should be easy to use in a clinical setting and it should not slow down an outpatient visit. Translation of the lateral compartment of the intact knee was 2.0 ± 0.2 mm and for the anterior cruciate ligament-deficient knee was 8.9 ± 0.9 mm (p software provides reliable, objective, and quantitative data on translation of the lateral knee compartment during the pivot shift test and meets the design constraints posed by the clinical setting.

  15. Structural biology computing: Lessons for the biomedical research sciences.

    Science.gov (United States)

    Morin, Andrew; Sliz, Piotr

    2013-11-01

    The field of structural biology, whose aim is to elucidate the molecular and atomic structures of biological macromolecules, has long been at the forefront of biomedical sciences in adopting and developing computational research methods. Operating at the intersection between biophysics, biochemistry, and molecular biology, structural biology's growth into a foundational framework on which many concepts and findings of molecular biology are interpreted1 has depended largely on parallel advancements in computational tools and techniques. Without these computing advances, modern structural biology would likely have remained an exclusive pursuit practiced by few, and not become the widely practiced, foundational field it is today. As other areas of biomedical research increasingly embrace research computing techniques, the successes, failures and lessons of structural biology computing can serve as a useful guide to progress in other biomedically related research fields. Copyright © 2013 Wiley Periodicals, Inc.

  16. Multi-slice computer tomography of left ventricular function with automated analysis software in comparison with conventional ventriculography

    Energy Technology Data Exchange (ETDEWEB)

    Gilard, Martine [Departments of Cardiology, EA 3878, Brest University Hospital, Brest (France)]. E-mail: martine.gilard@chu-brest.fr; Pennec, Pierre-Yves [Departments of Cardiology, EA 3878, Brest University Hospital, Brest (France); Cornily, Jean-Christophe [Departments of Cardiology, EA 3878, Brest University Hospital, Brest (France); Vinsonneau, Ulric [Service de cardiologie, hopital d' instruction des armees Clermont-Tonnerre, 29240 Brest Armees (France); Le Gal, Gregoire [Department of Internal Medicine, Brest University Hospital, Brest (France); Nonent, Michel [Departments of Radiology, Brest University Hospital, Brest (France); Mansourati, Jacques [Departments of Cardiology, EA 3878, Brest University Hospital, Brest (France); Boschat, Jacques [Departments of Cardiology, EA 3878, Brest University Hospital, Brest (France)

    2006-08-15

    Purpose: To evaluate the accuracy of left ventricular volumetric and functional parameters from multi-slice computed tomography using automated analysis software, and to correlate results with those of invasive left ventriculography. Materials and methods: In 145 consecutive patients (mean age, 61 years {+-} 12) known or suspected to have coronary artery disease, a 16-channel multi-slice computed tomography (Philips Mx8000 IDT 16) was performed using a standard technique. Using short-axis secondary multi-slice computed tomography reformations, we determined end-diastolic and end-systolic left ventricular volumes and ejection fraction with a commercially available analysis software capable of automated contour detection. Conventional left ventriculography was performed according to standard techniques within the following 24 h. Bland-Altman analysis was performed to calculate the limits of agreement and systematic errors between multi-slice computed tomography and conventional left ventriculography. Results: As determined by computer tomography, mean end-systolic (53 {+-} 29 mL) left ventricular volumes had an acceptable correlation with conventional ventriculography (67 {+-} 50 mL; r = 0.74; p < 0.001) and mean end-diastolic (119 {+-} 33 mL) left ventricular volumes a poor correlation with conventional ventriculography measurements (154 {+-} 69 mL; r = 0.41). Left ventricular ejection fraction (57% {+-} 14 versus 55% {+-} 14 for conventional ventriculography; r = 0.79) showed a very good correlation (p < 0.001). Bland-Altman analysis showed acceptable limits of agreement ({+-}9.2% for ejection fraction) without systematic errors. Conclusion: The use of a multi-slice computed tomography with an automatic calculation software has a good correlation with conventional ventriculography findings and could accurately assess left ventricular function, but should not be used for ventricular volumes, because of biased estimations.

  17. Approaches and Software for Multi-Objective Optimization of Nuclear Power Structures

    Directory of Open Access Journals (Sweden)

    Andrei A. Andrianov

    2012-04-01

    Full Text Available The work presents the approaches and software developed for multi-objective optimization of nuclear power structures: the modules for energy planning package MESSAGE intended for modeling purposes of developing nuclear power systems and multi-objective evaluation of its effectiveness and an integrated approach based on the method of system dynamics and parameter space investigation, allowing the problem of optimizing a nuclear power system structure in multi-objective formulation to be solved. Some results of implementation of these tools for multi-objective optimization of nuclear power structures are shown.

  18. Intellectual Property Protection for Computer Software: A Comparative Analysis of the United States and Japanese Intellectual Property Regimes

    OpenAIRE

    Arancibia, Rafael

    2003-01-01

    This thesis explores the reform of intellectual property regulation policies with respect to computer software within two advanced industrial nations after 1980. A comparative case analysis of the United States and Japan will provide insight as to how advanced industrial nations have responded to market forces, competing private interests, and international pressure for policy harmonization in the construction and implementation of intellectual property regulation reforms. This study will sho...

  19. Software Construction and Composition Tools for Petascale Computing SCW0837 Progress Report

    Energy Technology Data Exchange (ETDEWEB)

    Epperly, T W; Hochstein, L

    2011-09-12

    The majority of scientific software is distributed as source code. As the number of library dependencies and supported platforms increases, so does the complexity of describing the rules for configuring and building software. In this project, we have performed an empirical study of the magnitude of the build problem by examining the development history of two DOE-funded scientific software projects. We have developed MixDown, a meta-build tool, to simplify the task of building applications that depend on multiple third-party libraries. The results of this research indicate that the effort that scientific programmers spend takes a significant fraction of the total development effort and that the use of MixDown can significantly simplify the task of building software with multiple dependencies.

  20. Method and computer product to increase accuracy of time-based software verification for sensor networks

    Science.gov (United States)

    Foo Kune, Denis [Saint Paul, MN; Mahadevan, Karthikeyan [Mountain View, CA

    2011-01-25

    A recursive verification protocol to reduce the time variance due to delays in the network by putting the subject node at most one hop from the verifier node provides for an efficient manner to test wireless sensor nodes. Since the software signatures are time based, recursive testing will give a much cleaner signal for positive verification of the software running on any one node in the sensor network. In this protocol, the main verifier checks its neighbor, who in turn checks its neighbor, and continuing this process until all nodes have been verified. This ensures minimum time delays for the software verification. Should a node fail the test, the software verification downstream is halted until an alternative path (one not including the failed node) is found. Utilizing techniques well known in the art, having a node tested twice, or not at all, can be avoided.