WorldWideScience

Sample records for asc software quality

  1. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan : ASC software quality engineering practices Version 3.0.

    Energy Technology Data Exchange (ETDEWEB)

    Turgeon, Jennifer L.; Minana, Molly A.; Hackney, Patricia; Pilch, Martin M.

    2009-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in the US Department of Energy/National Nuclear Security Agency (DOE/NNSA) Quality Criteria, Revision 10 (QC-1) as 'conformance to customer requirements and expectations'. This quality plan defines the SNL ASC Program software quality engineering (SQE) practices and provides a mapping of these practices to the SNL Corporate Process Requirement (CPR) 001.3.6; 'Corporate Software Engineering Excellence'. This plan also identifies ASC management's and the software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals. This SNL ASC Software Quality Plan establishes the signatories commitments to improving software products by applying cost-effective SQE practices. This plan enumerates the SQE practices that comprise the development of SNL ASC's software products and explains the project teams opportunities for tailoring and implementing the practices.

  2. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1: ASC software quality engineering practices, Version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  3. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1 : ASC software quality engineering practices version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Minana, Molly A.; Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in DOE/AL Quality Criteria (QC-1) as conformance to customer requirements and expectations. This quality plan defines the ASC program software quality practices and provides mappings of these practices to the SNL Corporate Process Requirements (CPR 1.3.2 and CPR 1.3.6) and the Department of Energy (DOE) document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines (GP&G). This quality plan identifies ASC management and software project teams' responsibilities for cost-effective software engineering quality practices. The SNL ASC Software Quality Plan establishes the signatories commitment to improving software products by applying cost-effective software engineering quality practices. This document explains the project teams opportunities for tailoring and implementing the practices; enumerates the practices that compose the development of SNL ASC's software products; and includes a sample assessment checklist that was developed based upon the practices in this document.

  4. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan part 2 mappings for the ASC software quality engineering practices, version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR001.3.2 and CPR001.3.6 and to a Department of Energy document, ''ASCI Software Quality Engineering: Goals, Principles, and Guidelines''. This document also identifies ASC management and software project teams' responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  5. Structure of AscE and Induced Burial Regions in AscE and AscG upon Formation of the Chaperone Needle-subunit Complex of Type III Secretion System in Aeromonas Hydrophila

    Energy Technology Data Exchange (ETDEWEB)

    Tan, Y.; Yu, H; Leung, K; Sivaraman, J; Mok, Y

    2008-01-01

    In the type III secretion system (T3SS) of Aeromonas hydrophila, the putative needle complex subunit AscF requires both putative chaperones AscE and AscG for formation of a ternary complex to avoid premature assembly. Here we report the crystal structure of AscE at 2.7 A resolution and the mapping of buried regions of AscE, AscG, and AscF in the AscEG and AscEFG complexes using limited protease digestion. The dimeric AscE is comprised of two helix-turn-helix monomers packed in an antiparallel fashion. The N-terminal 13 residues of AscE are buried only upon binding with AscG, but this region is found to be nonessential for the interaction. AscE functions as a monomer and can be coexpressed with AscG or with both AscG and AscF to form soluble complexes. The AscE binding region of AscG in the AscEG complex is identified to be within the N-terminal 61 residues of AscG. The exposed C-terminal substrate-binding region of AscG in the AscEG complex is induced to be buried only upon binding to AscF. However, the N-terminal 52 residues of AscF remain exposed even in the ternary AscEFG complex. On the other hand, the 35-residue C-terminal region of AscF in the complex is resistant to protease digestion in the AscEFG complex. Site-directed mutagenesis showed that two C-terminal hydrophobic residues, Ile83 and Leu84, of AscF are essential for chaperone binding.

  6. Software Correlator for Radioastron Mission

    Science.gov (United States)

    Likhachev, Sergey F.; Kostenko, Vladimir I.; Girin, Igor A.; Andrianov, Andrey S.; Rudnitskiy, Alexey G.; Zharov, Vladimir E.

    In this paper, we discuss the characteristics and operation of Astro Space Center (ASC) software FX correlator that is an important component of space-ground interferometer for Radioastron project. This project performs joint observations of compact radio sources using 10m space radio telescope (SRT) together with ground radio telescopes at 92, 18, 6 and 1.3 cm wavelengths. In this paper, we describe the main features of space-ground VLBI data processing of Radioastron project using ASC correlator. Quality of implemented fringe search procedure provides positive results without significant losses in correlated amplitude. ASC Correlator has a computational power close to real time operation. The correlator has a number of processing modes: “Continuum”, “Spectral Line”, “Pulsars”, “Giant Pulses”,“Coherent”. Special attention is paid to peculiarities of Radioastron space-ground VLBI data processing. The algorithms of time delay and delay rate calculation are also discussed, which is a matter of principle for data correlation of space-ground interferometers. During five years of Radioastron SRT successful operation, ASC correlator showed high potential of satisfying steady growing needs of current and future ground and space VLBI science. Results of ASC software correlator operation are demonstrated.

  7. Pathfinding the Flight Advanced Stirling Convertor Design with the ASC-E3

    Science.gov (United States)

    Wong, Wayne A.; Wilson, Kyle; Smith, Eddie; Collins, Josh

    2012-01-01

    The Advanced Stirling Convertor (ASC) was initially developed by Sunpower, Inc. under contract to NASA Glenn Research Center (GRC) as a technology development project. The ASC technology fulfills NASA's need for high efficiency power convertors for future Radioisotope Power Systems (RPS). Early successful technology demonstrations between 2003 to 2005 eventually led to the expansion of the project including the decision in 2006 to use the ASC technology on the Advanced Stirling Radioisotope Generator (ASRG). Sunpower has delivered 22 ASC convertors of progressively mature designs to date to GRC. Currently, Sunpower with support from GRC, Lockheed Martin Space System Company (LMSSC), and the Department of Energy (DOE) is developing the flight ASC-F in parallel with the ASC-E3 pathfinders. Sunpower will deliver four pairs of ASC-E3 convertors to GRC which will be used for extended operation reliability assessment, independent validation and verification testing, system interaction tests, and to support LMSSC controller verification. The ASC-E3 and -F convertors are being built to the same design and processing documentation and the same product specification. The initial two pairs of ASC-E3 are built before the flight units and will validate design and processing changes prior to implementation on the ASC-F flight convertors. This paper provides a summary on development of the ASC technology and the status of the ASC-E3 build and how they serve the vital pathfinder role ahead of the flight build for ASRG. The ASRG is part of two of the three candidate missions being considered for selection for the Discovery 12 mission.

  8. 77 FR 25168 - Appraisal Subcommittee (ASC); ASC Rules of Operation; Amended

    Science.gov (United States)

    2012-04-27

    ... heads of the Bureau of Consumer Financial Protection and the Federal Housing Finance Agency. The ASC Rules of Operation serve as corporate bylaws outlining the ASC's purpose, functions, authority... Title XI. The ASC Rules of Operation serve as corporate bylaws outlining the ASC's purpose, functions...

  9. Performance Measurement of Advanced Stirling Convertors (ASC-E3)

    Science.gov (United States)

    Oriti, Salvatore M.

    2013-01-01

    NASA Glenn Research Center (GRC) has been supporting development of the Advanced Stirling Radioisotope Generator (ASRG) since 2006. A key element of the ASRG project is providing life, reliability, and performance testing data of the Advanced Stirling Convertor (ASC). The latest version of the ASC (ASC-E3, to represent the third cycle of engineering model test hardware) is of a design identical to the forthcoming flight convertors. For this generation of hardware, a joint Sunpower and GRC effort was initiated to improve and standardize the test support hardware. After this effort was completed, the first pair of ASC-E3 units was produced by Sunpower and then delivered to GRC in December 2012. GRC has begun operation of these units. This process included performance verification, which examined the data from various tests to validate the convertor performance to the product specification. Other tests included detailed performance mapping that encompassed the wide range of operating conditions that will exist during a mission. These convertors were then transferred to Lockheed Martin for controller checkout testing. The results of this latest convertor performance verification activity are summarized here.

  10. 75 FR 80813 - Appraisal Subcommittee (ASC); ASC Rules of Operation; Amended

    Science.gov (United States)

    2010-12-23

    ... name and must rely on the General Services Administration (GSA) for the procurement of office space by... administration, procurement, and other services, consistent with directives of the ASC. In executing this... Conduct Section 11.01. Ethics Provision. The ASC members of the ASC and its officers and employees shall...

  11. ASC-1

    DEFF Research Database (Denmark)

    Jakimoski, Goce; Khajuria, Samant

    2011-01-01

    uses leak extraction from diÆerent AES rounds to compute the key material that is XOR-ed with the message to compute the ciphertext. Unlike LEX, the ASC-1 operates in a CFB fashion to compute an authentication tag over the encrypted message. We argue that ASC-1 is secure by reducingits (IND-CCA , INT......The goal of the modes of operation for authenticated encryption is to achieve faster encryption and message authentication by performing both the encryption and the message authentication in a single pass as opposed to the traditional encrypt-then-mac approach, which requires two passes....... Unfortunately, the use of a block cipher as a building block limits the performance of the authenticated encryption schemes to at most one message block per block cipher evaluation. In this paper, we propose the authenticated encryption scheme ASC-1 (Authenticating Stream Cipher One). Similarly to LEX, ASC-1...

  12. Sandia software guidelines: Software quality planning

    Energy Technology Data Exchange (ETDEWEB)

    1987-08-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standard for Software Quality Assurance Plans, this volume identifies procedures to follow in producing a Software Quality Assurance Plan for an organization or a project, and provides an example project SQA plan. 2 figs., 4 tabs.

  13. Advanced Stirling Convertor (ASC) Technology Maturation

    Science.gov (United States)

    Wong, Wayne A.; Wilson, Scott; Collins, Josh; Wilson, Kyle

    2016-01-01

    The Advanced Stirling Convertor (ASC) development effort was initiated by NASA Glenn Research Center with contractor Sunpower, Inc., to develop high-efficiency thermal-to-electric power conversion technology for NASA Radioisotope Power Systems (RPSs). Early successful performance demonstrations led to the expansion of the project as well as adoption of the technology by the Department of Energy (DOE) and system integration contractor Lockheed Martin Space Systems Company as part of the Advanced Stirling Radioisotope Generator (ASRG) flight project. The ASRG integrates a pair of ASCs to convert the heat from a pair of General Purpose Heat Source (GPHS) modules into electrical power. The expanded NASA ASC effort included development of several generations of ASC prototypes or engineering units to help prepare the ASC technology and Sunpower for flight implementation. Sunpower later had two parallel contracts allowing the last of the NASA engineering units called ASC-E3 to serve as pathfinders for the ASC-F flight convertors being built for DOE. The ASC-E3 convertors utilized the ASC-F flight specifications and were built using the ASC-F design and process documentation. Shortly after the first ASC-F pair achieved initial operation, due to budget constraints, the DOE ASRG flight development contract was terminated. NASA continues to invest in the development of Stirling RPS technology including continued production of the ASC-E3 convertors, seven of which have been delivered with one additional unit in production. Starting in fiscal year 2015, Stirling Convertor Technology Maturation has been reorganized as an element of the RPS Stirling Cycle Technology Development (SCTD) Project and long-term plans for continued Stirling technology advancement are in reformulation. This paper provides a status on the ASC project, an overview of advancements made in the design and production of the ASC at Sunpower, and a summary of acceptance tests, reliability tests, and tactical

  14. Software Quality Assurance in Software Projects: A Study of Pakistan

    OpenAIRE

    Faisal Shafique Butt; Sundus Shaukat; M. Wasif Nisar; Ehsan Ullah Munir; Muhammad Waseem; Kashif Ayyub

    2013-01-01

    Software quality is specific property which tells what kind of standard software should have. In a software project, quality is the key factor of success and decline of software related organization. Many researches have been done regarding software quality. Software related organization follows standards introduced by Capability Maturity Model Integration (CMMI) to achieve good quality software. Quality is divided into three main layers which are Software Quality Assurance (SQA), Software Qu...

  15. Software quality assurance

    CERN Document Server

    Laporte, Claude Y

    2018-01-01

    This book introduces Software Quality Assurance (SQA) and provides an overview of standards used to implement SQA. It defines ways to assess the effectiveness of how one approaches software quality across key industry sectors such as telecommunications, transport, defense, and aerospace. * Includes supplementary website with an instructor's guide and solutions * Applies IEEE software standards as well as the Capability Maturity Model Integration for Development (CMMI) * Illustrates the application of software quality assurance practices through the use of practical examples, quotes from experts, and tips from the authors

  16. Software Quality Assurance Metrics

    Science.gov (United States)

    McRae, Kalindra A.

    2004-01-01

    Software Quality Assurance (SQA) is a planned and systematic set of activities that ensures conformance of software life cycle processes and products conform to requirements, standards and procedures. In software development, software quality means meeting requirements and a degree of excellence and refinement of a project or product. Software Quality is a set of attributes of a software product by which its quality is described and evaluated. The set of attributes includes functionality, reliability, usability, efficiency, maintainability, and portability. Software Metrics help us understand the technical process that is used to develop a product. The process is measured to improve it and the product is measured to increase quality throughout the life cycle of software. Software Metrics are measurements of the quality of software. Software is measured to indicate the quality of the product, to assess the productivity of the people who produce the product, to assess the benefits derived from new software engineering methods and tools, to form a baseline for estimation, and to help justify requests for new tools or additional training. Any part of the software development can be measured. If Software Metrics are implemented in software development, it can save time, money, and allow the organization to identify the caused of defects which have the greatest effect on software development. The summer of 2004, I worked with Cynthia Calhoun and Frank Robinson in the Software Assurance/Risk Management department. My task was to research and collect, compile, and analyze SQA Metrics that have been used in other projects that are not currently being used by the SA team and report them to the Software Assurance team to see if any metrics can be implemented in their software assurance life cycle process.

  17. Examining software complexity and quality for scientific software

    International Nuclear Information System (INIS)

    Kelly, D.; Shepard, T.

    2005-01-01

    Research has not found a simple relationship between software complexity and software quality, and particularly no relationship between commonly used software complexity metrics and the occurrence of software faults. A study with an example of scientific software from the nuclear power industry illustrates the importance of addressing cognitive complexity, the complexity related to understanding the intellectual content of the software. Simple practices such as aptly-named variables contributes more to high quality software than limiting code sizes. This paper examines the research into complexity and quality and reports on a longitudinal study using the example of nuclear software. (author)

  18. Software quality in 1997

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Inc., Burlington, MA (United States)

    1997-11-01

    For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Deployment (QFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels. Since software is on the critical path for many engineered products, and for internal business systems as well, the new approaches are starting to affect global competition and attract widespread international interest. It can be hypothesized that success in mastering software quality will be a key strategy for dominating global software markets in the 21st century.

  19. Proceedings of the Fifth Triennial Software Quality Forum 2000, Software for the Next Millennium, Software Quality Forum

    Energy Technology Data Exchange (ETDEWEB)

    Scientific Software Engineering Group, CIC-12

    2000-04-01

    The Software Quality Forum is a triennial conference held by the Software Quality Assurance Subcommittee for the Department of Energy's Quality Managers. The forum centers on key issues, information, and technology important in software development for the Nuclear Weapons Complex. This year it will be opened up to include local information technology companies and software vendors presenting their solutions, ideas, and lessons learned. The Software Quality Forum 2000 will take on a more hands-on, instructional tone than those previously held. There will be an emphasis on providing information, tools, and resources to assist developers in their goal of producing next generation software.

  20. Software cost/resource modeling: Software quality tradeoff measurement

    Science.gov (United States)

    Lawler, R. W.

    1980-01-01

    A conceptual framework for treating software quality from a total system perspective is developed. Examples are given to show how system quality objectives may be allocated to hardware and software; to illustrate trades among quality factors, both hardware and software, to achieve system performance objectives; and to illustrate the impact of certain design choices on software functionality.

  1. Software quality management

    International Nuclear Information System (INIS)

    Bishop, D.C.; Pymm, P.

    1991-01-01

    As programmable electronic (software-based) systems are increasingly being proposed as design solutions for high integrity applications in nuclear power stations, the need to adopt suitable quality management arrangements is paramount. The authors describe Scottish Nuclear's strategy for software quality management and, using the main on-line monitoring system at Torness Power Station as an example, explain how this strategy is put into practice. Particular attention is given to the topics of software quality planning and change control. (author)

  2. Software component quality evaluation

    Science.gov (United States)

    Clough, A. J.

    1991-01-01

    The paper describes a software inspection process that can be used to evaluate the quality of software components. Quality criteria, process application, independent testing of the process and proposed associated tool support are covered. Early results indicate that this technique is well suited for assessing software component quality in a standardized fashion. With automated machine assistance to facilitate both the evaluation and selection of software components, such a technique should promote effective reuse of software components.

  3. Modernization of software quality assurance

    Science.gov (United States)

    Bhaumik, Gokul

    1988-01-01

    The customers satisfaction depends not only on functional performance, it also depends on the quality characteristics of the software products. An examination of this quality aspect of software products will provide a clear, well defined framework for quality assurance functions, which improve the life-cycle activities of software development. Software developers must be aware of the following aspects which have been expressed by many quality experts: quality cannot be added on; the level of quality built into a program is a function of the quality attributes employed during the development process; and finally, quality must be managed. These concepts have guided our development of the following definition for a Software Quality Assurance function: Software Quality Assurance is a formal, planned approach of actions designed to evaluate the degree of an identifiable set of quality attributes present in all software systems and their products. This paper is an explanation of how this definition was developed and how it is used.

  4. Computer software quality assurance

    International Nuclear Information System (INIS)

    Ives, K.A.

    1986-06-01

    The author defines some criteria for the evaluation of software quality assurance elements for applicability to the regulation of the nuclear industry. The author then analyses a number of software quality assurance (SQA) standards. The major extracted SQA elements are then discussed, and finally specific software quality assurance recommendations are made for the nuclear industry

  5. Advanced Stirling Convertor (ASC) Development for NASA RPS

    Science.gov (United States)

    Wong, Wayne A.; Wilson, Scott; Collins, Josh

    2014-01-01

    Sunpower's Advanced Stirling Convertor (ASC) initiated development under contract to the NASA Glenn Research Center (GRC) and after a series of successful demonstrations, the ASC began transitioning from a technology development project to flight development project. The ASC has very high power conversion efficiency making it attractive for future Radioisotope Power Systems (RPS) in order to make best use of the low plutonium-238 fuel inventory in the U.S. In recent years, the ASC became part of the NASA-Department of Energy Advanced Stirling Radioisotope Generator (ASRG) Integrated Project. Sunpower held two parallel contracts to produce ASC convertors, one with the Department of Energy/Lockheed Martin to produce the ASC-F flight convertors, and one with NASA GRC for the production of ASC-E3 engineering units, the initial units of which served as production pathfinders. The integrated ASC technical team successfully overcame various technical challenges that led to the completion and delivery of the first two pairs of flight-like ASC-E3 by 2013. However, in late Fall 2013, the DOE initiated termination of the Lockheed Martin ASRG flight development contract driven primarily by budget constraints. NASA continues to recognize the importance of high efficiency ASC power conversion for RPS and continues investment in the technology including the continuation of ASC-E3 production at Sunpower and the assembly of the ASRG Engineering Unit #2. This paper provides a summary of ASC technical accomplishments, overview of tests at GRC, plans for continued ASC production at Sunpower, and status of Stirling technology development.

  6. Software quality engineering a practitioner's approach

    CERN Document Server

    Suryn, Witold

    2014-01-01

    Software quality stems from two distinctive, but associated, topics in software engineering: software functional quality and software structural quality. Software Quality Engineering studies the tenets of both of these notions, which focus on the efficiency and value of a design, respectively. The text addresses engineering quality on both the application and system levels with attention to Information Systems and Embedded Systems as well as recent developments. Targeted at graduate engineering students and software quality specialists, the book analyzes the relationship between functionality

  7. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  8. ASC deficiency suppresses proliferation and prevents medulloblastoma incidence.

    Science.gov (United States)

    Knight, E R W; Patel, E Y; Flowers, C A; Crowther, A J; Ting, J P; Miller, C R; Gershon, T R; Deshmukh, M

    2015-01-15

    Apoptosis-associated speck-like protein containing a caspase recruitment domain (ASC) is silenced by promoter methylation in many types of tumors, yet ASC's role in most cancers remains unknown. Here, we show that ASC is highly expressed in a model of medulloblastoma, the most common malignant pediatric brain cancer; ASC is also expressed in human medulloblastomas. Importantly, while ASC deficiency did not affect normal cerebellar development, ASC knockout mice on the Smoothened (ND2:SmoA1) transgenic model of medulloblastoma exhibited a profound reduction in medulloblastoma incidence and a delayed tumor onset. A similar decrease in tumorigenesis with ASC deficiency was also seen in the hGFAP-Cre:SmoM2 mouse model of medulloblastoma. Interestingly, hyperproliferation of the external granule layer (EGL) was comparable at P20 in both wild-type and ASC-deficient SmoA1 mice. However, while the apoptosis and differentiation markers remained unchanged at this age, proliferation makers were decreased, and the EGL was reduced in thickness and area by P60. This reduction in proliferation with ASC deficiency was also seen in isolated SmoA1 cerebellar granule precursor cells in vitro, indicating that the effect of ASC deletion on proliferation was cell autonomous. Interestingly, ASC-deficient SmoA1 cerebella exhibited disrupted expression of genes in the transforming growth factor-β pathway and increased level of nuclear Smad3. Taken together, these results demonstrate an unexpected role for ASC in Sonic hedgehog-driven medulloblastoma tumorigenesis, thus identifying ASC as a promising novel target for antitumor therapy.

  9. Software product quality measurement

    OpenAIRE

    Godliauskas, Eimantas

    2016-01-01

    This paper analyses Ruby product quality measures, suggesting three new measures for Ruby product quality measurement tool Rubocop to measure Ruby product quality characteristics defined in ISO 2502n standard series. This paper consists of four main chapters. The first chapter gives a brief view of software product quality and software product quality measurement. The second chapter analyses object oriented quality measures. The third chapter gives a brief view of the most popular Ruby qualit...

  10. Software quality - how is it achieved?

    International Nuclear Information System (INIS)

    Straker, E.A.

    1986-01-01

    Although software quality can't be quantified, the tools and techniques to achieve high quality are available. As management stresses the need for definable software quality programs from vendors and subcontractors and provides the incentives for these programs, the quality of software will improve. EPRI could provide the leadership in establishing guidelines for a balanced software quality program and through workshops provide training to utility staff and management on the methods for evaluating the characteristics of quality software. With the more complex systems discussed at this workshop and particularly with the trend toward the use of artificial intelligence, the importance of quality software will grow dramatically

  11. Authenticated Secure Container System (ASCS)

    International Nuclear Information System (INIS)

    1991-01-01

    Sandia National Laboratories developed an Authenticated Secure Container System (ASCS) for the International Atomic Energy Agency (IAEA). Agency standard weights and safeguards samples can be stored in the ASCS to provide continuity of knowledge. The ASCS consists of an optically clear cover, a base containing the Authenticated Item Monitoring System (AIMS) transmitter, and the AIMS receiver unit for data collection. The ASCS will provide the Inspector with information concerning the status of the system, during a surveillance period, such as state of health, tampering attempts, and movement of the container system. The secure container is located inside a Glove Box with the receiver located remotely from the Glove Box. AIMS technology uses rf transmission from the secure container to the receiver to provide a record of state of health and tampering. The data is stored in the receiver for analysis by the Inspector during a future inspection visit. 2 refs

  12. Software quality assurance handbook

    Energy Technology Data Exchange (ETDEWEB)

    1990-09-01

    There are two important reasons for Software Quality Assurance (SQA) at Allied-Signal Inc., Kansas City Division (KCD): First, the benefits from SQA make good business sense. Second, the Department of Energy has requested SQA. This handbook is one of the first steps in a plant-wide implementation of Software Quality Assurance at KCD. The handbook has two main purposes. The first is to provide information that you will need to perform software quality assurance activities. The second is to provide a common thread to unify the approach to SQA at KCD. 2 figs.

  13. Software quality testing process analysis

    OpenAIRE

    Mera Paz, Julián

    2016-01-01

    Introduction: This article is the result of reading, review, analysis of books, magazines and articles well known for their scientific and research quality, which have addressed the software quality testing process. The author, based on his work experience in software development companies, teaching and other areas, has compiled and selected information to argue and substantiate the importance of the software quality testing process. Methodology: the existing literature on the software qualit...

  14. Software Quality Perceptions of Stakeholders Involved in the Software Development Process

    Science.gov (United States)

    Padmanabhan, Priya

    2013-01-01

    Software quality is one of the primary determinants of project management success. Stakeholders involved in software development widely agree that quality is important (Barney and Wohlin 2009). However, they may differ on what constitutes software quality, and which of its attributes are more important than others. Although, software quality…

  15. Software Quality Assurance activities of ITER CODAC

    Energy Technology Data Exchange (ETDEWEB)

    Pande, Sopan, E-mail: sopan.pande@iter.org [ITER Organization, Route de Vinon sur Verdon, 13115 St Paul Lez Durance (France); DiMaio, Franck; Kim, Changseung; Kim, Joohan; Klotz, Wolf-Dieter; Makijarvi, Petri; Stepanov, Denis; Wallander, Anders [ITER Organization, Route de Vinon sur Verdon, 13115 St Paul Lez Durance (France)

    2013-10-15

    Highlights: ► Comprehensive and consistent software engineering and quality assurance of CODAC. ► Applicable to all CODAC software projects executed by ITER DAs and contractors. ► Configurable plans for cost effective application of SQA processes. ► CODAC software plans SQAP, SVVP, SDP, and SCMP. ► CODAC software processes based on IEEE 12207-2008. -- Abstract: Software as an integral part of the plant system I and C is crucial in the manufacturing and integrated operation of ITER plant systems. Software Quality Assurance is necessary to ensure the development and maintenance of consistently high quality I and C software throughout the lifetime of ITER. CODAC decided to follow IEEE 12207-2008 software lifecycle processes for Software Engineering and Software Quality Assurance. Software Development Plan, Software Configuration Management Plan and Software Verification and Validation Plan are the mainstay of Software Quality Assurance which is documented in the Software Quality Assurance Plan. This paper describes the Software Quality Assurance (SQA) activities performed by CODAC. The SQA includes development and maintenance of above plans, processes and resources. With the help of Verification and Validation Teams they gather evidence of process conformance and product conformance, and record process data for quality audits and perform process improvements.

  16. Software Quality Assurance activities of ITER CODAC

    International Nuclear Information System (INIS)

    Pande, Sopan; DiMaio, Franck; Kim, Changseung; Kim, Joohan; Klotz, Wolf-Dieter; Makijarvi, Petri; Stepanov, Denis; Wallander, Anders

    2013-01-01

    Highlights: ► Comprehensive and consistent software engineering and quality assurance of CODAC. ► Applicable to all CODAC software projects executed by ITER DAs and contractors. ► Configurable plans for cost effective application of SQA processes. ► CODAC software plans SQAP, SVVP, SDP, and SCMP. ► CODAC software processes based on IEEE 12207-2008. -- Abstract: Software as an integral part of the plant system I and C is crucial in the manufacturing and integrated operation of ITER plant systems. Software Quality Assurance is necessary to ensure the development and maintenance of consistently high quality I and C software throughout the lifetime of ITER. CODAC decided to follow IEEE 12207-2008 software lifecycle processes for Software Engineering and Software Quality Assurance. Software Development Plan, Software Configuration Management Plan and Software Verification and Validation Plan are the mainstay of Software Quality Assurance which is documented in the Software Quality Assurance Plan. This paper describes the Software Quality Assurance (SQA) activities performed by CODAC. The SQA includes development and maintenance of above plans, processes and resources. With the help of Verification and Validation Teams they gather evidence of process conformance and product conformance, and record process data for quality audits and perform process improvements

  17. SWiFT Software Quality Assurance Plan.

    Energy Technology Data Exchange (ETDEWEB)

    Berg, Jonathan Charles [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2016-01-01

    This document describes the software development practice areas and processes which contribute to the ability of SWiFT software developers to provide quality software. These processes are designed to satisfy the requirements set forth by the Sandia Software Quality Assurance Program (SSQAP). APPROVALS SWiFT Software Quality Assurance Plan (SAND2016-0765) approved by: Department Manager SWiFT Site Lead Dave Minster (6121) Date Jonathan White (6121) Date SWiFT Controls Engineer Jonathan Berg (6121) Date CHANGE HISTORY Issue Date Originator(s) Description A 2016/01/27 Jon Berg (06121) Initial release of the SWiFT Software Quality Assurance Plan

  18. Software quality assurance - seven years experience

    International Nuclear Information System (INIS)

    Malsbury, J.A.

    1987-01-01

    This paper describes seven years experience with software quality assurance at PPPL. It covers the early attempts of 1980 and 1981 to establish software quality assurance; the first attempt of 1982 to develop a complete software quality assurance plan; the significant modifications of this plan in 1985; and the future. In addition, the paper describes the role of the Quality Assurance organization within each plan. The scope of this paper is limited to discussions of the software development procedures used in the seven year period. Other software quality topics, such as configuration control or problem identification and resolution, are not discussed

  19. Methods of Software Quality Assurance under a Nuclear Quality Assurance Program

    International Nuclear Information System (INIS)

    Kim, Jang Yeol; Lee, Young Jun; Cha, Kyung Ho; Cheon, Se Woo; Lee, Jang Soo; Kwon, Kee Choon

    2005-01-01

    This paper addresses a substantial implementation of a software quality assurance under a nuclear quality assurance program. The relationship of the responsibility between a top-level nuclear quality assurance program such as ASME/NQA-1 and its lower level software quality assurance is described. Software quality assurance activities and software quality assurance procedures during the software development life cycle are also described

  20. The dynamic of modern software development project management and the software crisis of quality. An integrated system dynamics approach towards software quality improvement

    OpenAIRE

    Nasirikaljahi, Armindokht

    2012-01-01

    The software industry is plagued by cost-overruns, delays, poor customer satisfaction and quality issues that are costing clients and customers world-wide billions of dollars each year. The phenomenon is coined The Software Crisis", and poses a huge challenge for software project management. This thesis addresses one of the core issues of the software crisis, namely software quality. The challenges of software quality are central for understanding the other symptoms of the software crisis. Th...

  1. Software product quality control

    CERN Document Server

    Wagner, Stefan

    2013-01-01

    Quality is not a fixed or universal property of software; it depends on the context and goals of its stakeholders. Hence, when you want to develop a high-quality software system, the first step must be a clear and precise specification of quality. Yet even if you get it right and complete, you can be sure that it will become invalid over time. So the only solution is continuous quality control: the steady and explicit evaluation of a product's properties with respect to its updated quality goals.This book guides you in setting up and running continuous quality control in your environment. Star

  2. Quality in cytopathology: an analysis of the internal quality monitoring indicators of the Instituto Nacional de Câncer

    Directory of Open Access Journals (Sweden)

    Mario Lucio C. Araujo Jr

    2015-04-01

    Full Text Available Introduction: Quality control programs are required to ensure the effectiveness of Pap smear, which still remain a key strategy for control of cervical cancer worldwide. Objective: This study was based on the retrospective and quantitative analysis of the post-analytical phase indicators from the internal quality monitoring (IQM program for cytopathology laboratories, such as: positivity rate, atypical squamous cell (ASC/satisfactory exams ratio, ASC/abnormal test results ratio, ASC/squamous intraepithelial lesions (SIL ratio, percentage of tests compatible with high-grade squamous intraepithelial lesion (HSIL, and total of false negative. Materials and methods: The information was extracted from the computerized system of the Section for Integrated Technology in Cytopathology (Seção Integrada de Tecnologia em Citopatologia [SITEC], a reference institution for cancer cytopathology, from July 2013 to June 2014. From a total of 156,888 Pap smears, 157,454 were considered satisfactory for indicator analysis and 566 were excluded because they were considered unsatisfactory and/or rejected for analysis. The data was organized in tables using Microsoft Excel 2010 software, and categorized as indicators. Results: The averages for the indicators were: 7.2% for positivity rate, 56.9 for ASC/abnormal test ratio, 4.1 for ASC/satisfactory tests ratio, 1.4 for ASC/SIL ratio, 0.6% percentage for tests compatible with HSIL, and 2.1% for false-negative rate. Conclusion: The results show that an Internal Quality Monitoring Program is essencial to ensure quality for cytopathology laboratories, and a randomized review of at least 10% of the negative exams, as recommended by the Brazilian Ministry of Health/Instituto Nacional de Câncer (INCA, since is an effective method, especially for large laboratories.

  3. Ambulatory Surgical Center (ASC) Payment System

    Data.gov (United States)

    U.S. Department of Health & Human Services — This file contains a summary of service utilization by ASC supplier and is derived from 2011 ASC line item level data, updated through June 2012, that is, line items...

  4. Software Quality Assurance Audits Guidebooks

    Science.gov (United States)

    1990-01-01

    The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes that are used in software development. The Software Assurance Guidebook, NASA-GB-A201, issued in September, 1989, provides an overall picture of the NASA concepts and practices in software assurance. Second level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the second level Software Quality Assurance Audits Guidebook that describes software quality assurance audits in a way that is compatible with practices at NASA Centers.

  5. Integrated modeling of software cost and quality

    International Nuclear Information System (INIS)

    Rone, K.Y.; Olson, K.M.

    1994-01-01

    In modeling the cost and quality of software systems, the relationship between cost and quality must be considered. This explicit relationship is dictated by the criticality of the software being developed. The balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and the developers with respect to the processes being employed

  6. A Software Reuse Approach and Its Effect On Software Quality, An Empirical Study for The Software Industry

    OpenAIRE

    Mateen, Ahmed; Kausar, Samina; Sattar, Ahsan Raza

    2017-01-01

    Software reusability has become much interesting because of increased quality and reduce cost. A good process of software reuse leads to enhance the reliability, productivity, quality and the reduction of time and cost. Current reuse techniques focuses on the reuse of software artifact which grounded on anticipated functionality whereas, the non-functional (quality) aspect are also important. So, Software reusability used here to expand quality and productivity of software. It improves overal...

  7. Software metrics: Software quality metrics for distributed systems. [reliability engineering

    Science.gov (United States)

    Post, J. V.

    1981-01-01

    Software quality metrics was extended to cover distributed computer systems. Emphasis is placed on studying embedded computer systems and on viewing them within a system life cycle. The hierarchy of quality factors, criteria, and metrics was maintained. New software quality factors were added, including survivability, expandability, and evolvability.

  8. Assessing Software Quality Through Visualised Cohesion Metrics

    Directory of Open Access Journals (Sweden)

    Timothy Shih

    2001-05-01

    Full Text Available Cohesion is one of the most important factors for software quality as well as maintainability, reliability and reusability. Module cohesion is defined as a quality attribute that seeks for measuring the singleness of the purpose of a module. The module of poor quality can be a serious obstacle to the system quality. In order to design a good software quality, software managers and engineers need to introduce cohesion metrics to measure and produce desirable software. A highly cohesion software is thought to be a desirable constructing. In this paper, we propose a function-oriented cohesion metrics based on the analysis of live variables, live span and the visualization of processing element dependency graph. We give six typical cohesion examples to be measured as our experiments and justification. Therefore, a well-defined, well-normalized, well-visualized and well-experimented cohesion metrics is proposed to indicate and thus enhance software cohesion strength. Furthermore, this cohesion metrics can be easily incorporated with software CASE tool to help software engineers to improve software quality.

  9. SQIMSO: Quality Improvement for Small Software Organizations

    OpenAIRE

    Rabih Zeineddine; Nashat Mansour

    2005-01-01

    Software quality improvement process remains incomplete if it is not initiated and conducted through a wide improvement program that considers process quality improvement, product quality improvement and evolution of human resources. But, small software organizations are not capable of bearing the cost of establishing software process improvement programs. In this work, we propose a new software quality improvement model for small organizations, SQIMSO, based on three ...

  10. The development of the brake system of the BMW 850i including ABS and ASC. Entwicklung des Bremssystems des BMW 850i einschliesslich ABS und ASC

    Energy Technology Data Exchange (ETDEWEB)

    Kraft, H.J.; Leffler, H.

    1990-02-01

    The brake system of the new BMW 850i is described in the following. The brake actuation takes place via an hydraulic brake booster. The disc brakes at front and rear axle are arranged in diagnonal brake split. The 4-channel ABS is fitted as standard equipment. The ABS control unit also incorporates the algorithm for the Automatic Stability Control System ASC or ASC+T. The ASC+T shows improved traction compared with the pure stability system ASC and is standard in the BMW 850i with manual gear box. The automatic gear box equipped BMW 850i are supplied with ASC, the ASC+T is available as an option. Both systems, the ASC and the ASC+T are described with special view on the electronical and hydraulical network in the car. A performance comparison of the ASC-systems completes the description. (orig.).

  11. Software quality: Process or people

    Science.gov (United States)

    Palmer, Regina; Labaugh, Modenna

    1993-01-01

    This paper will present data related to software development processes and personnel involvement from the perspective of software quality assurance. We examine eight years of data collected from six projects. Data collected varied by project but usually included defect and fault density with limited use of code metrics, schedule adherence, and budget growth information. The data are a blend of AFSCP 800-14 and suggested productivity measures in Software Metrics: A Practioner's Guide to Improved Product Development. A software quality assurance database tool, SQUID, was used to store and tabulate the data.

  12. Software quality assurance plans for safety-critical software

    International Nuclear Information System (INIS)

    Liddle, P.

    2006-01-01

    Application software is defined as safety-critical if a fault in the software could prevent the system components from performing their nuclear-safety functions. Therefore, for nuclear-safety systems, the AREVA TELEPERM R XS (TXS) system is classified 1E, as defined in the Inst. of Electrical and Electronics Engineers (IEEE) Std 603-1998. The application software is classified as Software Integrity Level (SIL)-4, as defined in IEEE Std 7-4.3.2-2003. The AREVA NP Inc. Software Program Manual (SPM) describes the measures taken to ensure that the TELEPERM XS application software attains a level of quality commensurate with its importance to safety. The manual also describes how TELEPERM XS correctly performs the required safety functions and conforms to established technical and documentation requirements, conventions, rules, and standards. The program manual covers the requirements definition, detailed design, integration, and test phases for the TELEPERM XS application software, and supporting software created by AREVA NP Inc. The SPM is required for all safety-related TELEPERM XS system applications. The program comprises several basic plans and practices: 1. A Software Quality-Assurance Plan (SQAP) that describes the processes necessary to ensure that the software attains a level of quality commensurate with its importance to safety function. 2. A Software Safety Plan (SSP) that identifies the process to reasonably ensure that safety-critical software performs as intended during all abnormal conditions and events, and does not introduce any new hazards that could jeopardize the health and safety of the public. 3. A Software Verification and Validation (V and V) Plan that describes the method of ensuring the software is in accordance with the requirements. 4. A Software Configuration Management Plan (SCMP) that describes the method of maintaining the software in an identifiable state at all times. 5. A Software Operations and Maintenance Plan (SO and MP) that

  13. Software quality concepts and practice

    CERN Document Server

    Galin, Daniel

    2018-01-01

    The book presents a comprehensive discussion on software quality issues and software quality assurance (SQA) principles and practices, and lays special emphasis on implementing and managing SQA. Primarily designed to serve three audiences; universities and college students, vocational training participants, and software engineers and software development managers, the book may be applicable to all personnel engaged in a software projects Features: * A broad view of SQA. The book delves into SQA issues, going beyond the classic boundaries of custom-made software development to also cover in-house software development, subcontractors, and readymade software. * An up-to-date wide-range coverage of SQA and SQA related topics. Providing comprehensive coverage on multifarious SQA subjects, including topics, hardly explored till in SQA texts. * A systematic presentation of the SQA function and its tasks: establishing the SQA processes, planning, coordinating, follow-up, review and evaluation of SQA processes. * Fo...

  14. Software quality metrics aggregation in industry

    NARCIS (Netherlands)

    Mordal, K.; Anquetil, N.; Laval, J.; Serebrenik, A.; Vasilescu, B.N.; Ducasse, S.

    2013-01-01

    With the growing need for quality assessment of entire software systems in the industry, new issues are emerging. First, because most software quality metrics are defined at the level of individual software components, there is a need for aggregation methods to summarize the results at the system

  15. Report of experiments and evidence for ASC L2 milestone 4467 : demonstration of a legacy application's path to exascale.

    Energy Technology Data Exchange (ETDEWEB)

    Curry, Matthew L.; Ferreira, Kurt Brian; Pedretti, Kevin Thomas Tauke; Leung, Vitus Joseph; Moreland, Kenneth D.; Lofstead, Gerald Fredrick, II; Gentile, Ann C. (Sandia National Laboratories, Livermore, CA); Klundt, Ruth Ann; Ward, H. Lee; Laros, James H., III; Hemmert, Karl Scott; Fabian, Nathan D.; Levenhagen, Michael J.; Barrett, Brian W.; Brightwell, Ronald Brian; Barrett, Richard; Wheeler, Kyle Bruce; Kelly, Suzanne Marie; Rodrigues, Arun F.; Brandt, James M. (Sandia National Laboratories, Livermore, CA); Thompson, David (Sandia National Laboratories, Livermore, CA); VanDyke, John P.; Oldfield, Ron A.; Tucker, Thomas (Open Grid Computing, Inc., Austin, TX); Vaughan, Courtenay Thomas

    2012-03-01

    This report documents thirteen of Sandia's contributions to the Computational Systems and Software Environment (CSSE) within the Advanced Simulation and Computing (ASC) program between fiscal years 2009 and 2012. It describes their impact on ASC applications. Most contributions are implemented in lower software levels allowing for application improvement without source code changes. Improvements are identified in such areas as reduced run time, characterizing power usage, and Input/Output (I/O). Other experiments are more forward looking, demonstrating potential bottlenecks using mini-application versions of the legacy codes and simulating their network activity on Exascale-class hardware. The purpose of this report is to prove that the team has completed milestone 4467-Demonstration of a Legacy Application's Path to Exascale. Cielo is expected to be the last capability system on which existing ASC codes can run without significant modifications. This assertion will be tested to determine where the breaking point is for an existing highly scalable application. The goal is to stretch the performance boundaries of the application by applying recent CSSE RD in areas such as resilience, power, I/O, visualization services, SMARTMAP, lightweight LWKs, virtualization, simulation, and feedback loops. Dedicated system time reservations and/or CCC allocations will be used to quantify the impact of system-level changes to extend the life and performance of the ASC code base. Finally, a simulation of anticipated exascale-class hardware will be performed using SST to supplement the calculations. Determine where the breaking point is for an existing highly scalable application: Chapter 15 presented the CSSE work that sought to identify the breaking point in two ASC legacy applications-Charon and CTH. Their mini-app versions were also employed to complete the task. There is no single breaking point as more than one issue was found with the two codes. The results were

  16. Advanced Stirling Convertor (ASC) Technology Maturation in Preparation for Flight

    Science.gov (United States)

    Wong, Wayne A.; Cornell, Peggy A.

    2012-01-01

    The Advanced Stirling Convertor (ASC) is being developed by an integrated team of Sunpower and National Aeronautics and Space Administration s (NASA s) Glenn Research Center (GRC). The ASC development, funded by NASA s Science Mission Directorate, started as a technology development effort in 2003 and has since evolved through progressive convertor builds and successful testing to demonstrate high conversion efficiency, low mass, and capability to meet long-life Radioisotope Power System (RPS) requirements. The technology has been adopted by the Department of Energy and Lockheed Martin Space Systems Company s Advanced Stirling Radioisotope Generator (ASRG), which has been selected for potential flight demonstration on Discovery 12. This paper provides an overview of the status of ASC development including the most recent ASC-E2 convertors that have been delivered to GRC and an introduction to the ASC-E3 and ASC flight convertors that Sunpower will build next. The paper also describes the technology maturation and support tasks being conducted at GRC to support ASC and ASRG development in the areas of convertor and generator extended operation, high-temperature materials, heater head life assessment, organics, nondestructive inspection, spring fatigue testing, and other reliability verification tasks.

  17. lessons and challenges from software quality assessment

    African Journals Online (AJOL)

    DJFLEX

    www.globaljournalseries.com, Email: info@globaljournalseries.com ... ASSESSMENT: THE CASE OF SPACE SYSTEMS SOFTWARE. ... KEYWORDS: Software, Software Quality ,Quality Standard, Characteristics, ... and communication, etc.

  18. Applying Quality Costs in a Software Development Environment

    Directory of Open Access Journals (Sweden)

    I.P. Hollingsworth

    1999-05-01

    Full Text Available This paper shows how Quality Costs can be a measure of software quality. The relationship between Quality Costs and other software quality metrics is briefly explained, and software development oriented versions of the two principal Quality Cost models are described. Finally the paper discusses the major issues involved in setting up a software Quality Cost programme. The concepts are based on previous research on Quality Costs in manufacturing, coupled with work on software metrics and the work currently being undertaken by the authors in a number of industries.

  19. Software and Systems Test Track Architecture and Concept Definition

    Science.gov (United States)

    2007-05-01

    Light 11.0 11.0 11.0 ASC Flex Free Software Foundation 2.5.31 2.5.31 2.5.31 ASC Fluent Fluent Inc. 6.2.26 6.2.26 6.2.26 6.2.26 ASC FMD ...11 ERDC Fluent Fluent 6.2.16 ERDC Fortran 77/90 compiler Compaq/Cray/SGI 7.4 7.4.3m 7.4.4m 5.6 ERDC FTA Platform 1.1 1.1 1.1 ERDC GAMESS

  20. Advanced Stirling Convertor (ASC-E2) Characterization Testing

    Science.gov (United States)

    Williams, Zachary D.; Oriti, Salvatore M.

    2012-01-01

    Testing has been conducted on Advanced Stirling Convertors (ASCs)-E2 at NASA Glenn Research Center in support of the Advanced Stirling Radioisotope Generator (ASRG) project. This testing has been conducted to understand sensitivities of convertor parameters due to environmental and operational changes during operation of the ASRG in missions to space. This paper summarizes test results and explains the operation of the ASRG during space missions

  1. Pragmatic quality metrics for evolutionary software development models

    Science.gov (United States)

    Royce, Walker

    1990-01-01

    Due to the large number of product, project, and people parameters which impact large custom software development efforts, measurement of software product quality is a complex undertaking. Furthermore, the absolute perspective from which quality is measured (customer satisfaction) is intangible. While we probably can't say what the absolute quality of a software product is, we can determine the relative quality, the adequacy of this quality with respect to pragmatic considerations, and identify good and bad trends during development. While no two software engineers will ever agree on an optimum definition of software quality, they will agree that the most important perspective of software quality is its ease of change. We can call this flexibility, adaptability, or some other vague term, but the critical characteristic of software is that it is soft. The easier the product is to modify, the easier it is to achieve any other software quality perspective. This paper presents objective quality metrics derived from consistent lifecycle perspectives of rework which, when used in concert with an evolutionary development approach, can provide useful insight to produce better quality per unit cost/schedule or to achieve adequate quality more efficiently. The usefulness of these metrics is evaluated by applying them to a large, real world, Ada project.

  2. Experiences with Software Quality Metrics in the EMI middlewate

    OpenAIRE

    Alandes, M; Kenny, E M; Meneses, D; Pucciani, G

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristi...

  3. ASC-PROBA Interface Control Document

    DEFF Research Database (Denmark)

    Betto, Maurizio; Jørgensen, John Leif; Jørgensen, Finn E

    1999-01-01

    of Automation of the Technical University of Denmark. The document is structured as follows. First we present the ASC - heritage, system description, performance - then we address more specifically the environmental properties, like the EMC compatibility and thermal characteristics, and the design...... and reliability issues. Section 6 deals with the testing and the calibration procedures and in section 7 the mechanical and electrical interfaces are given. In section 8 and 9 we address issues like manufacturing, transportation and storage and to conclude we review the ASC specifications against the PROBA...

  4. Software Quality Assessment Tool Based on Meta-Models

    OpenAIRE

    Doneva Rositsa; Gaftandzhieva Silvia; Doneva Zhelyana; Staevsky Nevena

    2015-01-01

    In the software industry it is indisputably essential to control the quality of produced software systems in terms of capabilities for easy maintenance, reuse, portability and others in order to ensure reliability in the software development. But it is also clear that it is very difficult to achieve such a control through a ‘manual’ management of quality.There are a number of approaches for software quality assurance based typically on software quality models (e.g. ISO 9126, McCall’s, Boehm’s...

  5. Experiences with Software Quality Metrics in the EMI Middleware

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project t...

  6. Continuous improvement of software quality

    International Nuclear Information System (INIS)

    Sivertsen, Terje

    1999-04-01

    The present report is the first Halden Work Report delivered from the OECD Halden Reactor Project's research activity on formal methods and software quality. Of particular concern in this activity is to reach a consensus between regulators, licensees and the nuclear industry on questions related to the effective, industrial use of formal methods. The report gives considerable attention to the importance of continuous improvement as a characteristic of a living software quality system, and to the need of providing a basis for software process/product quality integration. In particular, the report discusses these aspects from the perspectives of defect prevention, formal methods, Total Quality Management (TQM), and Bayesian Belief Nets. Another concern is to promote controlled experiments on the use of new methods, techniques, and tools. This is achieved partly by reviewing suggestions on the collection and experimental use of data, and by surveying a number of metrics believed to have some potential for comparison studies (author) (ml)

  7. Requirements engineering: foundation for software quality

    NARCIS (Netherlands)

    Daneva, Maia; Pastor, Oscar

    2016-01-01

    Welcome to the proceedings of the 22nd edition of REFSQ: the International Working Conference on Requirements Engineering – Foundation for Software Quality! Requirements engineering (RE) has been recognized as a critical factor that impacts the quality of software, systems, and services. Since the

  8. A SYSTEMATIC STUDY OF SOFTWARE QUALITY MODELS

    OpenAIRE

    Dr.Vilas. M. Thakare; Ashwin B. Tomar

    2011-01-01

    This paper aims to provide a basis for software quality model research, through a systematic study ofpapers. It identifies nearly seventy software quality research papers from journals and classifies paper asper research topic, estimation approach, study context and data set. The paper results combined withother knowledge provides support for recommendations in future software quality model research, toincrease the area of search for relevant studies, carefully select the papers within a set ...

  9. Automatic sample changer control software for automation of neutron activation analysis process in Malaysian Nuclear Agency

    Science.gov (United States)

    Yussup, N.; Ibrahim, M. M.; Rahman, N. A. A.; Mokhtar, M.; Salim, N. A. A.; Soh@Shaari, S. C.; Azman, A.; Lombigit, L.; Azman, A.; Omar, S. A.

    2018-01-01

    Most of the procedures in neutron activation analysis (NAA) process that has been established in Malaysian Nuclear Agency (Nuclear Malaysia) since 1980s were performed manually. These manual procedures carried out by the NAA laboratory personnel are time consuming and inefficient especially for sample counting and measurement process. The sample needs to be changed and the measurement software needs to be setup for every one hour counting time. Both of these procedures are performed manually for every sample. Hence, an automatic sample changer system (ASC) that consists of hardware and software is developed to automate sample counting process for up to 30 samples consecutively. This paper describes the ASC control software for NAA process which is designed and developed to control the ASC hardware and call GammaVision software for sample measurement. The software is developed by using National Instrument LabVIEW development package.

  10. Software Quality in the Objectory Software Development Process

    NARCIS (Netherlands)

    van den Berg, Klaas; Demeyer, S.; Bosch, J.

    In this paper we discuss how software quality assurance is realized in Rational Objectory. Although much support is given through guidelines and checkpoints, the tool fails to provide clear goals and metrics for quality assessments and it only partially supports the phases in a measurement program.

  11. Experiences with Software Quality Metrics in the EMI middleware

    International Nuclear Information System (INIS)

    Alandes, M; Meneses, D; Pucciani, G; Kenny, E M

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project to extract “code metrics” on the status of the software products and “process metrics” related to the quality of the development and support process such as reaction time to critical bugs, requirements tracking and delays in product releases.

  12. A research review of quality assessment for software

    Science.gov (United States)

    1991-01-01

    Measures were recommended to assess the quality of software submitted to the AdaNet program. The quality factors that are important to software reuse are explored and methods of evaluating those factors are discussed. Quality factors important to software reuse are: correctness, reliability, verifiability, understandability, modifiability, and certifiability. Certifiability is included because the documentation of many factors about a software component such as its efficiency, portability, and development history, constitute a class for factors important to some users, not important at all to other, and impossible for AdaNet to distinguish between a priori. The quality factors may be assessed in different ways. There are a few quantitative measures which have been shown to indicate software quality. However, it is believed that there exists many factors that indicate quality and have not been empirically validated due to their subjective nature. These subjective factors are characterized by the way in which they support the software engineering principles of abstraction, information hiding, modularity, localization, confirmability, uniformity, and completeness.

  13. Evaluating the Effect of Software Quality Characteristics on Health Care Quality Indicators

    Directory of Open Access Journals (Sweden)

    Sakineh Aghazadeh

    2015-07-01

    Full Text Available Introduction: Various types of software are used in health care organizations to manage information and care processes. The quality of software has been an important concern for both health authorities and designers of Health Information Technology. Thus, assessing the effect of software quality on the performance quality of healthcare institutions is essential. Method: The most important health care quality indicators in relation to software quality characteristics are provided via an already performed literature review. ISO 9126 standard model is used for definition and integration of various characteristics of software quality. The effects of software quality characteristics and sub-characteristics on the healthcare indicators are evaluated through expert opinion analyses. A questionnaire comprising of 126 questions of 10-point Likert scale was used to gather opinions of experts in the field of Medical/Health Informatics. The data was analyzed using Structural Equation Modeling. Results: Our findings showed that software Maintainability was rated as the most effective factor on user satisfaction (R2 =0.89 and Functionality as the most important and independent variable affecting patient care quality (R2 =0.98. Efficiency was considered as the most effective factor on workflow (R2 =0.97, and Maintainability as the most important factor that affects healthcare communication (R2 =0.95. Usability and Efficiency were rated as the most effectual factor affecting patient satisfaction (R2 =0.80, 0.81. Reliability, Maintainability, and Efficiency were considered as the main factors affecting care costs (R2 =0.87, 0.74, 0.87. Conclusion: We presented a new model based on ISO standards. The model demonstrates and weighs the relations between software quality characteristics and healthcare quality indicators. The clear relationships between variables and the type of the metrics and measurement methods used in the model make it a reliable method to assess

  14. Achieving software quality by GQM measurement

    NARCIS (Netherlands)

    Solingen, van D.M.; Veenendaal, van E.P.W.M.; Veenendaal, van E.P.W.M.; McMullan, J.

    1997-01-01

    Software development is a discipline with specific management difficulties. Collecting relevant data during development is a way to overcome these difficulties. Such data collection for sqftware development is termed "Software Measurement". Software measurement is a powerful aid to quality

  15. Software metrics to improve software quality in HEP

    International Nuclear Information System (INIS)

    Lancon, E.

    1996-01-01

    The ALEPH reconstruction program maintainability has been evaluated with a case tool implementing an ISO standard methodology based on software metrics. It has been found that the overall quality of the program is good and has shown improvement over the past five years. Frequently modified routines exhibits lower quality; most buys were located in routines with particularly low quality. Implementing from the beginning a quality criteria could have avoided time losses due to bug corrections. (author)

  16. Quality measures and assurance for AI (Artificial Intelligence) software

    Science.gov (United States)

    Rushby, John

    1988-01-01

    This report is concerned with the application of software quality and evaluation measures to AI software and, more broadly, with the question of quality assurance for AI software. Considered are not only the metrics that attempt to measure some aspect of software quality, but also the methodologies and techniques (such as systematic testing) that attempt to improve some dimension of quality, without necessarily quantifying the extent of the improvement. The report is divided into three parts Part 1 reviews existing software quality measures, i.e., those that have been developed for, and applied to, conventional software. Part 2 considers the characteristics of AI software, the applicability and potential utility of measures and techniques identified in the first part, and reviews those few methods developed specifically for AI software. Part 3 presents an assessment and recommendations for the further exploration of this important area.

  17. R D software quality assurance

    Energy Technology Data Exchange (ETDEWEB)

    Hood, F.C.

    1991-10-01

    Research software quality assurance (QA) requirements must be adequate to strengthen development or modification objectives, but flexible enough not to restrict creativity. Application guidelines are needed for the different kinds of research and development (R D) software activities to assure project objectives are achieved.

  18. Criteria and tools for scientific software quality measurements

    Energy Technology Data Exchange (ETDEWEB)

    Tseng, M Y [Previse Inc., Willowdale ON (Canada)

    1995-12-01

    Not all software used in the nuclear industry needs the rigorous formal verification, reliability testing and quality assessment that are being applied to safety critical software. Recently, however, there is increasing recognition that systematic and objective quality assessment of the scientific software used in design and safety analyses of nuclear facilities is necessary to support safety and licensing decisions. Because of the complexity and large size of these programs and the resource constraints faced by the AECB reviewer, it is desirable that appropriate automated tools are used wherever practical. To objectively assess the quality of software, a set of attributes of a software product by which its quality is described and evaluated must be established. These attributes must be relevant to the application domain of software under evaluation. To effectively assess the quality of software, metrics defining quantitative scale and method appropriate to determine the value of attributes need to be applied. To cost-effectively perform the evaluation, use of suitable automated tools is desirable. In this project, criteria for evaluating the quality of scientific software are presented; metrics for which those criteria can be evaluated are identified; a survey of automated tools to measure those metrics was conducted and the most appropriate tool (QA Fortran) was acquired; and the tool usage was demonstrated on three sample programs. (author) 5 refs.

  19. Criteria and tools for scientific software quality measurements

    International Nuclear Information System (INIS)

    Tseng, M.Y.

    1995-12-01

    Not all software used in the nuclear industry needs the rigorous formal verification, reliability testing and quality assessment that are being applied to safety critical software. Recently, however, there is increasing recognition that systematic and objective quality assessment of the scientific software used in design and safety analyses of nuclear facilities is necessary to support safety and licensing decisions. Because of the complexity and large size of these programs and the resource constraints faced by the AECB reviewer, it is desirable that appropriate automated tools are used wherever practical. To objectively assess the quality of software, a set of attributes of a software product by which its quality is described and evaluated must be established. These attributes must be relevant to the application domain of software under evaluation. To effectively assess the quality of software, metrics defining quantitative scale and method appropriate to determine the value of attributes need to be applied. To cost-effectively perform the evaluation, use of suitable automated tools is desirable. In this project, criteria for evaluating the quality of scientific software are presented; metrics for which those criteria can be evaluated are identified; a survey of automated tools to measure those metrics was conducted and the most appropriate tool (QA Fortran) was acquired; and the tool usage was demonstrated on three sample programs. (author) 5 refs

  20. Test Hardware Design for Flightlike Operation of Advanced Stirling Convertors (ASC-E3)

    Science.gov (United States)

    Oriti, Salvatore M.

    2012-01-01

    NASA Glenn Research Center (GRC) has been supporting development of the Advanced Stirling Radioisotope Generator (ASRG) since 2006. A key element of the ASRG project is providing life, reliability, and performance testing of the Advanced Stirling Convertor (ASC). For this purpose, the Thermal Energy Conversion branch at GRC has been conducting extended operation of a multitude of free-piston Stirling convertors. The goal of this effort is to generate long-term performance data (tens of thousands of hours) simultaneously on multiple units to build a life and reliability database. The test hardware for operation of these convertors was designed to permit in-air investigative testing, such as performance mapping over a range of environmental conditions. With this, there was no requirement to accurately emulate the flight hardware. For the upcoming ASC-E3 units, the decision has been made to assemble the convertors into a flight-like configuration. This means the convertors will be arranged in the dual-opposed configuration in a housing that represents the fit, form, and thermal function of the ASRG. The goal of this effort is to enable system level tests that could not be performed with the traditional test hardware at GRC. This offers the opportunity to perform these system-level tests much earlier in the ASRG flight development, as they would normally not be performed until fabrication of the qualification unit. This paper discusses the requirements, process, and results of this flight-like hardware design activity.

  1. Using software metrics and software reliability models to attain acceptable quality software for flight and ground support software for avionic systems

    Science.gov (United States)

    Lawrence, Stella

    1992-01-01

    This paper is concerned with methods of measuring and developing quality software. Reliable flight and ground support software is a highly important factor in the successful operation of the space shuttle program. Reliability is probably the most important of the characteristics inherent in the concept of 'software quality'. It is the probability of failure free operation of a computer program for a specified time and environment.

  2. Immunocytoexpression profile of ProExC in smears interpreted as ASC-US, ASC-H, and cervical intraepithelial lesion

    Directory of Open Access Journals (Sweden)

    Zeynep Tosuner

    2017-01-01

    Full Text Available Aims: We aimed to investigate the immunocytoexpression profiles of a novel assay ProEx C for topoisomerase II alpha (TOP2A and minichromosome maintenance protein 2 (MCM2 in abnormal interpreted smears. Settings and Design: Screening programs with Papanicolaou smear and high risk group human papilloma virus testing have yielded a dramatic reduction of cervical cancer incidence. However, both of these tests have limited specificity for the detection of clinically significant cervical high grade lesions. ProEx C for topoisomerase II alpha (TOP2A and minichromosome maintenance protein 2 (MCM2 has been considered to have tight association with high grade intraepithelial lesions. Materials and Methods: A total number of 54 SurePath cervical cytology specimens of patients previously interpreted as atypical squamous cells-undetermined significance (ASC-US, atypical squamous cells-cannot exclude high grade squamous intraepithelial lesion (ASC-H, low grade squamous intraepithelial lesion (LSIL, and high grade squamous intraepithelial lesion (HSIL were included in our study. Results and Conclusions: ProEx C was positive in 14 of HSILs (100%, 3 of 19 LSILs (16%, 2 of 4 ASC-Hs, and none of ASC-USs (0%. The ProEx C test showed very intense nuclear staining in all cytologically abnormal cells. Further studies are indicated to evaluate the diagnostic role of ProEx C.

  3. Software Quality Assurance for Nuclear Safety Systems

    International Nuclear Information System (INIS)

    Sparkman, D R; Lagdon, R

    2004-01-01

    The US Department of Energy has undertaken an initiative to improve the quality of software used to design and operate their nuclear facilities across the United States. One aspect of this initiative is to revise or create new directives and guides associated with quality practices for the safety software in its nuclear facilities. Safety software includes the safety structures, systems, and components software and firmware, support software and design and analysis software used to ensure the safety of the facility. DOE nuclear facilities are unique when compared to commercial nuclear or other industrial activities in terms of the types and quantities of hazards that must be controlled to protect workers, public and the environment. Because of these differences, DOE must develop an approach to software quality assurance that ensures appropriate risk mitigation by developing a framework of requirements that accomplishes the following goals: (sm b ullet) Ensures the software processes developed to address nuclear safety in design, operation, construction and maintenance of its facilities are safe (sm b ullet) Considers the larger system that uses the software and its impacts (sm b ullet) Ensures that the software failures do not create unsafe conditions Software designers for nuclear systems and processes must reduce risks in software applications by incorporating processes that recognize, detect, and mitigate software failure in safety related systems. It must also ensure that fail safe modes and component testing are incorporated into software design. For nuclear facilities, the consideration of risk is not necessarily sufficient to ensure safety. Systematic evaluation, independent verification and system safety analysis must be considered for software design, implementation, and operation. The software industry primarily uses risk analysis to determine the appropriate level of rigor applied to software practices. This risk-based approach distinguishes safety

  4. Measuring health care process quality with software quality measures.

    Science.gov (United States)

    Yildiz, Ozkan; Demirörs, Onur

    2012-01-01

    Existing quality models focus on some specific diseases, clinics or clinical areas. Although they contain structure, process, or output type measures, there is no model which measures quality of health care processes comprehensively. In addition, due to the not measured overall process quality, hospitals cannot compare quality of processes internally and externally. To bring a solution to above problems, a new model is developed from software quality measures. We have adopted the ISO/IEC 9126 software quality standard for health care processes. Then, JCIAS (Joint Commission International Accreditation Standards for Hospitals) measurable elements were added to model scope for unifying functional requirements. Assessment (diagnosing) process measurement results are provided in this paper. After the application, it was concluded that the model determines weak and strong aspects of the processes, gives a more detailed picture for the process quality, and provides quantifiable information to hospitals to compare their processes with multiple organizations.

  5. Building quality into performance and safety assessment software

    International Nuclear Information System (INIS)

    Wojciechowski, L.C.

    2011-01-01

    Quality assurance is integrated throughout the development lifecycle for performance and safety assessment software. The software used in the performance and safety assessment of a Canadian deep geological repository (DGR) follows the CSA quality assurance standard CSA-N286.7 [1], Quality Assurance of Analytical, Scientific and Design Computer Programs for Nuclear Power Plants. Quality assurance activities in this standard include tasks such as verification and inspection; however, much more is involved in producing a quality software computer program. The types of errors found with different verification methods are described. The integrated quality process ensures that defects are found and corrected as early as possible. (author)

  6. The Effects of Development Team Skill on Software Product Quality

    Science.gov (United States)

    Beaver, Justin M.; Schiavone, Guy A.

    2006-01-01

    This paper provides an analysis of the effect of the skill/experience of the software development team on the quality of the final software product. A method for the assessment of software development team skill and experience is proposed, and was derived from a workforce management tool currently in use by the National Aeronautics and Space Administration. Using data from 26 smallscale software development projects, the team skill measures are correlated to 5 software product quality metrics from the ISO/IEC 9126 Software Engineering Product Quality standard. in the analysis of the results, development team skill is found to be a significant factor in the adequacy of the design and implementation. In addition, the results imply that inexperienced software developers are tasked with responsibilities ill-suited to their skill level, and thus have a significant adverse effect on the quality of the software product. Keywords: software quality, development skill, software metrics

  7. 12: Assuring the quality of critical software

    International Nuclear Information System (INIS)

    Jacky, J.; Kalet, I.

    1987-01-01

    The authors recommend quality assurance procedures for radiation therapy software. Software quality assurance deals with preventing, detecting and repairing programming errors. Error detection difficulties are most severe in computer-based control systems, for example therapy machine control systems, because it may be impossible for users to confirm correct operation while treatments are in progress, or to intervene if things go wrong. Software quality assurance techniques observed in other industries in which public safety is at risk are reviewed. In some of these industries software must be approved or certified before it can be used. Approval is subject to technical reviews and audits by experts other than the program authors. The main obstacles to adoption of these techniques in the radiation therapy field are costs, lack of familiarity and doubts regarding efficacy. 18 refs

  8. Software Quality Control at Belle II

    Science.gov (United States)

    Ritter, M.; Kuhr, T.; Hauth, T.; Gebard, T.; Kristof, M.; Pulvermacher, C.; Belle Software Group, II

    2017-10-01

    Over the last seven years the software stack of the next generation B factory experiment Belle II has grown to over one million lines of C++ and Python code, counting only the part included in offline software releases. There are several thousand commits to the central repository by about 100 individual developers per year. To keep a coherent software stack of high quality that it can be sustained and used efficiently for data acquisition, simulation, reconstruction, and analysis over the lifetime of the Belle II experiment is a challenge. A set of tools is employed to monitor the quality of the software and provide fast feedback to the developers. They are integrated in a machinery that is controlled by a buildbot master and automates the quality checks. The tools include different compilers, cppcheck, the clang static analyzer, valgrind memcheck, doxygen, a geometry overlap checker, a check for missing or extra library links, unit tests, steering file level tests, a sophisticated high-level validation suite, and an issue tracker. The technological development infrastructure is complemented by organizational means to coordinate the development.

  9. The medical software quality deployment method.

    Science.gov (United States)

    Hallberg, N; Timpka, T; Eriksson, H

    1999-03-01

    The objective of this study was to develop a Quality Function Deployment (QFD) model for design of information systems in health-care environments. Consecutive blocked-subject case studies were conducted, based on action research methods. Starting with a QFD model for software development, a model for information system design, the Medical Software Quality Deployment (MSQD) model, was developed. The MSQD model was divided into the pre-study phase, in which the customer categories and their power to influence the design are determined; the data collection phase, in which the voice of customers (VoC) is identified by observations and interviews and quantified by Critical. Incident questionnaires; the need specification phase, where the VoC is specified into ranked customer needs; and the design phase where the customer needs are transformed stepwise to technical requirements and design attributes. QFD showed to be useful for integrating the values of different customer categories in software development for health-care settings. In the later design phases, other quality methods should be used for software implementation and testing.

  10. Software quality assurance and software safety in the Biomed Control System

    International Nuclear Information System (INIS)

    Singh, R.P.; Chu, W.T.; Ludewigt, B.A.; Marks, K.M.; Nyman, M.A.; Renner, T.R.; Stradtner, R.

    1989-01-01

    The Biomed Control System is a hardware/software system used for the delivery, measurement and monitoring of heavy-ion beams in the patient treatment and biology experiment rooms in the Bevalac at the Lawrence Berkeley Laboratory (LBL). This paper describes some aspects of this system including historical background philosophy, configuration management, hardware features that facilitate software testing, software testing procedures, the release of new software quality assurance, safety and operator monitoring. 3 refs

  11. CHALLENGES OF SOFTWARE QUALITY ASSURANCE AND TESTING

    Directory of Open Access Journals (Sweden)

    Md.Shahadat Hossain

    2018-02-01

    Full Text Available Uncertainty exists in Software Company over the world. Software quality problem is leading issue for the software industry. The issue exists from 40 years or 50 years long. The industry is suffering and closing for this issue. In this circumstance, it is important to address and remove its root cause. Otherwise, day by day industry economic loss will increase. I figure out some vital challenges of software quality assurance and testing which have been facing by software industries. The research focused on several small and medium software companies of the world. This paper represents different category of challenges along with responsible stakeholders. This research finds out that testing tools are available testing elements are available testing process has improved but still software has some testing challenges. My research figured out the bottleneck of challenges and explained in this paper. Here software engineers have scope to improve & overcome those challenges. This paper suggests systematic approach to solve the problem.

  12. Software quality assurance plan for GCS

    Science.gov (United States)

    Duncan, Stephen E.; Bailey, Elizabeth K.

    1990-01-01

    The software quality assurance (SQA) function for the Guidance and Control Software (GCS) project which is part of a software error studies research program is described. The SQA plan outlines all of the procedures, controls, and audits to be carried out by the SQA organization to ensure adherence to the policies, procedures, and standards for the GCS project.

  13. Software quality assurance plan for PORFLOW-3D

    International Nuclear Information System (INIS)

    Maheras, S.J.

    1993-03-01

    This plan describes the steps taken by the Idaho National Engineering Laboratory Subsurface and Environmental Modeling Unit personnel to implement software quality assurance procedures for the PORFLOW-3D computer code. PORFLOW-3D was used to conduct radiological performance assessments at the Savannah River Site. software quality assurance procedures for PORFLOW-3D include software acquisition, installation, testing, operation, maintenance, and retirement. Configuration control and quality assurance procedures are also included or referenced in this plan

  14. ASC FY17 Implementation Plan, Rev. 1

    Energy Technology Data Exchange (ETDEWEB)

    Hamilton, P. G. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-06-14

    The Stockpile Stewardship Program (SSP) is an integrated technical program for maintaining the safety, surety, and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational capabilities to support these programs. The Advanced Simulation and Computing Program (ASC) is a cornerstone of the SSP, providing simulation capabilities and computational resources that support annual stockpile assessment and certification, study advanced nuclear weapons design and manufacturing processes, analyze accident scenarios and weapons aging, and provide the tools to enable stockpile Life Extension Programs (LEPs) and the resolution of Significant Finding Investigations (SFIs). This requires a balance of resources, including technical staff, hardware, simulation software, and computer science solutions.

  15. Software for Optimizing Quality Assurance of Other Software

    Science.gov (United States)

    Feather, Martin; Cornford, Steven; Menzies, Tim

    2004-01-01

    Software assurance is the planned and systematic set of activities that ensures that software processes and products conform to requirements, standards, and procedures. Examples of such activities are the following: code inspections, unit tests, design reviews, performance analyses, construction of traceability matrices, etc. In practice, software development projects have only limited resources (e.g., schedule, budget, and availability of personnel) to cover the entire development effort, of which assurance is but a part. Projects must therefore select judiciously from among the possible assurance activities. At its heart, this can be viewed as an optimization problem; namely, to determine the allocation of limited resources (time, money, and personnel) to minimize risk or, alternatively, to minimize the resources needed to reduce risk to an acceptable level. The end result of the work reported here is a means to optimize quality-assurance processes used in developing software.

  16. Software process improvement, quality assurance and measurement

    NARCIS (Netherlands)

    Trienekens, J.J.M.; Kusters, R.J.; Balla, K.; Kontogiannis, K.; Zou, Y.; Di Penta, M.

    2006-01-01

    The aim of this workshop was to present and discuss emergent software quality improvement approaches, with an emphasis on practical applications. Different views on the improvement of software processes, software products, and their interrelations, have been addressed during the workshop.

  17. Software Quality Certification: identifying the real obstacles

    Directory of Open Access Journals (Sweden)

    Megan Baker

    1996-05-01

    Full Text Available A case study of software certification reveals the real difficulty of certifying quality beyond superficial assessment - readers are invited to form their own conclusions. AS 3563 Software Quality Management System is the Australian version of ISO 9001, developed specifically for the software industry. For many Australian software houses, gaining certification with AS 3563 is a priority since certification has become a prerequisite to doing business with government departments and major corporations. However, the process of achieving registration with this standard is a lengthy and resource intensive process, and may have little impact on actual software quality. This case study recounts the experience of the consulting arm of one of Australia's accounting firms in its quest for certification. By using a number of specific management strategies this company was able to successfully implement AS 3563 in less than half the time usually taken to achieve certification - a feat for which its management should be congratulated. However, because the focus of the project was on gaining certification, few internal benefits have been realised despite the successful implementation of the standard.

  18. Evaluating predictive models of software quality

    International Nuclear Information System (INIS)

    Ciaschini, V; Canaparo, M; Ronchieri, E; Salomoni, D

    2014-01-01

    Applications from High Energy Physics scientific community are constantly growing and implemented by a large number of developers. This implies a strong churn on the code and an associated risk of faults, which is unavoidable as long as the software undergoes active evolution. However, the necessities of production systems run counter to this. Stability and predictability are of paramount importance; in addition, a short turn-around time for the defect discovery-correction-deployment cycle is required. A way to reconcile these opposite foci is to use a software quality model to obtain an approximation of the risk before releasing a program to only deliver software with a risk lower than an agreed threshold. In this article we evaluated two quality predictive models to identify the operational risk and the quality of some software products. We applied these models to the development history of several EMI packages with intent to discover the risk factor of each product and compare it with its real history. We attempted to determine if the models reasonably maps reality for the applications under evaluation, and finally we concluded suggesting directions for further studies.

  19. Evaluating Predictive Models of Software Quality

    Science.gov (United States)

    Ciaschini, V.; Canaparo, M.; Ronchieri, E.; Salomoni, D.

    2014-06-01

    Applications from High Energy Physics scientific community are constantly growing and implemented by a large number of developers. This implies a strong churn on the code and an associated risk of faults, which is unavoidable as long as the software undergoes active evolution. However, the necessities of production systems run counter to this. Stability and predictability are of paramount importance; in addition, a short turn-around time for the defect discovery-correction-deployment cycle is required. A way to reconcile these opposite foci is to use a software quality model to obtain an approximation of the risk before releasing a program to only deliver software with a risk lower than an agreed threshold. In this article we evaluated two quality predictive models to identify the operational risk and the quality of some software products. We applied these models to the development history of several EMI packages with intent to discover the risk factor of each product and compare it with its real history. We attempted to determine if the models reasonably maps reality for the applications under evaluation, and finally we concluded suggesting directions for further studies.

  20. A survey of Canadian medical physicists: software quality assurance of in-house software.

    Science.gov (United States)

    Salomons, Greg J; Kelly, Diane

    2015-01-05

    This paper reports on a survey of medical physicists who write and use in-house written software as part of their professional work. The goal of the survey was to assess the extent of in-house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple-choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software-related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines.

  1. Continuous software quality analysis for the ATLAS experiment

    CERN Document Server

    Washbrook, Andrew; The ATLAS collaboration

    2017-01-01

    The software for the ATLAS experiment on the Large Hadron Collider at CERN has evolved over many years to meet the demands of Monte Carlo simulation, particle detector reconstruction and data analysis. At present over 3.8 million lines of C++ code (and close to 6 million total lines of code) are maintained by an active worldwide developer community. In order to run the experiment software efficiently at hundreds of computing centres it is essential to maintain a high level of software quality standards. The methods proposed to improve software quality practices by incorporating checks into the new ATLAS software build infrastructure.

  2. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Fishler, B

    2011-03-18

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  3. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    International Nuclear Information System (INIS)

    Fishler, B.

    2011-01-01

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  4. Revision of ASCE 4

    International Nuclear Information System (INIS)

    Nelson, T.A.; Murray, R.C.; Short, S.A.

    1995-01-01

    The original version of ASCE Standard 4, ''Seismic Analysis of Safety-Related Nuclear Structures'' was published in September 1986. It is ASCE policy to update its standards on a five year interval and the Working Group on Seismic Analysis of Safety Related Nuclear Structures was reconvened to formulate the revisions. The goal in updating the standard is to make sure that it is still relevant and that it incorporates the state of the practice in seismic engineering or, in some cases, where it has been demonstrated that state-of-the-art improvements need to be made to standard practice; new improvements are included. The contents of the new standard cover the same areas as the original version, with some additions. The contents are as follows: Input - response spectra and time histories; modeling of structures; analysis of structures; soil-structure interaction; input for subsystem analysis; special structures - buried pipes and conduits, earth-retaining walls, above-ground vertical tanks, raceways, and base-isolated structures; and an appendix providing seismic probabilistic risk assessment and margin assessment

  5. Study on Software Quality Improvement based on Rayleigh Model and PDCA Model

    OpenAIRE

    Ning Jingfeng; Hu Ming

    2013-01-01

    As the software industry gradually becomes mature, software quality is regarded as the life of a software enterprise. This article discusses how to improve the quality of software, applies Rayleigh model and PDCA model to the software quality management, combines with the defect removal effectiveness index, exerts PDCA model to solve the problem of quality management objectives when using the Rayleigh model in bidirectional quality improvement strategies of software quality management, a...

  6. Software Quality in the Trenches: Two Case Studies of Quality Assurance Practices in Free/Libre and Open Source Software (FLOSS)

    OpenAIRE

    Vestbø, Tor Arne

    2007-01-01

    When proponents of open source software are asked to explain the success of their movement they typically point to the quality of the software produced, which is in turn attributed to the rather unconventional development model of releasing unfinished versions of the software and having users look over the code and report and fix bugs. This thesis investigates the open source quality assurance model from a knowledge management perspective based on the assumption that debugging involves a h...

  7. Software quality for 1997 - what works and what doesn`t?

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Burlington, MA (United States)

    1997-11-01

    This presentation provides a view of software quality for 1997 - what works and what doesn`t. For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Development (WFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels.

  8. Quality assurance of EDP software in practical application

    International Nuclear Information System (INIS)

    Winkler, H.

    1982-01-01

    Alongside the specific properties of the soft software, it is mainly points outside the traditional testing field which apply for the quality assurance thereof. Measures for quality assurance must in particular, start in the development. This presupposes a partial-result orientated development process of software. Due to the high qualitative demands, implements for testing and inspection are of great importance. The problems in software quality assurance are typical for a young technical field where the necessity of which is indisputed, but which has to effect on an empirical-pragmatical level still, due to insufficient scientific foundation. (orig.) [de

  9. Lessons and challenges from software quality assessment: The ...

    African Journals Online (AJOL)

    Lessons and challenges from software quality assessment: The case of space systems software. ... esoteric software technologies and paradigms such as object oriented development, etc. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  10. Quantification frameworks and their application for evaluating the software quality factor using quality characteristic value

    International Nuclear Information System (INIS)

    Kim, C.; Chung, C.H.; Won-Ahn, K.

    2004-01-01

    Many problems, related with safety, frequently occur because Digital Instrument and Control Systems are widely used and expanding their ranges to many applications in Nuclear Power Plants. It, however, does not hold a general position to estimate an appropriate software quality. Thus, the Quality Characteristic Value, a software quality factor through each software life cycle, is suggested in this paper. The Quality Characteristic Value is obtained as following procedure: 1) Scoring Quality Characteristic Factors (especially correctness, traceability, completeness, and understandability) onto Software Verification and Validation results, 2) Deriving the diamond-shaped graphs by setting values of Factors at each axis and lining every points, and lastly 3) Measuring the area of the graph for Quality Characteristic Value. In this paper, this methodology is applied to Plant Control System. In addition, the series of quantification frameworks exhibit some good characteristics in the view of software quality factor. More than any thing else, it is believed that introduced framework may be applicable to regulatory guide, software approval procedures, due to its soundness and simple characteristics. (authors)

  11. Training, Quality Assurance Factors, and Tools Investigation: a Work Report and Suggestions on Software Quality Assurance

    Science.gov (United States)

    Lee, Pen-Nan

    1991-01-01

    Previously, several research tasks have been conducted, some observations were obtained, and several possible suggestions have been contemplated involving software quality assurance engineering at NASA Johnson. These research tasks are briefly described. Also, a brief discussion is given on the role of software quality assurance in software engineering along with some observations and suggestions. A brief discussion on a training program for software quality assurance engineers is provided. A list of assurance factors as well as quality factors are also included. Finally, a process model which can be used for searching and collecting software quality assurance tools is presented.

  12. A framework for business oriented software quality approaches

    NARCIS (Netherlands)

    Trienekens, J.J.M.; Veenendaal, van E.P.W.M.; McMullan, J.

    1997-01-01

    The importance of software for business systems continues to grow. Software products play an increasingly important role in industry and society. The need f~r delivering "quality products" and "quality services" has become as relevant for companies in the field qf sqfiware development as for any

  13. Reducing the risk of failure: Software Quality assurance standards and methods

    International Nuclear Information System (INIS)

    Elphick, J.; Cope, H.

    1992-01-01

    An effective Software Quality Assurance (SQA) program provides an overall approach to software engineering and the establishment of proven methods for the production of reliable software. And, in the authors experience, the overall costs for the software life are diminished with the application of quality methods. In their experience, the issues for implementing quality standards and practices are many. This paper addresses those issues as well as the lessons learned from developing and implementing a number of software quality assurance programs. Their experience includes the development and implementation of their own NRC accepted SQA program and an SQA program for an engineering software developer, as well as developing SQA procedures, standards, and methods for utilities, medical and commercial clients. Some of the issues addressed in this paper are: setting goals and defining quality; applying the software life cycle; addressing organizational issues; providing flexibility and increasing productivity; producing effective documentation; maintaining quality records; Imposing software configuration management; conducting reviews, audits, and controls; verification and validation; and controlling software procurement

  14. Differential splicing of the apoptosis-associated speck like protein containing a caspase recruitment domain (ASC regulates inflammasomes

    Directory of Open Access Journals (Sweden)

    Rojanasakul Yon

    2010-05-01

    Full Text Available Abstract Background The apoptotic speck-like protein containing a caspase recruitment domain (ASC is the essential adaptor protein for caspase 1 mediated interleukin (IL-1β and IL-18 processing in inflammasomes. It bridges activated Nod like receptors (NLRs, which are a family of cytosolic pattern recognition receptors of the innate immune system, with caspase 1, resulting in caspase 1 activation and subsequent processing of caspase 1 substrates. Hence, macrophages from ASC deficient mice are impaired in their ability to produce bioactive IL-1β. Furthermore, we recently showed that ASC translocates from the nucleus to the cytosol in response to inflammatory stimulation in order to promote an inflammasome response, which triggers IL-1β processing and secretion. However, the precise regulation of inflammasomes at the level of ASC is still not completely understood. In this study we identified and characterized three novel ASC isoforms for their ability to function as an inflammasome adaptor. Methods To establish the ability of ASC and ASC isoforms as functional inflammasome adaptors, IL-1β processing and secretion was investigated by ELISA in inflammasome reconstitution assays, stable expression in THP-1 and J774A1 cells, and by restoring the lack of endogenous ASC in mouse RAW264.7 macrophages. In addition, the localization of ASC and ASC isoforms was determined by immunofluorescence staining. Results The three novel ASC isoforms, ASC-b, ASC-c and ASC-d display unique and distinct capabilities to each other and to full length ASC in respect to their function as an inflammasome adaptor, with one of the isoforms even showing an inhibitory effect. Consistently, only the activating isoforms of ASC, ASC and ASC-b, co-localized with NLRP3 and caspase 1, while the inhibitory isoform ASC-c, co-localized only with caspase 1, but not with NLRP3. ASC-d did not co-localize with NLRP3 or with caspase 1 and consistently lacked the ability to function as an

  15. Quality assurance (QA) procedures for software: Evaluation of an ADC quality system

    International Nuclear Information System (INIS)

    Efstathopoulos, E. P.; Benekos, O.; Molfetas, M.; Charou, E.; Kottou, S.; Argentos, S.; Kelekis, N. L.

    2005-01-01

    Image viewing and processing software in computed radiography manipulates image contrast in such a way that all relevant image features are rendered to an appropriate degree of visibility, and improves image quality using enhancement algorithms. The purpose of this study was to investigate procedures for the quality assessment of image processing software for computed radiography with the use of existing test objects and to assess the influence that processing introduces on physical image quality characteristics. Measurements of high-contrast resolution, low-contrast resolution, spatial resolution, grey scale (characteristic curve) and geometric distortion were performed 'subjectively' by three independent observers and 'objectively' by the use of criteria based on pixel intensity values. Results show quality assessment is possible without the need for human evaluators, using digital images. It was discovered that the processing software evaluated in this study was able to improve some aspects of image quality, without introducing geometric distortion. (authors)

  16. Quality Attributes for Mission Flight Software: A Reference for Architects

    Science.gov (United States)

    Wilmot, Jonathan; Fesq, Lorraine; Dvorak, Dan

    2016-01-01

    In the international standards for architecture descriptions in systems and software engineering (ISO/IEC/IEEE 42010), "concern" is a primary concept that often manifests itself in relation to the quality attributes or "ilities" that a system is expected to exhibit - qualities such as reliability, security and modifiability. One of the main uses of an architecture description is to serve as a basis for analyzing how well the architecture achieves its quality attributes, and that requires architects to be as precise as possible about what they mean in claiming, for example, that an architecture supports "modifiability." This paper describes a table, generated by NASA's Software Architecture Review Board, which lists fourteen key quality attributes, identifies different important aspects of each quality attribute and considers each aspect in terms of requirements, rationale, evidence, and tactics to achieve the aspect. This quality attribute table is intended to serve as a guide to software architects, software developers, and software architecture reviewers in the domain of mission-critical real-time embedded systems, such as space mission flight software.

  17. Systems and software quality the next step for industrialisation

    CERN Document Server

    Wieczorek, Martin; Bons, Heinz

    2014-01-01

    Software and systems quality is playing an increasingly important role in the growth of almost all - profit and non-profit - organisations. Quality is vital to the success of enterprises in their markets. Most small trade and repair businesses use software systems in their administration and marketing processes. Every doctor's surgery is managing its patients using software. Banking is no longer conceivable without software. Aircraft, trucks and cars use more and more software to handle their increasingly complex technical systems. Innovation, competition and cost pressure are always present i

  18. Operational excellence (six sigma) philosophy: Application to software quality assurance

    Energy Technology Data Exchange (ETDEWEB)

    Lackner, M.

    1997-11-01

    This report contains viewgraphs on operational excellence philosophy of six sigma applied to software quality assurance. This report outlines the following: goal of six sigma; six sigma tools; manufacturing vs administrative processes; Software quality assurance document inspections; map software quality assurance requirements document; failure mode effects analysis for requirements document; measuring the right response variables; and questions.

  19. The 7 C's for Creating Living Software: A Research Perspective for Quality-Oriented Software Engineering

    NARCIS (Netherlands)

    Aksit, Mehmet

    2004-01-01

    This article proposes the 7 C's for realizing quality-oriented software engineering practices. All the desired qualities of this approach are expressed in short by the term living software. The 7 C's are: Concern-oriented processes, Canonical models, Composable models, Certifiable models,

  20. A software perspective of environmental data quality

    International Nuclear Information System (INIS)

    Banerjee, B.

    1995-01-01

    Because of the large amount of complex data in environmental projects, particularly large decontamination and decommissioning projects, the quality of the data has a profound impact on the success and cost of the mission. In every phase of the life cycle of the project, including regulatory intervention and legal proceedings, maintaining the quality of data and presenting data in a timely and meaningful manner are critical. In this paper, a systemic view of data quality management from a software engineering perspective is presented. A method of evaluation evolves from this view. This method complements the principles of the data quality objective. When graded adequately, the method of evaluation establishes a paradigm for ensuring data quality for new and renewed projects. This paper also demonstrates that incorporating good practices of software engineering into the data management process leads to continuous improvement of data quality

  1. A survey of Canadian medical physicists: software quality assurance of in‐house software

    Science.gov (United States)

    Kelly, Diane

    2015-01-01

    This paper reports on a survey of medical physicists who write and use in‐house written software as part of their professional work. The goal of the survey was to assess the extent of in‐house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple‐choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software‐related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines. PACS number: 87.55.Qr PMID:25679168

  2. Software Quality and Copyright: Issues in Computer-Assisted Instruction.

    Science.gov (United States)

    Helm, Virginia

    The two interconnected problems of educational quality and piracy are described and analyzed in this book, which begins with an investigation of the accusations regarding the alleged dismal quality of educational software. The reality behind accusations of rampant piracy and the effect of piracy on the quality of educational software is examined…

  3. Dimorfismo de ascósporos em Glomerella cingulata f.sp. phaseoli

    OpenAIRE

    Castro, Renata A.; Mendes-Costa, Maria C.; Souza, Elaine A.

    2006-01-01

    O objetivo deste trabalho foi caracterizar linhagens de Glomerella cingulata f. sp. phaseoli quanto ao crescimento micelial, formação de peritécio e dimorfismo de ascósporos. Foram avaliadas quatro linhagens nos meios de cultura de folhas de feijoeiro e meio M3 e em condições com e sem fotoperíodo. O comprimento dos ascósporos da linhagem normal e da mutante foi determinado. Houve diferença significativa quanto ao crescimento micelial e tamanho dos ascósporos. Constatou-se, pela primeira vez,...

  4. On the Role of Software Quality Management in Software Process Improvement

    DEFF Research Database (Denmark)

    Wiedemann Jacobsen, Jan; Kuhrmann, Marco; Münch, Jürgen

    2016-01-01

    Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. SPI addresses many aspects ranging from individual developer skills to entire organizations. It comprises, for instance, the optimization of specific activities...... and a strong focus on custom review, testing, and documentation techniques, whereas a set of five selected improvement measures is almost equally addressed....

  5. USDA-ASCS 1936-1939 Air Photos

    Data.gov (United States)

    Minnesota Department of Natural Resources — This data set is a digital version of aerial photographs taken during the 1936-1939 time frame for the USDA-ASCS. These photos were originally recorded at a scale of...

  6. The AscSimulationMode command

    DEFF Research Database (Denmark)

    Jørgensen, John Leif

    1998-01-01

    Complex instruments like the ASC may be quite difficult to test in closed loops. This problem is augmented by the fact, that no direct stimulation of the CHU is possible that will render the full performance, noise-spectrum and real-timeliness with high fidelity. In order to circumvent this impasse...

  7. Develop Quality Characteristics Based Quality Evaluation Process for Ready to Use Software Products

    OpenAIRE

    Daiju Kato; Hiroshi Ishikawa

    2016-01-01

    The users who use ready to use software product had better get the products’ quality information with classified by some kind of global standard metrics or technique for their evaluation. But many of those software products’ co mpanies don’t provide the quality information because of the products are developed b y their own development and evaluation process. But those users want to get quality i...

  8. Measuring the Software Product Quality during the Software Development Life-Cycle: An ISO Standards Perspective

    OpenAIRE

    Rafa E. Al-Qutaish

    2009-01-01

    Problem statement: The International Organization for Standardization (ISO) published a set of international standards related to the software engineering, such as ISO 12207 and ISO 9126. However, there is a set of cross-references between the two standards. Approach: The ISO 9126 on software product quality and ISO 12207 on software life cycle processes had been analyzed to invistigate the relationships between them and to make a mapping from the ISO 9126 quality characteristics to the ISO 1...

  9. Software metrics: The key to quality software on the NCC project

    Science.gov (United States)

    Burns, Patricia J.

    1993-01-01

    Network Control Center (NCC) Project metrics are captured during the implementation and testing phases of the NCCDS software development lifecycle. The metrics data collection and reporting function has interfaces with all elements of the NCC project. Close collaboration with all project elements has resulted in the development of a defined and repeatable set of metrics processes. The resulting data are used to plan and monitor release activities on a weekly basis. The use of graphical outputs facilitates the interpretation of progress and status. The successful application of metrics throughout the NCC project has been instrumental in the delivery of quality software. The use of metrics on the NCC Project supports the needs of the technical and managerial staff. This paper describes the project, the functions supported by metrics, the data that are collected and reported, how the data are used, and the improvements in the quality of deliverable software since the metrics processes and products have been in use.

  10. Continuous Software Quality analysis for the ATLAS experiment

    CERN Document Server

    Washbrook, Andrew; The ATLAS collaboration

    2017-01-01

    The regular application of software quality tools in large collaborative projects is required to reduce code defects to an acceptable level. If left unchecked the accumulation of defects invariably results in performance degradation at scale and problems with the long-term maintainability of the code. Although software quality tools are effective for identification there remains a non-trivial sociological challenge to resolve defects in a timely manner. This is a ongoing concern for the ATLAS software which has evolved over many years to meet the demands of Monte Carlo simulation, detector reconstruction and data analysis. At present over 3.8 million lines of C++ code (and close to 6 million total lines of code) are maintained by a community of hundreds of developers worldwide. It is therefore preferable to address code defects before they are introduced into a widely used software release. Recent wholesale changes to the ATLAS software infrastructure have provided an ideal opportunity to apply software quali...

  11. Quality Market: Design and Field Study of Prediction Market for Software Quality Control

    Science.gov (United States)

    Krishnamurthy, Janaki

    2010-01-01

    Given the increasing competition in the software industry and the critical consequences of software errors, it has become important for companies to achieve high levels of software quality. While cost reduction and timeliness of projects continue to be important measures, software companies are placing increasing attention on identifying the user…

  12. Absence of the inflammasome adaptor ASC reduces hypoxia-induced pulmonary hypertension in mice.

    Science.gov (United States)

    Cero, Fadila Telarevic; Hillestad, Vigdis; Sjaastad, Ivar; Yndestad, Arne; Aukrust, Pål; Ranheim, Trine; Lunde, Ida Gjervold; Olsen, Maria Belland; Lien, Egil; Zhang, Lili; Haugstad, Solveig Bjærum; Løberg, Else Marit; Christensen, Geir; Larsen, Karl-Otto; Skjønsberg, Ole Henning

    2015-08-15

    Pulmonary hypertension is a serious condition that can lead to premature death. The mechanisms involved are incompletely understood although a role for the immune system has been suggested. Inflammasomes are part of the innate immune system and consist of the effector caspase-1 and a receptor, where nucleotide-binding oligomerization domain-like receptor pyrin domain-containing 3 (NLRP3) is the best characterized and interacts with the adaptor protein apoptosis-associated speck-like protein containing a caspase-recruitment domain (ASC). To investigate whether ASC and NLRP3 inflammasome components are involved in hypoxia-induced pulmonary hypertension, we utilized mice deficient in ASC and NLRP3. Active caspase-1, IL-18, and IL-1β, which are regulated by inflammasomes, were measured in lung homogenates in wild-type (WT), ASC(-/-), and NLRP3(-/-) mice, and phenotypical changes related to pulmonary hypertension and right ventricular remodeling were characterized after hypoxic exposure. Right ventricular systolic pressure (RVSP) of ASC(-/-) mice was significantly lower than in WT exposed to hypoxia (40.8 ± 1.5 mmHg vs. 55.8 ± 2.4 mmHg, P right ventricular remodeling. RVSP of NLRP3(-/-) mice exposed to hypoxia was not significantly altered compared with WT hypoxia. Whereas hypoxia increased protein levels of caspase-1, IL-18, and IL-1β in WT and NLRP3(-/-) mice, this response was absent in ASC(-/-) mice. Moreover, ASC(-/-) mice displayed reduced muscularization and collagen deposition around arteries. In conclusion, hypoxia-induced elevated right ventricular pressure and remodeling were attenuated in mice lacking the inflammasome adaptor protein ASC, suggesting that inflammasomes play an important role in the pathogenesis of pulmonary hypertension. Copyright © 2015 the American Physiological Society.

  13. A multiobjective module-order model for software quality enhancement

    NARCIS (Netherlands)

    Khoshgoftaar, TM; Liu, Y; Seliya, N

    2004-01-01

    The knowledge, prior to system operations, of which program modules are problematic is valuable to a software quality assurance team, especially when there is a constraint on software quality enhancement resources. A cost-effective approach for allocating such resources is to obtain a prediction in

  14. Automating the management of software projects in a developing it ...

    African Journals Online (AJOL)

    Software project management is the control of the transformation of users' ... Model from The American Systems Corporation (ASC) was used for risk management. ... Multi-site development approach facilitates large projects by using simple ...

  15. Quality assurance for software important to safety

    International Nuclear Information System (INIS)

    2000-01-01

    Software applications play an increasingly relevant role in nuclear power plant systems. This is particularly true of software important to safety used in both: calculations for the design, testing and analysis of nuclear reactor systems (design, engineering and analysis software); and monitoring, control and safety functions as an integral part of the reactor systems (monitoring, control and safety system software). Computer technology is advancing at a fast pace, offering new possibilities in nuclear reactor design, construction, commissioning, operation, maintenance and decommissioning. These advances also present new issues which must be considered both by the utility and by the regulatory organization. Refurbishment of ageing instrumentation and control systems in nuclear power plants and new safety related application areas have emerged, with direct (e.g. interfaces with safety systems) and indirect (e.g. operator intervention) implications for safety. Currently, there exist several international standards and guides on quality assurance for software important to safety. However, none of the existing documents provides comprehensive guidance to the developer, manager and regulator during all phases of the software life-cycle. The present publication was developed taking into account the large amount of available documentation, the rapid development of software systems and the need for updated guidance on h ow to do it . It provides information and guidance for defining and implementing quality assurance programmes covering the entire life-cycle of software important to safety. Expected users are managers, performers and assessors from nuclear utilities, regulatory bodies, suppliers and technical support organizations involved with the development and use of software applied in nuclear power plants

  16. Process model for building quality software on internet time ...

    African Journals Online (AJOL)

    The competitive nature of the software construction market and the inherently exhilarating nature of software itself have hinged the success of any software development project on four major pillars: time to market, product quality, innovation and documentation. Unfortunately, however, existing software development models ...

  17. Master Pump Shutdown MPS Software Quality Assurance Plan (SQAP)

    International Nuclear Information System (INIS)

    BEVINS, R.R.

    2000-01-01

    The MPSS Software Quality Assurance (SQAP) describes the tools and strategy used in the development of the MPSS software. The document also describes the methodology for controlling and managing changes to the software

  18. Software Design Improvements. Part 2; Software Quality and the Design and Inspection Process

    Science.gov (United States)

    Lalli, Vincent R.; Packard, Michael H.; Ziemianski, Tom

    1997-01-01

    The application of assurance engineering techniques improves the duration of failure-free performance of software. The totality of features and characteristics of a software product are what determine its ability to satisfy customer needs. Software in safety-critical systems is very important to NASA. We follow the System Safety Working Groups definition for system safety software as: 'The optimization of system safety in the design, development, use and maintenance of software and its integration with safety-critical systems in an operational environment. 'If it is not safe, say so' has become our motto. This paper goes over methods that have been used by NASA to make software design improvements by focusing on software quality and the design and inspection process.

  19. Single-Molecule Fluorescence Reveals the Oligomerization and Folding Steps Driving the Prion-like Behavior of ASC.

    Science.gov (United States)

    Gambin, Yann; Giles, Nichole; O'Carroll, Ailís; Polinkovsky, Mark; Hunter, Dominic; Sierecki, Emma

    2018-02-16

    Single-molecule fluorescence has the unique ability to quantify small oligomers and track conformational changes at a single-protein level. Here we tackled one of the most extreme protein behaviors, found recently in an inflammation pathway. Upon danger recognition in the cytosol, NLRP3 recruits its signaling adaptor, ASC. ASC start polymerizing in a prion-like manner and the system goes in "overdrive" by producing a single micron-sized "speck." By precisely controlling protein expression levels in an in vitro translation system, we could trigger the polymerization of ASC and mimic formation of specks in the absence of inflammasome nucleators. We utilized single-molecule spectroscopy to fully characterize prion-like behaviors and self-propagation of ASC fibrils. We next used our controlled system to monitor the conformational changes of ASC upon fibrillation. Indeed, ASC consists of a PYD and CARD domains, separated by a flexible linker. Individually, both domains have been found to form fibrils, but the structure of the polymers formed by the full-length ASC proteins remains elusive. For the first time, using single-molecule Förster resonance energy transfer, we studied the relative positions of the CARD and PYD domains of full-length ASC. An unexpectedly large conformational change occurred upon ASC fibrillation, suggesting that the CARD domain folds back onto the PYD domain. However, contradicting current models, the "prion-like" conformer was not initiated by binding of ASC to the NLRP3 platform. Rather, using a new method, hybrid between Photon Counting Histogram and Number and Brightness analysis, we showed that NLRP3 forms hexamers with self-binding affinities around 300nM. Overall our data suggest a new mechanism, where NLRP3 can initiate ASC polymerization simply by increasing the local concentration of ASC above a supercritical level. Copyright © 2017. Published by Elsevier Ltd.

  20. Image quality dependence on image processing software in ...

    African Journals Online (AJOL)

    Image quality dependence on image processing software in computed radiography. ... Agfa CR readers use MUSICA software, and an upgrade with significantly different image ... Full Text: EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  1. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    De Guio, Federico

    2013-01-01

    The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release Validation, to constantly validate the functionality and the performance of the reconstruction software; in Monte Carlo productions. The central tool to deliver Data Quality information is a web site for browsing data quality histograms (DQM GUI). In this contribution the usage of the DQM Software in the different environments and its integration in the CMS Reconstruction Software Framework and in all production workflows are presented.

  2. A software quality model and metrics for risk assessment

    Science.gov (United States)

    Hyatt, L.; Rosenberg, L.

    1996-01-01

    A software quality model and its associated attributes are defined and used as the model for the basis for a discussion on risk. Specific quality goals and attributes are selected based on their importance to a software development project and their ability to be quantified. Risks that can be determined by the model's metrics are identified. A core set of metrics relating to the software development process and its products is defined. Measurements for each metric and their usability and applicability are discussed.

  3. Development and application of new quality model for software projects.

    Science.gov (United States)

    Karnavel, K; Dillibabu, R

    2014-01-01

    The IT industry tries to employ a number of models to identify the defects in the construction of software projects. In this paper, we present COQUALMO and its limitations and aim to increase the quality without increasing the cost and time. The computation time, cost, and effort to predict the residual defects are very high; this was overcome by developing an appropriate new quality model named the software testing defect corrective model (STDCM). The STDCM was used to estimate the number of remaining residual defects in the software product; a few assumptions and the detailed steps of the STDCM are highlighted. The application of the STDCM is explored in software projects. The implementation of the model is validated using statistical inference, which shows there is a significant improvement in the quality of the software projects.

  4. Quality Assurance in Software Development: An Exploratory Investigation in Software Project Failures and Business Performance

    Science.gov (United States)

    Ichu, Emmanuel A.

    2010-01-01

    Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…

  5. Analysis of quality raw data of second generation sequencers with Quality Assessment Software.

    Science.gov (United States)

    Ramos, Rommel Tj; Carneiro, Adriana R; Baumbach, Jan; Azevedo, Vasco; Schneider, Maria Pc; Silva, Artur

    2011-04-18

    Second generation technologies have advantages over Sanger; however, they have resulted in new challenges for the genome construction process, especially because of the small size of the reads, despite the high degree of coverage. Independent of the program chosen for the construction process, DNA sequences are superimposed, based on identity, to extend the reads, generating contigs; mismatches indicate a lack of homology and are not included. This process improves our confidence in the sequences that are generated. We developed Quality Assessment Software, with which one can review graphs showing the distribution of quality values from the sequencing reads. This software allow us to adopt more stringent quality standards for sequence data, based on quality-graph analysis and estimated coverage after applying the quality filter, providing acceptable sequence coverage for genome construction from short reads. Quality filtering is a fundamental step in the process of constructing genomes, as it reduces the frequency of incorrect alignments that are caused by measuring errors, which can occur during the construction process due to the size of the reads, provoking misassemblies. Application of quality filters to sequence data, using the software Quality Assessment, along with graphing analyses, provided greater precision in the definition of cutoff parameters, which increased the accuracy of genome construction.

  6. A Study on the Software Quality Assurance Plan

    International Nuclear Information System (INIS)

    Kim, Hyun Tae

    2006-01-01

    On 25 August 2006, the CMMI V1.2 (Capability Maturity Model Integration Version 1.2) was released with the new title CMMI-DEV (CMMI for Development) which supersedes the CMMI-SE/SW (CMMI for systems engineering and software engineering) V1.1. This study discusses the application of IEEE Std 730-2002, IEEE Standard for Software Quality Assurance Plans, for the implementation of the Process and Product Quality Assurance (PPQA) process area (PA) of the CMMI-DEV

  7. Alignment of Software Product Quality Goals in Two Outsourcing Relationships

    OpenAIRE

    Barney, Sebastian; Wohlin, Claes

    2010-01-01

    Background: Issues with software product quality are commonly reported when organisations engage in outsourcing relationships. To address this issue, value-based software engineering literature emphasises the need for all success-critical stakeholder groups to work towards a mutually agreed goal. Aim: This paper presents a case study that aims to compare and contrast the priority two groups place on software product quality — stakeholders internal to the development organisation, and stakehol...

  8. Results of the EC research project REQUEST on software quality and reliability

    International Nuclear Information System (INIS)

    Kersken, M.; Saglietti, F.

    1990-01-01

    GRS work in software safety was mainly concerned with the qualitative assessment of software reliability and quality. As a supplement to these activities the work within the REQUEST project emphasized the quantitative determination of the respective parameters. The three-level quality model COQUAMO serves for the computation - and partly for the prediction - of quality factors during the software life cycle. PERFIDE controls the application of software reliability models during the test phase and in early operational life. Specific attention was paid to the assessment of fault-tolerant diverse software systems. (orig.) [de

  9. SQuAVisiT : A Software Quality Assessment and Visualisation Toolset

    NARCIS (Netherlands)

    Roubtsov, Serguei; Telea, Alexandru; Holten, Danny

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the Software Quality Assessment and Visualisation Toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  10. SQuAVisiT: a software quality assessment and visualisation toolset

    NARCIS (Netherlands)

    Roubtsov, S.; Telea, A.C.; Holten, D.H.R.

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the software quality assessment and visualisation toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  11. Early experiences building a software quality prediction model

    Science.gov (United States)

    Agresti, W. W.; Evanco, W. M.; Smith, M. C.

    1990-01-01

    Early experiences building a software quality prediction model are discussed. The overall research objective is to establish a capability to project a software system's quality from an analysis of its design. The technical approach is to build multivariate models for estimating reliability and maintainability. Data from 21 Ada subsystems were analyzed to test hypotheses about various design structures leading to failure-prone or unmaintainable systems. Current design variables highlight the interconnectivity and visibility of compilation units. Other model variables provide for the effects of reusability and software changes. Reported results are preliminary because additional project data is being obtained and new hypotheses are being developed and tested. Current multivariate regression models are encouraging, explaining 60 to 80 percent of the variation in error density of the subsystems.

  12. Required number of records for ASCE/SEI 7 ground-motion scaling procedure

    Science.gov (United States)

    Reyes, Juan C.; Kalkan, Erol

    2011-01-01

    The procedures and criteria in 2006 IBC (International Council of Building Officials, 2006) and 2007 CBC (International Council of Building Officials, 2007) for the selection and scaling ground-motions for use in nonlinear response history analysis (RHA) of structures are based on ASCE/SEI 7 provisions (ASCE, 2005, 2010). According to ASCE/SEI 7, earthquake records should be selected from events of magnitudes, fault distance, and source mechanisms that comply with the maximum considered earthquake, and then scaled so that the average value of the 5-percent-damped response spectra for the set of scaled records is not less than the design response spectrum over the period range from 0.2Tn to 1.5Tn sec (where Tn is the fundamental vibration period of the structure). If at least seven ground-motions are analyzed, the design values of engineering demand parameters (EDPs) are taken as the average of the EDPs determined from the analyses. If fewer than seven ground-motions are analyzed, the design values of EDPs are taken as the maximum values of the EDPs. ASCE/SEI 7 requires a minimum of three ground-motions. These limits on the number of records in the ASCE/SEI 7 procedure are based on engineering experience, rather than on a comprehensive evaluation. This study statistically examines the required number of records for the ASCE/SEI 7 procedure, such that the scaled records provide accurate, efficient, and consistent estimates of" true" structural responses. Based on elastic-perfectly-plastic and bilinear single-degree-of-freedom systems, the ASCE/SEI 7 scaling procedure is applied to 480 sets of ground-motions. The number of records in these sets varies from three to ten. The records in each set were selected either (i) randomly, (ii) considering their spectral shapes, or (iii) considering their spectral shapes and design spectral-acceleration value, A(Tn). As compared to benchmark (that is, "true") responses from unscaled records using a larger catalog of ground

  13. Proteomics Quality Control: Quality Control Software for MaxQuant Results.

    Science.gov (United States)

    Bielow, Chris; Mastrobuoni, Guido; Kempa, Stefan

    2016-03-04

    Mass spectrometry-based proteomics coupled to liquid chromatography has matured into an automatized, high-throughput technology, producing data on the scale of multiple gigabytes per instrument per day. Consequently, an automated quality control (QC) and quality analysis (QA) capable of detecting measurement bias, verifying consistency, and avoiding propagation of error is paramount for instrument operators and scientists in charge of downstream analysis. We have developed an R-based QC pipeline called Proteomics Quality Control (PTXQC) for bottom-up LC-MS data generated by the MaxQuant software pipeline. PTXQC creates a QC report containing a comprehensive and powerful set of QC metrics, augmented with automated scoring functions. The automated scores are collated to create an overview heatmap at the beginning of the report, giving valuable guidance also to nonspecialists. Our software supports a wide range of experimental designs, including stable isotope labeling by amino acids in cell culture (SILAC), tandem mass tags (TMT), and label-free data. Furthermore, we introduce new metrics to score MaxQuant's Match-between-runs (MBR) functionality by which peptide identifications can be transferred across Raw files based on accurate retention time and m/z. Last but not least, PTXQC is easy to install and use and represents the first QC software capable of processing MaxQuant result tables. PTXQC is freely available at https://github.com/cbielow/PTXQC .

  14. Automated Improvement of Software Architecture Models for Performance and Other Quality Attributes

    OpenAIRE

    Koziolek, Anne

    2013-01-01

    Quality attributes, such as performance or reliability, are crucial for the success of a software system and largely influenced by the software architecture. Their quantitative prediction supports systematic, goal-oriented software design and forms a base of an engineering approach to software design. This thesis proposes a method and tool to automatically improve component-based software architecture (CBA) models based on such quantitative quality prediction techniques.

  15. Structure and assembly of the mouse ASC inflammasome by combined NMR spectroscopy and cryo-electron microscopy

    Science.gov (United States)

    Sborgi, Lorenzo; Ravotti, Francesco; Dandey, Venkata P.; Dick, Mathias S.; Mazur, Adam; Reckel, Sina; Chami, Mohamed; Scherer, Sebastian; Huber, Matthias; Böckmann, Anja; Egelman, Edward H.; Stahlberg, Henning; Broz, Petr; Meier, Beat H.; Hiller, Sebastian

    2015-01-01

    Inflammasomes are multiprotein complexes that control the innate immune response by activating caspase-1, thus promoting the secretion of cytokines in response to invading pathogens and endogenous triggers. Assembly of inflammasomes is induced by activation of a receptor protein. Many inflammasome receptors require the adapter protein ASC [apoptosis-associated speck-like protein containing a caspase-recruitment domain (CARD)], which consists of two domains, the N-terminal pyrin domain (PYD) and the C-terminal CARD. Upon activation, ASC forms large oligomeric filaments, which facilitate procaspase-1 recruitment. Here, we characterize the structure and filament formation of mouse ASC in vitro at atomic resolution. Information from cryo-electron microscopy and solid-state NMR spectroscopy is combined in a single structure calculation to obtain the atomic-resolution structure of the ASC filament. Perturbations of NMR resonances upon filament formation monitor the specific binding interfaces of ASC-PYD association. Importantly, NMR experiments show the rigidity of the PYD forming the core of the filament as well as the high mobility of the CARD relative to this core. The findings are validated by structure-based mutagenesis experiments in cultured macrophages. The 3D structure of the mouse ASC-PYD filament is highly similar to the recently determined human ASC-PYD filament, suggesting evolutionary conservation of ASC-dependent inflammasome mechanisms. PMID:26464513

  16. The software product assurance metrics study: JPL's software systems quality and productivity

    Science.gov (United States)

    Bush, Marilyn W.

    1989-01-01

    The findings are reported of the Jet Propulsion Laboratory (JPL)/Software Product Assurance (SPA) Metrics Study, conducted as part of a larger JPL effort to improve software quality and productivity. Until recently, no comprehensive data had been assembled on how JPL manages and develops software-intensive systems. The first objective was to collect data on software development from as many projects and for as many years as possible. Results from five projects are discussed. These results reflect 15 years of JPL software development, representing over 100 data points (systems and subsystems), over a third of a billion dollars, over four million lines of code and 28,000 person months. Analysis of this data provides a benchmark for gauging the effectiveness of past, present and future software development work. In addition, the study is meant to encourage projects to record existing metrics data and to gather future data. The SPA long term goal is to integrate the collection of historical data and ongoing project data with future project estimations.

  17. Effectiveness of Software Quality Assurance in Offshore Development Enterprises in Sri Lanka

    OpenAIRE

    Malinda G. Sirisena

    2014-01-01

    The aim of this research is to evaluate the effectiveness of software quality assurance approaches of Sri Lankan offshore software development organizations, and to propose a framework which could be used across all offshore software development organizations. An empirical study was conducted using derived framework from popular software quality evaluation models. The research instrument employed was a questionnaire survey among thirty seven Sri Lankan registered offshore software develop...

  18. Round table discussion: Quality control and standardization of nuclear medicine software

    International Nuclear Information System (INIS)

    Anon.

    1988-01-01

    In summary the round table came to the following important conclusions: Nuclear medicine software systems need better documentation, especially regarding details of algorithms and limitations, and user friendliness could be considerably improved. Quality control of software is an integral part of quality assurance in nuclear medicine and should be performed at all levels of the software. Quality control of applications software should preferably be performed with assistance of generally accepted software phantoms. A basic form of standardization was welcomed and partly regarded as essential by all participants. Some areas such as patient study files could be standardized in the near future, whereas other areas such as the standardization of clinical applications programs or acquisition protocols still present major difficulties. An international cooperation in the field of standardization of software and other topics has already been started on the European level and should be continued and supported. (orig.)

  19. Quality assurance of nuclear medicine computer software

    International Nuclear Information System (INIS)

    Cradduck, T.D.

    1986-01-01

    Although quality assurance activities have become well established for the hardware found in nuclear medicine little attention has been paid to computer software. This paper outlines some of the problems that exist and indicates some of the solutions presently under development. The major thrust has been towards establishment of programming standards and comprehensive documentation. Some manufacturers have developed installation verification procedures which programmers are urged to use as models for their own programs. Items that tend to cause erroneous results are discussed with the emphasis for error detection and correction being placed on proper education and training of the computer operator. The concept of interchangeable data files or 'software phantoms' for purposes of quality assurance is discussed. (Author)

  20. A pattern framework for software quality assessment and tradeoff analysis

    NARCIS (Netherlands)

    Folmer, Eelke; Boscht, Jan

    The earliest design decisions often have a significant impact on software quality and are the most costly to revoke. One of the challenges in architecture design is to reduce the frequency of retrofit problems in software designs; not being able to improve the quality of a system cost effectively, a

  1. Comparison of Simulated Microgravity and Hydrostatic Pressure for Chondrogenesis of hASC.

    Science.gov (United States)

    Mellor, Liliana F; Steward, Andrew J; Nordberg, Rachel C; Taylor, Michael A; Loboa, Elizabeth G

    2017-04-01

    Cartilage tissue engineering is a growing field due to the lack of regenerative capacity of native tissue. The use of bioreactors for cartilage tissue engineering is common, but the results are controversial. Some studies suggest that microgravity bioreactors are ideal for chondrogenesis, while others show that mimicking hydrostatic pressure is crucial for cartilage formation. A parallel study comparing the effects of loading and unloading on chondrogenesis has not been performed. The goal of this study was to evaluate chondrogenesis of human adipose-derived stem cells (hASC) under two different mechanical stimuli relative to static culture: microgravity and cyclic hydrostatic pressure (CHP). Pellets of hASC were cultured for 14 d under simulated microgravity using a rotating wall vessel bioreactor or under CHP (7.5 MPa, 1 Hz, 4 h · d-1) using a hydrostatic pressure vessel. We found that CHP increased mRNA expression of Aggrecan, Sox9, and Collagen II, caused a threefold increase in sulfated glycosaminoglycan production, and resulted in stronger vimentin staining intensity and organization relative to microgravity. In addition, Wnt-signaling patterns were altered in a manner that suggests that simulated microgravity decreases chondrogenic differentiation when compared to CHP. Our goal was to compare chondrogenic differentiation of hASC using a microgravity bioreactor and a hydrostatic pressure vessel, two commonly used bioreactors in cartilage tissue engineering. Our results indicate that CHP promotes hASC chondrogenesis and that microgravity may inhibit hASC chondrogenesis. Our findings further suggest that cartilage formation and regeneration might be compromised in space due to the lack of mechanical loading.Mellor LF, Steward AJ, Nordberg RC, Taylor MA, Loboa EG. Comparison of simulated microgravity and hydrostatic pressure for chondrogenesis of hASC. Aerosp Med Hum Perform. 2017; 88(4):377-384.

  2. Sandia National Laboratories ASCI Applications Software Quality Engineering Practices; TOPICAL

    International Nuclear Information System (INIS)

    ZEPPER, JOHN D.; ARAGON, KATHRYN MARY; ELLIS, MOLLY A.; BYLE, KATHLEEN A.; EATON, DONNA SUE

    2002-01-01

    This document provides a guide to the deployment of the software verification activities, software engineering practices, and project management principles that guide the development of Accelerated Strategic Computing Initiative (ASCI) applications software at Sandia National Laboratories (Sandia). The goal of this document is to identify practices and activities that will foster the development of reliable and trusted products produced by the ASCI Applications program. Document contents include an explanation of the structure and purpose of the ASCI Quality Management Council, an overview of the software development lifecycle, an outline of the practices and activities that should be followed, and an assessment tool. These sections map practices and activities at Sandia to the ASCI Software Quality Engineering: Goals, Principles, and Guidelines, a Department of Energy document

  3. Software quality assurance plan for viscometer

    International Nuclear Information System (INIS)

    Gimera, M.

    1994-01-01

    The in situ viscometer is a portable instrument designed to raise and lower a sphere (rheometer ball) through layers of tank waste material while recording ball position, velocity, and cable tension. In the field, the viscometer attaches to a decontamination spool piece which in turn is designed to attach to any 4-inch, 150-pound flange (typical of many available tank risers). The motion of the ball and collection of data is controlled by instrumentation and control equipment housed in a separate remote control console. This document covers the product, Viscometer Data Acquisition Software. This document provides the software quality assurance plan, verification and validation plan, and configuration management plan for developing the software for the instrumentation that will be used to obtain rheology data from Tank SY-101

  4. EMMC guidance on quality assurance for academic materials modelling software engineering

    OpenAIRE

    European Materials Modelling Council

    2015-01-01

    Proposed recommendations for software development in LEIT projects. This document presents the advice of software owners, commercial and academic, on what academic software could do to generate better quality software, ready to be used by third parties.

  5. The impact of software quality characteristics on healthcare outcome: a literature review.

    Science.gov (United States)

    Aghazadeh, Sakineh; Pirnejad, Habibollah; Moradkhani, Alireza; Aliev, Alvosat

    2014-01-01

    The aim of this study was to discover the effect of software quality characteristics on healthcare quality and efficiency indicators. Through a systematic literature review, we selected and analyzed 37 original research papers to investigate the impact of the software indicators (coming from the standard ISO 9126 quality characteristics and sub-characteristics) on some of healthcare important outcome indicators and finally ranked these software indicators. The results showed that the software characteristics usability, reliability and efficiency were mostly favored in the studies, indicating their importance. On the other hand, user satisfaction, quality of patient care, clinical workflow efficiency, providers' communication and information exchange, patient satisfaction and care costs were among the healthcare outcome indicators frequently evaluated in relation to the mentioned software characteristics. Regression Logistic Method was the most common assessment methodology, and Confirmatory Factor Analysis and Structural Equation Modeling were performed to test the structural model's fit. The software characteristics were considered to impact the healthcare outcome indicators through other intermediate factors (variables).

  6. MicroASC instrument onboard Juno spacecraft utilizing inertially controlled imaging

    DEFF Research Database (Denmark)

    Pedersen, David Arge Klevang; Jørgensen, Andreas Härstedt; Benn, Mathias

    2016-01-01

    This contribution describes the post-processing of the raw image data acquired by the microASC instrument during the Earth-fly-by of the Juno spacecraft. The images show a unique view of the Earth and Moon system as seen from afar. The procedure utilizes attitude measurements and inter......-calibration of the Camera Head Units of the microASC system to trigger the image capturing. The triggering is synchronized with the inertial attitude and rotational phase of the sensor acquiring the images. This is essentially works as inertially controlled imaging facilitating image acquisition from unexplored...

  7. Recommendations for a Software Quality Assurance Plan for the CMR Facility at LANL

    International Nuclear Information System (INIS)

    Adams, K.; Matthews, S. D.; McQueen, M. A.

    1998-01-01

    The Nuclear Materials Technology (NMT) organizations 1 and 3 within the Chemical and Metallurgical Research (CMR) facility at the Los Alamos National Laboratory are working to achieve Waste Isolation Pilot Plant (WIPP) certification to enable them to transport their TRU waste to WIPP. This document is intended to provide not only recommendations to address the necessary software quality assurance activities to enable the NMT-1 and NMT-3 organizations to be WIPP compliant but is also meant to provide a template for the final Software Quality Assurance Plan (SQAP). This document specifically addresses software quality assurance for all software used in support of waste characterization and analysis. Since NMT-1 and NMT-3 currently have several operational software products that are used for waste characterization and analysis, these software quality assurance recommendations apply to the operations, maintenance and retirement of the software and the creation and development of any new software required for waste characterization and analyses

  8. Human Factors in Software Development Processes: Measuring System Quality

    DEFF Research Database (Denmark)

    Abrahão, Silvia; Baldassarre, Maria Teresa; Caivano, Danilo

    2016-01-01

    Software Engineering and Human-Computer Interaction look at the development process from different perspectives. They apparently use very different approaches, are inspired by different principles and address different needs. But, they definitively have the same goal: develop high quality software...... in the most effective way. The second edition of the workshop puts particular attention on efforts of the two communities in enhancing system quality. The research question discussed is: who, what, where, when, why, and how should we evaluate?...

  9. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    Batinkov, Atanas Ivanov

    2013-01-01

    The Data Quality Monitoring Software proved to be a central tool in the Compact Muon Solenoid experiment. Its flexibility allowed its integration in several environments: online, for real-time detector monitoring; offline, for the final, fine-grained data certification. The usage of the Data Quality Monitoring software in the different environments and its integration in the Compact Muon Solenoid reconstruction software framework and in all production workflows are presented. The main technical challenges and the adopted solutions to them will be also discussed with emphasis on functionality, long-term robustness and performance.

  10. Software Quality Improvement in the OMC Team

    CERN Document Server

    Maier, Viktor

    Physicists use self-written software as a tool to fulfill their tasks and often the developed software is used for several years or even decades. If a software product lives for a long time, it has to be changed and adapted to external influences. This implies that the source code has to be read, understood and modified. The same applies to the software of the Optics Measurements and Corrections (OMC) team at CERN. Their task is to track, analyze and correct the beams in the LHC and other accelerators. To solve this task, they revert to a self-written software base with more than 150,000 physical lines of code. The base is subject to continuous changes as well. Their software does its job and is effective, but runs regrettably not efficient because some parts of the source code are in a bad shape and has a low quality. The implementation could be faster and more memory efficient. In addition it is difficult to read and understand the code. Source code files and functions are too big and identifiers do not rev...

  11. Continuous integration and quality control for scientific software

    Science.gov (United States)

    Neidhardt, A.; Ettl, M.; Brisken, W.; Dassing, R.

    2013-08-01

    Modern software has to be stable, portable, fast and reliable. This is going to be also more and more important for scientific software. But this requires a sophisticated way to inspect, check and evaluate the quality of source code with a suitable, automated infrastructure. A centralized server with a software repository and a version control system is one essential part, to manage the code basis and to control the different development versions. While each project can be compiled separately, the whole code basis can also be compiled with one central “Makefile”. This is used to create automated, nightly builds. Additionally all sources are inspected automatically with static code analysis and inspection tools, which check well-none error situations, memory and resource leaks, performance issues, or style issues. In combination with an automatic documentation generator it is possible to create the developer documentation directly from the code and the inline comments. All reports and generated information are presented as HTML page on a Web server. Because this environment increased the stability and quality of the software of the Geodetic Observatory Wettzell tremendously, it is now also available for scientific communities. One regular customer is already the developer group of the DiFX software correlator project.

  12. A perspective on software quality management using microcomputers in safety-related activities

    International Nuclear Information System (INIS)

    Braudt, T.E.; Pratl, M.J.

    1992-01-01

    Software Quality Management, often referred to as Software Quality Assurance or SQA, is a belief or mindset in establishing and protecting the value of software as a corporate asset. It is often expressed in terms of a basic methodology for ensuring adequate controls to maintain the integrity of the configuration of a software system. SQA applies to all activities germane to the acquisition, installation, operation and maintenance of software systems and is key to calculational accuracy and completeness in an Engineering and/or Scientific arena. Simply, it is a vital management tool for ensuring cost-effective utilization of information management resources. The basis principles of SQA apply equally to software applications in microcomputer environments and mainframe environments alike. Regardless of the nature of the computing environment, divisions of responsibilities or logistical difficulties, quality measures must be established to ensure accuracy, completeness, reliability, and reproducibility of the results of the software application. The extent to which these measures are applied should be based upon regulation, economics and practicality

  13. IMPROVING (SOFTWARE) PATENT QUALITY THROUGH THE ADMINISTRATIVE PROCESS.

    Science.gov (United States)

    Rai, Arti K

    2013-11-24

    The available evidence indicates that patent quality, particularly in the area of software, needs improvement. This Article argues that even an agency as institutionally constrained as the U.S. Patent and Trademark Office ("PTO") could implement a portfolio of pragmatic, cost-effective quality improvement strategies. The argument in favor of these strategies draws upon not only legal theory and doctrine but also new data from a PTO software examination unit with relatively strict practices. Strategies that resolve around Section 112 of the patent statute could usefully be deployed at the initial examination stage. Other strategies could be deployed within the new post-issuance procedures available to the agency under the America Invents Act. Notably, although the strategies the Article discusses have the virtue of being neutral as to technology, they are likely to have a very significant practical impact in the area of software.

  14. Quality control in diagnostic radiology: software (Visual Basic 6) and database applications

    International Nuclear Information System (INIS)

    Md Saion Salikin; Muhammad Farid Abdul Khalid

    2002-01-01

    Quality Assurance programme in diagnostic Radiology is being implemented by the Ministry of Health (MoH) in Malaysia. Under this program the performance of an x-ray machine used for diagnostic purpose is tested by using the approved procedure which is commonly known as Quality Control in diagnostic radiology. The quality control or performance tests are carried out b a class H licence holder issued the Atomic Energy Licensing Act 1984. There are a few computer applications (software) that are available in the market which can be used for this purpose. A computer application (software) using Visual Basics 6 and Microsoft Access, is being developed to expedite data handling, analysis and storage as well as report writing of the quality control tests. In this paper important features of the software for quality control tests are explained in brief. A simple database is being established for this purpose which is linked to the software. Problems encountered in the preparation of database are discussed in this paper. A few examples of practical usage of the software and database applications are presented in brief. (Author)

  15. Improving Software Quality and Management Through Use of Service Level Agreements

    Science.gov (United States)

    2005-03-01

    many who believe that the quality of the development process is the best predictor of software product quality. ( Fenton ) Repeatable software processes...reduced errors per KLOC for small projects ( Fenton ), and the quality management metric (QMM) (Machniak, Osmundson). There are also numerous IEEE 14...attention to cosmetic user interface issues and any problems that may arise with the prototype. (Sawyer) The validation process is also another check

  16. Software quality assurance (SQA) for Savannah River reactors

    Energy Technology Data Exchange (ETDEWEB)

    Schaumann, C.M.

    1990-01-01

    Over the last 25 years, the Savannah River Site (SRS) has developed a strong Software Quality Assurance (SQA) program. It provides the information and management controls required of a high quality auditable system. The SRS SQA program provides the framework to meet the requirements in increasing regulation.

  17. Quality of Design, Analysis and Reporting of Software Engineering Experiments:A Systematic Review

    OpenAIRE

    By Kampenes, Vigdis

    2007-01-01

    Background: Like any research discipline, software engineering research must be of a certain quality to be valuable. High quality research in software engineering ensures that knowledge is accumulated and helpful advice is given to the industry. One way of assessing research quality is to conduct systematic reviews of the published research literature. Objective: The purpose of this work was to assess the quality of published experiments in software engineering with respect to the validit...

  18. [Effects of chronic fluoxetine treatment on manifestation of sexual motivation and social behavior in mice of ASC line].

    Science.gov (United States)

    Tikhonova, M A; Otroshchenko, E A; Kulikov, A V

    2010-02-01

    Sexual dysfunctions are the typical symptoms accompanying depressive disorders. However antidepressants which improve general state of the patients have no effect on sexual disorders. Mice of ASC (Antidepressant Sensitive Catalepsy) line with high hereditary predisposition to catalepsy were proposed as a model of genetically associated depressive-like condition. The work was aimed at comparison of behavioral indices of sexual motivation and social interest of ASC mice with those of mice of parental inbred AKR and CBA strains, and at the study of the effects of chronic fluoxetine treatment in doses of 10 and 20 mg/kg on these parameters in ASC mice. ASC males demonstrated reduced sexual motivation which was not corrected by fluoxetine. ASC mice did not differ in the expression of social interest and aggression towards juvenile intruder from mice of parental strains. Fluoxetine failed to alter social behavior of ASC mice in social interaction test but its higher dose decreased percentage of aggressors. ASC mouse line seems to be a perspective model to study genetic mechanisms of sexual dysfunctions associated with depressive conditions.

  19. The software quality control for gamma spectrometry

    International Nuclear Information System (INIS)

    Monte, L.

    1986-01-01

    One of major problems with wich the quality control program of an environmental measurements laboratory is confronted is the evaluation of the performances of software packages for the analysis of gamma-ray spectra. A program of tests for evaluating the performances of the software package (SPECTRAN-F, Canberra Inc.) used by our laboratory is being carried out. In this first paper the results of a preliminary study concerning the evaluation of the performance of the doublet analysis routine are presented

  20. Software quality assurance for safety analysis and risk management at the Savannah River Site

    International Nuclear Information System (INIS)

    Ades, M.J.; Toffer, H.; Crowe, R.D.

    1991-01-01

    As part of its Reactor Operations Improvement Program at the Savannah River Site (SRS), Westinghouse Savannah River Company (WSRC), in cooperation with the Westinghouse Hanford Company, has developed and implemented quality assurance for safety-related software for technical programs essential to the safety and reliability of reactor operations. More specifically, the quality assurance process involved the development and implementation of quality standards and attendant procedures based on industry software quality standards. These procedures were then applied to computer codes in reactor safety and probabilistic risk assessment analyses. This paper provides a review of the major aspects of the WSRC safety-related software quality assurance. In particular, quality assurance procedures are described for the different life cycle phases of the software that include the Requirements, Software Design and Implementation, Testing and Installation, Operation and Maintenance, and Retirement Phases. For each phase, specific provisions are made to categorize the range of activities, the level of responsibilities, and the documentation needed to assure the control of the software. The software quality assurance procedures developed and implemented are evolutionary in nature, and thus, prone to further refinements. These procedures, nevertheless, represent an effective controlling tool for the development, production, and operation of safety-related software applicable to reactor safety and probabilistic risk assessment analyses

  1. Quality assurance of the modernized Dukovany I and C safety system software

    International Nuclear Information System (INIS)

    Karpeta, C.

    2005-01-01

    The approach to quality assurance of the software that implements the instrumentation and control functions for safety category A as per IEC 61226, which has been adopted within the 'NPP Dukovany I and C Refurbishment' project, is described. A survey of the requirements for software quality assurance of the systems that initiate protection interventions in the event of anticipated operational occurrences or accident conditions is given. The software development process applied by the system designers and manufacturers, from the software requirements specification phase to the software testing phase, is outlined. Basic information on technical audits of the software development process is also provided. (orig.)

  2. Acute stress among adolescents and female rape victims measured by ASC-Kids: a pilot study.

    Science.gov (United States)

    Nilsson, Doris; Nordenstam, Carin; Green, Sara; Wetterhall, Annika; Lundin, Tom; Svedin, Carl Göran

    2015-01-01

    Rape is considered a stressful trauma and often with durable consequences. How the aftermath of rape is for young adolescents' girls considering acute stress is an overlooked field and remains to be studied. In this study, we wanted to investigate acute stress among adolescent victims of rape and the psychometric properties of the Acute Stress Checklist for Children (ASC-Kids). A clinical sample (n = 79) of raped girls, 13-17 years old who had turned to a special rape victim unit for treatment, answered the ASC-Kids. ASC-Kids was also given to a group of minor stressed, non-raped adolescents in the same age range (n = 154) together with the University of California at Los Angeles Post-traumatic Stress Disorder Reaction Index (UCLA PTSD RI), and the Sense of Coherence Scale 13 (SOC-13). The scores from the groups were compared and showed significant differences in mean values on all the diagnostic criteria of acute stress disorder. In the clinical group, 36.7% obtained full ASD criteria. ASC-Kids could discriminate well between groups. Cronbach's alpha was found to be excellent, and the correlation between the UCLA PTSD RI and ASC-Kids found to be good; both ASC-Kids and UCLA PTSD RI had a good and moderate negative correlation with SOC-13. Adolescent female rape victims were shown to have a very high level of acute stress, and the ASC-Kids was found to have sound psychometrics and can be a valuable screening instrument to support clinicians in their assessments of an indication of adolescents after potentially stressful events such as rape.

  3. Linear accelerator quality assurance using EPIQA software

    International Nuclear Information System (INIS)

    Bozhikov, S.; Sokerov, H.; Tonev, A.; Ivanova, K.

    2012-01-01

    Unlike treatment with static fields, using a dynamic multileaf collimator (dMLC), there are significant dosimetric issues which must be assessed before dynamic therapy can be implemented. The advanced techniques require some additional commissioning and quality assurance tests. The results of standard quality assurance (QA) machine tests and commissioning tests for volume modulated arc therapy (VMAT) using electronic portal image device (EPID) and 'EPIQA' software are presented. (authors)

  4. How many records should be used in ASCE/SEI-7 ground motion scaling procedure?

    Science.gov (United States)

    Reyes, Juan C.; Kalkan, Erol

    2012-01-01

    U.S. national building codes refer to the ASCE/SEI-7 provisions for selecting and scaling ground motions for use in nonlinear response history analysis of structures. Because the limiting values for the number of records in the ASCE/SEI-7 are based on engineering experience, this study examines the required number of records statistically, such that the scaled records provide accurate, efficient, and consistent estimates of “true” structural responses. Based on elastic–perfectly plastic and bilinear single-degree-of-freedom systems, the ASCE/SEI-7 scaling procedure is applied to 480 sets of ground motions; the number of records in these sets varies from three to ten. As compared to benchmark responses, it is demonstrated that the ASCE/SEI-7 scaling procedure is conservative if fewer than seven ground motions are employed. Utilizing seven or more randomly selected records provides more accurate estimate of the responses. Selecting records based on their spectral shape and design spectral acceleration increases the accuracy and efficiency of the procedure.

  5. Quality factors in the life cycle of software oriented to safety systems in nuclear power plants

    International Nuclear Information System (INIS)

    Nunez McLeod, J.E.; Rivera, S.S.

    1997-01-01

    The inclusion of software in safety related systems for nuclear power plants, makes it necessary to include the software quality assurance concept. The software quality can be defined as the adjustment degree between the software and the specified requirements and user expectations. To guarantee a certain software quality level it is necessary to make a systematic and planned set of tasks, that constitute a software quality guaranty plan. The application of such a plan involves activities that should be performed all along the software life cycle, and that can be evaluated through the so called quality factors, due to the fact that the quality itself cannot be directly measured, but indirectly as some of it manifestations. In this work, a software life cycle model is proposed, for nuclear power plant safety related systems. A set os software quality factors is also proposed , with its corresponding classification according to the proposed model. (author) [es

  6. MRI-tracking of transplanted human ASC in a SCID mouse model

    Energy Technology Data Exchange (ETDEWEB)

    Siegmund, Birte J.; Kasten, Annika [Department of Oral and Maxillofacial Surgery, Facial Plastic Surgery, Rostock University Medical Center (Germany); Kühn, Jens-Peter [Institute of Diagnostic Radiology and Neuroradiology, Greifswald University Medical Center (Germany); Winter, Karsten [Institute of Anatomy, Faculty of Medicine, University of Leipzig (Germany); Grüttner, Cordula [Micromod Partikeltechnologie GmbH, Rostock (Germany); Frerich, Bernhard, E-mail: bernhard.frerich@med.uni-rostock.de [Department of Oral and Maxillofacial Surgery, Facial Plastic Surgery, Rostock University Medical Center (Germany)

    2017-04-01

    Background: Regarding strategies improving the efficacy of stem cell transplantation in adipose tissue engineering, cell tracking might be useful. Here we report the in vivo tracking of adipose tissue derived stem cells (ASC) by means of nanoparticle labeling and magnetic resonance imaging (MRI). Here we report the in vivo tracking of adipose tissue derived stromal cells (ASC) by means of nanoparticle labeling and magnetic resonance imaging (MRI). Materials and methods: Human ASC were amplified and labeled with two types of magnetic nanoparticles (MNP), BNF starch and nanomag®-D-spio. Adipose tissue constructs were fabricated by seeding collagen scaffolds with labeled and unlabeled ASCs. Constructs were implanted subcutaneously in the back of severe combined immunodeficient (SCID) mice (n =69, group 1: control with cells w/o label, group 2: BNF starch labeled cells, group 3: nanomag®-D-spio labeled cells). MRI scans were performed at 24 hours, four, twelve and 28 days and four months in a 7.1 T animal device. Explanted constructs were analyzed histomorphometrically. Results: MRI scans showed high contrast of the labeled cells in t2-tse-sequence compared to unlabeled controls. Loss of volume of the implants was observed over time due to partial loss for transplanted cells without significant difference (level of significance p<0.017). Compared to histomorphometry, there was found a positiv correlations in measurement of implant size with a significant at day four (correlation coefficient =0.643; p=0.024) and day twelve (correlation coefficient =0.687; p=0.010). Additional Prussian blue stain showed iron in all implants. Significant differences between the three groups (significance level p<0.017) were found after twelve days between control group and group 3 (p=0.008) and after 28 days between control group and group 2 and 3 (p=0.011). Conclusion: Both MNPs might be suitable for tracking of ASC in vivo and show long term stability over 4 months. - Highlights:

  7. Quality assessment with the AGIR software results and experience

    International Nuclear Information System (INIS)

    Rauch, D.; Kotter, E.; Kurtz, C.; Schaefer, O.; Ehritt-Braun, C.; Burger, D.; Schaper, J.; Uhrmeister, P.

    2001-01-01

    Purpose: To evaluate whether a new software from the working group for interventional radiology (AGIR) is an appropriate tool for quality assurance in interventional radiology, and presentation of results acquired within the quality improvement process in 1999. Patients and methods: AGIR-defined parameters such as patient data, risk profile, given interventions as well as complications were registered by a recently developed software. Based on monthly data analyses, possible complications were identified and discussed in morbidity and mortality conferences. Results: 1014 interventions were performed in our institution in 1999. According to criteria established by AGIR, the complication rate was 2.7%. In addition and according to SCVIR criteria, complications were distinguished quantitatively in five classes and semiquantitatively in minor and major groups. The result was a minor complication rate of 1.8%, and a major rate of 0.9%. There were no cases of death associated with the intervention. Further strategies were developed in order to reduce the complication rate. Conclusion: Extensive quality assurance methods can be integrated in daily routine work. These methods lead to an intensive transparency of treatment results, and allow the implementation of continuous quality improvements. The development of the software is a first step in establishing a nation-wide quality assurance system. Nevertheless, modification and additional definition of the AGIR predefined parameters are required, for example, to avoid unnecessary procedures. (orig.) [de

  8. Research of quality control during development of NPP DCS 1E classified software

    International Nuclear Information System (INIS)

    Shi Weihua; Lu Zhenguo; Xie Qi

    2012-01-01

    The Nuclear safety depends on right behavior of 1E software, which is a important part of 1E DCS system. Nowadays, user focus on good function of 1E system, but pay little attention to quality control of 1E software. In fact, it's declared in IEC61513 and IEC60880 that 1E software should under strict quality control during all stages of development. This article is related to the practice of 1E DCS system quality control and explores the QC surveillance for 1E software from the user's point of view. (authors)

  9. Software application for quality control protocol of mammography systems

    International Nuclear Information System (INIS)

    Kjosevski, Vladimir; Gershan, Vesna; Ginovska, Margarita; Spasevska, Hristina

    2010-01-01

    Considering the fact that the Quality Control of the technological process of the mammographic system involves testing of a large number of parameters, it is clearly evident that there is a need for using the information technology for gathering, processing and storing of all the parameters that are result of this process. The main goal of this software application is facilitation and automation of the gathering, processing, storing and presenting process of the data related to the qualification of the physical and technical parameters during the quality control of the technological process of the mammographic system. The software application along with its user interface and database has been made with the Microsoft Access 2003 application which is part of the Microsoft Office 2003 software packet and has been chosen as a platform for developing because it is the most commonly used office application today among the computer users in the country. This is important because it will provide the end users a familiar environment to work in, without the need for additional training and improving the computer skills that they posses. Most importantly, the software application is easy to use, fast in calculating the parameters needed and it is an excellent way to store and display the results. There is a possibility for up scaling this software solution so it can be used by many different users at the same time over the Internet. It is highly recommended that this system is implemented as soon as possible in the quality control process of the mammographic systems due to its many advantages.(Author)

  10. Filmes plásticos e ácido ascórbico na qualidade de araticum minimamente processado Plastic packaging film and ascorbic acid treatment on the quality of fresh cut araticum

    Directory of Open Access Journals (Sweden)

    Manoel Soares Soares Júnior

    2007-12-01

    Full Text Available O objetivo deste trabalho foi avaliar os efeitos do ácido ascórbico e do tipo de filme plástico como embalagem na qualidade do araticum minimamente processado e mantido sob refrigeração. O ácido ascórbico não evitou o escurecimento do araticum minimamente processado. Independentemente do tipo de embalagem, a acidez titulável aumentou com o tempo. A embalagem de policloreto de vinila ou polietileno de baixa densidade promoveu uma significativa perda de massa se comparada com a a laminada a vácuo. A embalagem laminada a vácuo propiciou vida-de-prateleira mais longa ao produto, o qual permaneceu com aparência adequada e qualidade comercial até o sétimo dia do armazenamento. A vida-de-prateleira dos demais tratamentos alcançou somente três dias.This study was aimed at evaluating the effect of ascorbic acid and type of plastic packaging film on the quality of refrigerated fresh cut araticum. Ascorbic acid did not prevent fresh cut araticum from darkening. Regardless of the type of plastic packaging, the fruit titrable acidity increased with time. Packing with polyvynil chloride or with low density polyethylene promoted a significant mass loss compared to laminate vacuum packaging. Laminate vacuum packaging increased the shelf life of the product up 7 days, maintaining its commercial quality and appearance. The shelf life of the other treatments reached 3 days only.

  11. Software quality assurance plan for the National Ignition Facility integrated computer control system

    Energy Technology Data Exchange (ETDEWEB)

    Woodruff, J.

    1996-11-01

    Quality achievement is the responsibility of the line organizations of the National Ignition Facility (NIF) Project. This Software Quality Assurance Plan (SQAP) applies to the activities of the Integrated Computer Control System (ICCS) organization and its subcontractors. The Plan describes the activities implemented by the ICCS section to achieve quality in the NIF Project`s controls software and implements the NIF Quality Assurance Program Plan (QAPP, NIF-95-499, L-15958-2) and the Department of Energy`s (DOE`s) Order 5700.6C. This SQAP governs the quality affecting activities associated with developing and deploying all control system software during the life cycle of the NIF Project.

  12. Software quality assurance plan for the National Ignition Facility integrated computer control system

    International Nuclear Information System (INIS)

    Woodruff, J.

    1996-11-01

    Quality achievement is the responsibility of the line organizations of the National Ignition Facility (NIF) Project. This Software Quality Assurance Plan (SQAP) applies to the activities of the Integrated Computer Control System (ICCS) organization and its subcontractors. The Plan describes the activities implemented by the ICCS section to achieve quality in the NIF Project's controls software and implements the NIF Quality Assurance Program Plan (QAPP, NIF-95-499, L-15958-2) and the Department of Energy's (DOE's) Order 5700.6C. This SQAP governs the quality affecting activities associated with developing and deploying all control system software during the life cycle of the NIF Project

  13. Lessons learned from development and quality assurance of software systems at the Halden Project

    International Nuclear Information System (INIS)

    Bjorlo, T.J.; Berg, O.; Pehrsen, M.; Dahll, G.; Sivertsen, T.

    1996-01-01

    The OECD Halden Reactor Project has developed a number of software systems within the research programmes. These programmes have comprised a wide range of topics, like studies of software for safety-critical applications, development of different operator support systems, and software systems for building and implementing graphical user interfaces. The systems have ranged from simple prototypes to installations in process plants. In the development of these software systems, Halden has gained much experience in quality assurance of different types of software. This paper summarises the accumulated experience at the Halden Project in quality assurance of software systems. The different software systems being developed at the Halden Project may be grouped into three categories. These are plant-specific software systems (one-of-a-kind deliveries), generic software products, and safety-critical software systems. This classification has been found convenient as the categories have different requirements to the quality assurance process. In addition, the experience from use of software development tools and proprietary software systems at Halden, is addressed. The paper also focuses on the experience gained from the complete software life cycle, starting with the software planning phase and ending with software operation and maintenance

  14. Recommendations for a software quality assurance policy for the nuclear waste disposal risk assessment programme

    International Nuclear Information System (INIS)

    Hill, I.E.

    1985-05-01

    This study reviewed a number of published standards for software quality assurance, and included a series of interviews with software developers aimed at exploring their attitudes to software quality assurance. Recommendations for software quality assurance policy are made based on the above investigations. This document provides a summary of the recommendations made in the full report on project, reference MR-CDS-4. (author)

  15. MCNP trademark Software Quality Assurance plan

    International Nuclear Information System (INIS)

    Abhold, H.M.; Hendricks, J.S.

    1996-04-01

    MCNP is a computer code that models the interaction of radiation with matter. MCNP is developed and maintained by the Transport Methods Group (XTM) of the Los Alamos National Laboratory (LANL). This plan describes the Software Quality Assurance (SQA) program applied to the code. The SQA program is consistent with the requirements of IEEE-730.1 and the guiding principles of ISO 900

  16. Faster Defect Resolution with Higher Technical Quality of Software

    NARCIS (Netherlands)

    Luijten, B.; Visser, J.

    2010-01-01

    We performed an empirical study of the relation between technical quality of software products and the defect resolution performance of their maintainers. In particular, we tested the hypothesis that ratings for source code maintainability, as employed by the SIG quality model, are correlated with

  17. Evolvability as a Quality Attribute of Software Architectures

    NARCIS (Netherlands)

    Ciraci, S.; van den Broek, P.M.; Duchien, Laurence; D'Hondt, Maja; Mens, Tom

    We review the definition of evolvability as it appears on the literature. In particular, the concept of software evolvability is compared with other system quality attributes, such as adaptability, maintainability and modifiability.

  18. System Quality Management in Software Testing Laboratory that Chooses Accreditation

    Directory of Open Access Journals (Sweden)

    Yanet Brito R.

    2013-12-01

    Full Text Available The evaluation of software products will reach full maturity when executed by the scheme and provides third party certification. For the validity of the certification, the independent laboratory must be accredited for that function, using internationally recognized standards. This brings with it a challenge for the Industrial Laboratory Testing Software (LIPS, responsible for testing the products developed in Cuban Software Industry, define strategies that will permit it to offer services with a high level of quality. Therefore it is necessary to establish a system of quality management according to NC-ISO/IEC 17025: 2006 to continuously improve the operational capacity and technical competence of the laboratory, with a view to future accreditation of tests performed. This article discusses the process defined in the LIPS for the implementation of a Management System of Quality, from the current standards and trends, as a necessary step to opt for the accreditation of the tests performed.

  19. Software Quality Evaluation Models Applicable in Health Information and Communications Technologies. A Review of the Literature.

    Science.gov (United States)

    Villamor Ordozgoiti, Alberto; Delgado Hito, Pilar; Guix Comellas, Eva María; Fernandez Sanchez, Carlos Manuel; Garcia Hernandez, Milagros; Lluch Canut, Teresa

    2016-01-01

    Information and Communications Technologies in healthcare has increased the need to consider quality criteria through standardised processes. The aim of this study was to analyse the software quality evaluation models applicable to healthcare from the perspective of ICT-purchasers. Through a systematic literature review with the keywords software, product, quality, evaluation and health, we selected and analysed 20 original research papers published from 2005-2016 in health science and technology databases. The results showed four main topics: non-ISO models, software quality evaluation models based on ISO/IEC standards, studies analysing software quality evaluation models, and studies analysing ISO standards for software quality evaluation. The models provide cost-efficiency criteria for specific software, and improve use outcomes. The ISO/IEC25000 standard is shown as the most suitable for evaluating the quality of ICTs for healthcare use from the perspective of institutional acquisition.

  20. Function of microRNAs in the Osteogenic Differentiation and Therapeutic Application of Adipose-Derived Stem Cells (ASCs

    Directory of Open Access Journals (Sweden)

    Walter M. Hodges

    2017-12-01

    Full Text Available Traumatic wounds with segmental bone defects represent substantial reconstructive challenges. Autologous bone grafting is considered the gold standard for surgical treatment in many cases, but donor site morbidity and associated post-operative complications remain a concern. Advances in regenerative techniques utilizing mesenchymal stem cell populations from bone and adipose tissue have opened the door to improving bone repair in the limbs, spine, and craniofacial skeleton. The widespread availability, ease of extraction, and lack of immunogenicity have made adipose-derived stem cells (ASCs particularly attractive as a stem cell source for regenerative strategies. Recently it has been shown that small, non-coding miRNAs are involved in the osteogenic differentiation of ASCs. Specifically, microRNAs such as miR-17, miR-23a, and miR-31 are expressed during the osteogenic differentiation of ASCs, and appear to play a role in inhibiting various steps in bone morphogenetic protein-2 (BMP2 mediated osteogenesis. Importantly, a number of microRNAs including miR-17 and miR-31 that act to attenuate the osteogenic differentiation of ASCs are themselves stimulated by transforming growth factor β-1 (TGFβ-1. In addition, transforming growth factor β-1 is also known to suppress the expression of microRNAs involved in myogenic differentiation. These data suggest that preconditioning strategies to reduce TGFβ-1 activity in ASCs may improve the therapeutic potential of ASCs for musculoskeletal application. Moreover, these findings support the isolation of ASCs from subcutaneous fat depots that tend to have low endogenous levels of TGFβ-1 expression.

  1. Antiproliferative effect of ASC-J9 delivered by PLGA nanoparticles against estrogen-dependent breast cancer cells.

    Science.gov (United States)

    Verderio, Paolo; Pandolfi, Laura; Mazzucchelli, Serena; Marinozzi, Maria Rosaria; Vanna, Renzo; Gramatica, Furio; Corsi, Fabio; Colombo, Miriam; Morasso, Carlo; Prosperi, Davide

    2014-08-04

    Among polymeric nanoparticles designed for cancer therapy, PLGA nanoparticles have become one of the most popular polymeric devices for chemotherapeutic-based nanoformulations against several kinds of malignant diseases. Promising properties, including long-circulation time, enhanced tumor localization, interference with "multidrug" resistance effects, and environmental biodegradability, often result in an improvement of the drug bioavailability and effectiveness. In the present work, we have synthesized 1,7-bis(3,4-dimethoxyphenyl)-5-hydroxyhepta-1,4,6-trien-3-one (ASC-J9) and developed uniform ASC-J9-loaded PLGA nanoparticles of about 120 nm, which have been prepared by a single-emulsion process. Structural and morphological features of the nanoformulation were analyzed, followed by an accurate evaluation of the in vitro drug release kinetics, which exhibited Fickian law diffusion over 10 days. The intracellular degradation of ASC-J9-bearing nanoparticles within estrogen-dependent MCF-7 breast cancer cells was correlated to a time- and dose-dependent activity of the released drug. A cellular growth inhibition associated with a specific cell cycle G2/M blocking effect caused by ASC-J9 release inside the cytosol allowed us to put forward a hypothesis on the action mechanism of this nanosystem, which led to the final cell apoptosis. Our study was accomplished using Annexin V-based cell death analysis, MTT assessment of proliferation, radical scavenging activity, and intracellular ROS evaluation. Moreover, the intracellular localization of nanoformulated ASC-J9 was confirmed by a Raman optical imaging experiment designed ad hoc. PLGA nanoparticles and ASC-J9 proved also to be safe for a healthy embryo fibroblast cell line (3T3-L1), suggesting a possible clinical translation of this potential nanochemotherapeutic to expand the inherently poor bioavailability of hydrophobic ASC-J9 that could be proposed for the treatment of malignant breast cancer.

  2. Spectrum analysis on quality requirements consideration in software design documents.

    Science.gov (United States)

    Kaiya, Haruhiko; Umemura, Masahiro; Ogata, Shinpei; Kaijiri, Kenji

    2013-12-01

    Software quality requirements defined in the requirements analysis stage should be implemented in the final products, such as source codes and system deployment. To guarantee this meta-requirement, quality requirements should be considered in the intermediate stages, such as the design stage or the architectural definition stage. We propose a novel method for checking whether quality requirements are considered in the design stage. In this method, a technique called "spectrum analysis for quality requirements" is applied not only to requirements specifications but also to design documents. The technique enables us to derive the spectrum of a document, and quality requirements considerations in the document are numerically represented in the spectrum. We can thus objectively identify whether the considerations of quality requirements in a requirements document are adapted to its design document. To validate the method, we applied it to commercial software systems with the help of a supporting tool, and we confirmed that the method worked well.

  3. Development of innovative computer software to facilitate the setup and computation of water quality index.

    Science.gov (United States)

    Nabizadeh, Ramin; Valadi Amin, Maryam; Alimohammadi, Mahmood; Naddafi, Kazem; Mahvi, Amir Hossein; Yousefzadeh, Samira

    2013-04-26

    Developing a water quality index which is used to convert the water quality dataset into a single number is the most important task of most water quality monitoring programmes. As the water quality index setup is based on different local obstacles, it is not feasible to introduce a definite water quality index to reveal the water quality level. In this study, an innovative software application, the Iranian Water Quality Index Software (IWQIS), is presented in order to facilitate calculation of a water quality index based on dynamic weight factors, which will help users to compute the water quality index in cases where some parameters are missing from the datasets. A dataset containing 735 water samples of drinking water quality in different parts of the country was used to show the performance of this software using different criteria parameters. The software proved to be an efficient tool to facilitate the setup of water quality indices based on flexible use of variables and water quality databases.

  4. Over-expression, purification and characterization of an Asc-1 homologue from Gloeobacter violaceus

    DEFF Research Database (Denmark)

    Wang, Xiaole; Hald, Helle; Ernst, Heidi Asschenfeldt

    2010-01-01

    The human alanine-serine-cysteine transporter 1 (Asc-1) belongs to the slc7a family of solute carrier transporters. Asc-1 mediates the uptake of D-serine in an exchanger-type fashion, coupling the process to the release of alanine and cysteine. Among the bacterial Asc-1 homologues, one transporter...... of auto-induction was crucial for obtaining high yields and purity of the transporter. The transporter was purified with yields in the range of 0.2-0.4 mg per L culture and eluted in a single peak from a size-exclusion column. The circular dichroism spectrum revealed a folded and apparently all...

  5. Fundamentals of civil engineering an introduction to the ASCE body of knowledge

    CERN Document Server

    McCuen, Richard H; Wong, Melanie K

    2011-01-01

    While the ASCE Body of Knowledge (BOK2) is the codified source for all technical and non-technical information necessary for those seeking to attain licensure in civil engineering, recent graduates have notoriously been lacking in the non-technical aspects even as they excel in the technical. Fundamentals of Civil Engineering: An Introduction to the ASCE Body of Knowledge addresses this shortfall and helps budding engineers develop the knowledge, skills, and attitudes suggested and implied by the BOK2. Written as a resource for all of the non-technical outcomes not specifically covered in the

  6. Establishing software quality assurance

    International Nuclear Information System (INIS)

    Malsbury, J.

    1983-01-01

    This paper is concerned with four questions about establishing software QA: What is software QA. Why have software QA. What is the role of software QA. What is necessary to ensure the success of software QA

  7. Application of software quality assurance to a specific scientific code development task

    International Nuclear Information System (INIS)

    Dronkers, J.J.

    1986-03-01

    This paper describes an application of software quality assurance to a specific scientific code development program. The software quality assurance program consists of three major components: administrative control, configuration management, and user documentation. The program attempts to be consistent with existing local traditions of scientific code development while at the same time providing a controlled process of development

  8. Atypical squamous cells, cannot exclude high grade squamous intraepithelial (ASC-H in HIV-positive women

    Directory of Open Access Journals (Sweden)

    Michelow Pam

    2010-01-01

    Full Text Available Objective: South Africa has very high rates of both HIV infection and cervical pathology. The management of ASC-H is colposcopy and directed biopsy, but with so many women diagnosed with HSIL and a dearth of colposcopy centres in South Africa, women with cytologic diagnosis of ASC-H may not be prioritized for colposcopy. The aim of this study was to determine if HIV-positive women with a cytologic diagnosis of ASC-H should undergo immediate colposcopy or whether colposcopy can be delayed, within the context of an underfunded health care setting with so many competing health needs. Materials and Methods: A computer database search was performed from the archives of an NGO-administered clinic that offers comprehensive HIV care. All women with a cytologic diagnosis of ASC-H on cervical smears from September 2005 until August 2009 were identified. Histologic follow up was sought in all patients. Results: A total of 2111 cervical smears were performed and 41 diagnosed as ASC-H (1.94%. No histologic follow up data was available in 15 cases. Follow up histologic results were as follows: three negative (11.5%, five koilocytosis and/ or CIN1 (19.2%, ten CIN2 (38.5% and eight CIN3 (30.8%. There were no cases of invasive carcinoma on follow up. Conclusion: The current appropriate management of HIV-positive women in low-resource settings with a diagnosis of ASC-H on cervical smear is colposcopy, despite the costs involved. In the future and if cost-effective in developing nations, use of novel markers may help select which HIV-positive women can be managed conservatively and which ones referred for more active treatment. More research in this regard is warranted.

  9. A pioneering application of NQA-1 quality assurance standards in the development of software

    International Nuclear Information System (INIS)

    Weisbin, A.N.

    1988-01-01

    The application of NQA-1 Quality Assurance Standards to computer software programs has been recent at the Oak Ridge National Laboratory. One reason for systematically applying quality assurance to computer software is the extensive use of results from computer programs. to characterize potential sites for nuclear waste repositories leading ultimately to important policy making decisions. Because data from these programs characterize the likely radioactivity profile for many hundreds of years, experimental validation is not feasible. The Sensitivity and Uncertainty Analysis Methods Development Project (SUAMDP) was developed to formulate and utilize efficient and comprehensive methods for determining sensitivities of calculated results with respect to changes in all input parameters. The computerized methodology was embodied in the Gradient Enhanced Software System (GRESS). Due to the fact that GRESS was to be used in the site characterization for waste storage, stringent NQA-1 requirements were imposed by the sponsor. A working relationship between the Oak Ridge National Laboratory (ORNL) Quality Department and the research scientists developing GRESS was essential in achieving understanding and acceptance of the quality assurance requirements as applied to the SUAMDP. The relationship resulted in the SUAMDP becoming the first software project at ORNL to develop a comprehensive NQA-1 Quality Assurance Plan; this plan now serves as a model for software quality assurance at ORNL. This paper describes the evolution of this plan and its impact on the application of quality assurance procedures to software

  10. Adopting software quality measures for healthcare processes.

    Science.gov (United States)

    Yildiz, Ozkan; Demirörs, Onur

    2009-01-01

    In this study, we investigated the adoptability of software quality measures for healthcare process measurement. Quality measures of ISO/IEC 9126 are redefined from a process perspective to build a generic healthcare process quality measurement model. Case study research method is used, and the model is applied to a public hospital's Entry to Care process. After the application, weak and strong aspects of the process can be easily observed. Access audibility, fault removal, completeness of documentation, and machine utilization are weak aspects and these aspects are the candidates for process improvement. On the other hand, functional completeness, fault ratio, input validity checking, response time, and throughput time are the strong aspects of the process.

  11. Handbook of software quality assurance techniques applicable to the nuclear industry

    Energy Technology Data Exchange (ETDEWEB)

    Bryant, J.L.; Wilburn, N.P.

    1987-08-01

    Pacific Northwest Laboratory is conducting a research project to recommend good engineering practices in the application of 10 CFR 50, Appendix B requirements to assure quality in the development and use of computer software for the design and operation of nuclear power plants for NRC and industry. This handbook defines the content of a software quality assurance program by enumerating the techniques applicable. Definitions, descriptions, and references where further information may be obtained are provided for each topic.

  12. Handbook of software quality assurance techniques applicable to the nuclear industry

    International Nuclear Information System (INIS)

    Bryant, J.L.; Wilburn, N.P.

    1987-08-01

    Pacific Northwest Laboratory is conducting a research project to recommend good engineering practices in the application of 10 CFR 50, Appendix B requirements to assure quality in the development and use of computer software for the design and operation of nuclear power plants for NRC and industry. This handbook defines the content of a software quality assurance program by enumerating the techniques applicable. Definitions, descriptions, and references where further information may be obtained are provided for each topic

  13. ASC Addresses Unit Commanders' Concerns through LBE and Reset Programs

    National Research Council Canada - National Science Library

    Young, Mark E

    2008-01-01

    .... Army Sustainment Command (ASC), part of the U.S. Army Materiel Command (AMC) team, is available to assist, identify, and resolve equipment and maintenance problems as well as materiel readiness issues for combatant commanders...

  14. Digital radiography: optimization of image quality and dose using multi-frequency software.

    Science.gov (United States)

    Precht, H; Gerke, O; Rosendahl, K; Tingberg, A; Waaler, D

    2012-09-01

    New developments in processing of digital radiographs (DR), including multi-frequency processing (MFP), allow optimization of image quality and radiation dose. This is particularly promising in children as they are believed to be more sensitive to ionizing radiation than adults. To examine whether the use of MFP software reduces the radiation dose without compromising quality at DR of the femur in 5-year-old-equivalent anthropomorphic and technical phantoms. A total of 110 images of an anthropomorphic phantom were imaged on a DR system (Canon DR with CXDI-50 C detector and MLT[S] software) and analyzed by three pediatric radiologists using Visual Grading Analysis. In addition, 3,500 images taken of a technical contrast-detail phantom (CDRAD 2.0) provide an objective image-quality assessment. Optimal image-quality was maintained at a dose reduction of 61% with MLT(S) optimized images. Even for images of diagnostic quality, MLT(S) provided a dose reduction of 88% as compared to the reference image. Software impact on image quality was found significant for dose (mAs), dynamic range dark region and frequency band. By optimizing image processing parameters, a significant dose reduction is possible without significant loss of image quality.

  15. Rules of thumb to increase the software quality through testing

    Science.gov (United States)

    Buttu, M.; Bartolini, M.; Migoni, C.; Orlati, A.; Poppi, S.; Righini, S.

    2016-07-01

    The software maintenance typically requires 40-80% of the overall project costs, and this considerable variability mostly depends on the software internal quality: the more the software is designed and implemented to constantly welcome new changes, the lower will be the maintenance costs. The internal quality is typically enforced through testing, which in turn also affects the development and maintenance costs. This is the reason why testing methodologies have become a major concern for any company that builds - or is involved in building - software. Although there is no testing approach that suits all contexts, we infer some general guidelines learned during the Development of the Italian Single-dish COntrol System (DISCOS), which is a project aimed at producing the control software for the three INAF radio telescopes (the Medicina and Noto dishes, and the newly-built SRT). These guidelines concern both the development and the maintenance phases, and their ultimate goal is to maximize the DISCOS software quality through a Behavior-Driven Development (BDD) workflow beside a continuous delivery pipeline. We consider different topics and patterns; they involve the proper apportion of the tests (from end-to-end to low-level tests), the choice between hardware simulators and mockers, why and how to apply TDD and the dependency injection to increase the test coverage, the emerging technologies available for test isolation, bug fixing, how to protect the system from the external resources changes (firmware updating, hardware substitution, etc.) and, eventually, how to accomplish BDD starting from functional tests and going through integration and unit tests. We discuss pros and cons of each solution and point out the motivations of our choices either as a general rule or narrowed in the context of the DISCOS project.

  16. Software quality assurance plan for void fraction instrument

    International Nuclear Information System (INIS)

    Gimera, M.

    1994-01-01

    Waste Tank SY-101 has been the focus of extensive characterization work over the past few years. The waste continually generates gases, most notably hydrogen, which are periodically released from the waste. Gas can be trapped in tank waste in three forms: as void gas (bubbles), dissolved gas, or absorbed gas. Void fraction is the volume percentage of a given sample that is comprised of void gas. The void fraction instrument (VFI) acquires the data necessary to calculate void fraction. This document covers the product, Void Fraction Data Acquisition Software. The void fraction software being developed will have the ability to control the void fraction instrument hardware and acquire data necessary to calculate the void fraction in samples. This document provides the software quality assurance plan, verification and validation plan, and configuration management plan for developing the software for the instrumentation that will be used to obtain void fraction data from Tank SY-101

  17. 78 FR 16474 - Extension of the Period for Comments on the Enhancement of Quality of Software-Related Patents

    Science.gov (United States)

    2013-03-15

    ...] Extension of the Period for Comments on the Enhancement of Quality of Software-Related Patents AGENCY... announcing the formation of a partnership with the software community to enhance the quality of software... quality of software-related patents and the preparation of software-related patent applications including...

  18. The American Satellite Company (ASC) satellite deployed from payload bay

    Science.gov (United States)

    1985-01-01

    The American Satellite Company (ASC) communications satellite is deployed from the payload bay of the Shuttle Discovery. A portion of the cloudy surface of the earth can be seen to the left of the frame.

  19. Quality factors quantification/assurance for software related to safety in nuclear power plants

    International Nuclear Information System (INIS)

    Nunez McLeod, J.E.; Rivera, S.S.

    1997-01-01

    Quality assurance plan is needed to guarantee the software quality. The use of such a plan involves activities that should take place all along the life cycle, and which can be evaluated using the so called quality factors. This is due to the fact that the quality itself cannot be measured, but some of its manifestations can be used for this purpose. In the present work, a methodology to quantify a set of quality factors is proposed, for software based systems to be used in safety related areas in nuclear power plants. (author) [es

  20. Automated Theorem Proving in High-Quality Software Design

    Science.gov (United States)

    Schumann, Johann; Swanson, Keith (Technical Monitor)

    2001-01-01

    The amount and complexity of software developed during the last few years has increased tremendously. In particular, programs are being used more and more in embedded systems (from car-brakes to plant-control). Many of these applications are safety-relevant, i.e. a malfunction of hardware or software can cause severe damage or loss. Tremendous risks are typically present in the area of aviation, (nuclear) power plants or (chemical) plant control. Here, even small problems can lead to thousands of casualties and huge financial losses. Large financial risks also exist when computer systems are used in the area of telecommunication (telephone, electronic commerce) or space exploration. Computer applications in this area are not only subject to safety considerations, but also security issues are important. All these systems must be designed and developed to guarantee high quality with respect to safety and security. Even in an industrial setting which is (or at least should be) aware of the high requirements in Software Engineering, many incidents occur. For example, the Warshaw Airbus crash, was caused by an incomplete requirements specification. Uncontrolled reuse of an Ariane 4 software module was the reason for the Ariane 5 disaster. Some recent incidents in the telecommunication area, like illegal "cloning" of smart-cards of D2GSM handies, or the extraction of (secret) passwords from German T-online users show that also in this area serious flaws can happen. Due to the inherent complexity of computer systems, most authors claim that only a rigorous application of formal methods in all stages of the software life cycle can ensure high quality of the software and lead to real safe and secure systems. In this paper, we will have a look, in how far automated theorem proving can contribute to a more widespread application of formal methods and their tools, and what automated theorem provers (ATPs) must provide in order to be useful.

  1. FY17 ASC CSSE L2 Milestone 6018: Power Usage Characteristics of Workloads Running on Trinity.

    Energy Technology Data Exchange (ETDEWEB)

    Pedretti, Kevin [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-09-01

    The overall goal of this work was to utilize the Advanced Power Management (APM) capabilities of the ATS-1 Trinity platform to understand the power usage behavior of ASC workloads running on Trinity and gain insight into the potential for utilizing power management techniques on future ASC platforms.

  2. Intracellular invasion of Orientia tsutsugamushi activates inflammasome in asc-dependent manner.

    Directory of Open Access Journals (Sweden)

    Jung-Eun Koo

    Full Text Available Orientia tsutsugamushi, a causative agent of scrub typhus, is an obligate intracellular bacterium, which escapes from the endo/phagosome and replicates in the host cytoplasm. O. tsutsugamushi infection induces production of pro-inflammatory mediators including interleukin-1β (IL-1β, which is secreted mainly from macrophages upon cytosolic stimuli by activating cysteine protease caspase-1 within a complex called the inflammasome, and is a key player in initiating and maintaining the inflammatory response. However, the mechanism for IL-1β maturation upon O. tsutsugamushi infection has not been identified. In this study, we show that IL-1 receptor signaling is required for efficient host protection from O. tsutsugamushi infection. Live Orientia, but not heat- or UV-inactivated Orientia, activates the inflammasome through active bacterial uptake and endo/phagosomal maturation. Furthermore, Orientia-stimulated secretion of IL-1β and activation of caspase-1 are ASC- and caspase-1- dependent since IL-1β production was impaired in Asc- and caspase-1-deficient macrophages but not in Nlrp3-, Nlrc4- and Aim2-deficient macrophages. Therefore, live O. tsutsugamushi triggers ASC inflammasome activation leading to IL-1β production, which is a critical innate immune response for effective host defense.

  3. Quality of Service Attributes for Software as a Service

    Directory of Open Access Journals (Sweden)

    Lukas Burkon

    2013-07-01

    Full Text Available Software as a Service (SaaS has been developing for over ten years and, is reaching a mature level, where quality and its monitoring and management become significant. Although, SaaS is derived from the ASP model, SaaS background and architecture is different and therefore also SaaS quality management is based on different concepts. This paper is focused on the difference between traditional IT outsourcing and SaaS and proposes a set of quality attributes appropriate for the management of the SaaS quality.

  4. A Framework for Evaluating the Software Product Quality of Pregnancy Monitoring Mobile Personal Health Records.

    Science.gov (United States)

    Idri, Ali; Bachiri, Mariam; Fernández-Alemán, José Luis

    2016-03-01

    Stakeholders' needs and expectations are identified by means of software quality requirements, which have an impact on software product quality. In this paper, we present a set of requirements for mobile personal health records (mPHRs) for pregnancy monitoring, which have been extracted from literature and existing mobile apps on the market. We also use the ISO/IEC 25030 standard to suggest the requirements that should be considered during the quality evaluation of these mPHRs. We then go on to design a checklist in which we contrast the mPHRs for pregnancy monitoring requirements with software product quality characteristics and sub-characteristics in order to calculate the impact of these requirements on software product quality, using the ISO/IEC 25010 software product quality standard. The results obtained show that the requirements related to the user's actions and the app's features have the most impact on the external sub-characteristics of the software product quality model. The only sub-characteristic affected by all the requirements is Appropriateness of Functional suitability. The characteristic Operability is affected by 95% of the requirements while the lowest degrees of impact were identified for Compatibility (15%) and Transferability (6%). Lastly, the degrees of the impact of the mPHRs for pregnancy monitoring requirements are discussed in order to provide appropriate recommendations for the developers and stakeholders of mPHRs for pregnancy monitoring.

  5. Software support for environmental measurement in quality at educational institutions

    Directory of Open Access Journals (Sweden)

    Alena Pauliková

    2016-03-01

    Full Text Available The analysed theme of this article is based on the training of environmental measurements for workplaces. This is very important for sustainable quality in technical educational institutions. Applied kinds of software, which are taught at technical educational institutions, have to offer the professional and methodical knowledge concerning conditions of working ambient for students of selected technical specialisations. This skill is performed in such a way that the graduates, after entering the practical professional life, will be able to participate in solutions for actual problems that are related to environmental protection by means of software support. Nowadays, during the training processit is also obligatory to introduce technical science. Taking into consideration the above-mentioned facts it is possible to say that information technology support for environmental study subjects is a relevant aspect, which should be integrated into the university educational process. There is an effective progress that further highlights the focus on the quality of university education not only for environmental engineers. Actual trends require an increasing number of software/hardware educated engineers who can participate in qualitative university preparation, i.e.IT environmentalists. The Department of Environmental Engineering at the Faculty of Mechanical Engineering, TechnicalUniversity in Košice, Slovakia is an institution specified and intended for quality objectivisation. This institution introduced into the study programmes (“Environmental Management” and “Technology of Environmental Protection” study subjects with the software support, which are oriented towards outdoor and indoor ambient and in this way the Department of Process and Environmental Engineering is integrated effectively and intensively into the area of measurement training with regard to the requirement of quality educational processes.

  6. Probability-Based Design Criteria of the ASCE 7 Tsunami Loads and Effects Provisions (Invited)

    Science.gov (United States)

    Chock, G.

    2013-12-01

    Mitigation of tsunami risk requires a combination of emergency preparedness for evacuation in addition to providing structural resilience of critical facilities, infrastructure, and key resources necessary for immediate response and economic and social recovery. Critical facilities would include emergency response, medical, tsunami refuges and shelters, ports and harbors, lifelines, transportation, telecommunications, power, financial institutions, and major industrial/commercial facilities. The Tsunami Loads and Effects Subcommittee of the ASCE/SEI 7 Standards Committee is developing a proposed new Chapter 6 - Tsunami Loads and Effects for the 2016 edition of the ASCE 7 Standard. ASCE 7 provides the minimum design loads and requirements for structures subject to building codes such as the International Building Code utilized in the USA. In this paper we will provide a review emphasizing the intent of these new code provisions and explain the design methodology. The ASCE 7 provisions for Tsunami Loads and Effects enables a set of analysis and design methodologies that are consistent with performance-based engineering based on probabilistic criteria. . The ASCE 7 Tsunami Loads and Effects chapter will be initially applicable only to the states of Alaska, Washington, Oregon, California, and Hawaii. Ground shaking effects and subsidence from a preceding local offshore Maximum Considered Earthquake will also be considered prior to tsunami arrival for Alaska and states in the Pacific Northwest regions governed by nearby offshore subduction earthquakes. For national tsunami design provisions to achieve a consistent reliability standard of structural performance for community resilience, a new generation of tsunami inundation hazard maps for design is required. The lesson of recent tsunami is that historical records alone do not provide a sufficient measure of the potential heights of future tsunamis. Engineering design must consider the occurrence of events greater than

  7. A systematic literature review of open source software quality assessment models.

    Science.gov (United States)

    Adewumi, Adewole; Misra, Sanjay; Omoregbe, Nicholas; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    Many open source software (OSS) quality assessment models are proposed and available in the literature. However, there is little or no adoption of these models in practice. In order to guide the formulation of newer models so they can be acceptable by practitioners, there is need for clear discrimination of the existing models based on their specific properties. Based on this, the aim of this study is to perform a systematic literature review to investigate the properties of the existing OSS quality assessment models by classifying them with respect to their quality characteristics, the methodology they use for assessment, and their domain of application so as to guide the formulation and development of newer models. Searches in IEEE Xplore, ACM, Science Direct, Springer and Google Search is performed so as to retrieve all relevant primary studies in this regard. Journal and conference papers between the year 2003 and 2015 were considered since the first known OSS quality model emerged in 2003. A total of 19 OSS quality assessment model papers were selected. To select these models we have developed assessment criteria to evaluate the quality of the existing studies. Quality assessment models are classified into five categories based on the quality characteristics they possess namely: single-attribute, rounded category, community-only attribute, non-community attribute as well as the non-quality in use models. Our study reflects that software selection based on hierarchical structures is found to be the most popular selection method in the existing OSS quality assessment models. Furthermore, we found that majority (47%) of the existing models do not specify any domain of application. In conclusion, our study will be a valuable contribution to the community and helps the quality assessment model developers in formulating newer models and also to the practitioners (software evaluators) in selecting suitable OSS in the midst of alternatives.

  8. Quality of Service Attributes for Software as a Service

    OpenAIRE

    Lukas Burkon

    2013-01-01

    Software as a Service (SaaS) has been developing for over ten years and, is reaching a mature level, where quality and its monitoring and management become significant. Although, SaaS is derived from the ASP model, SaaS background and architecture is different and therefore also SaaS quality management is based on different concepts. This paper is focused on the difference between traditional IT outsourcing and SaaS and proposes a set of quality attributes appropriate for the management of th...

  9. Analyst Tools and Quality Control Software for the ARM Data System

    Energy Technology Data Exchange (ETDEWEB)

    Moore, S.T.

    2004-12-14

    ATK Mission Research develops analyst tools and automated quality control software in order to assist the Atmospheric Radiation Measurement (ARM) Data Quality Office with their data inspection tasks. We have developed a web-based data analysis and visualization tool, called NCVweb, that allows for easy viewing of ARM NetCDF files. NCVweb, along with our library of sharable Interactive Data Language procedures and functions, allows even novice ARM researchers to be productive with ARM data with only minimal effort. We also contribute to the ARM Data Quality Office by analyzing ARM data streams, developing new quality control metrics, new diagnostic plots, and integrating this information into DQ HandS - the Data Quality Health and Status web-based explorer. We have developed several ways to detect outliers in ARM data streams and have written software to run in an automated fashion to flag these outliers.

  10. The role of colposcopy and typization of human papillomavirus in further diagnostic proceedings in patients with ASC-US cytological finding of the uterine cervix

    Directory of Open Access Journals (Sweden)

    Živadinović Radomir

    2009-01-01

    Full Text Available Background/Aim. Bethesda system of classification of cytological findings was introduced in 2001 two subcategories in the category of atypical squamous cells (ASC findings: ASC of undetermined significance (ASC-US and ASC which cannot exclude high-grade intraepithelial lesions (ASC-H. The aim of our study was to assess a possible association of these two subcategories with pathologic biopsy finding and to find out the best further diagnostic proceedings. Methods. At the Clinic of Gynecology and Obstetrics, Niš 130 patients with ASC findings were analyzed. Colposcopy was performed in all study participants. Patients with pathological colposcopic findings underwent cervical biopsy. In 10 patients with pathologic histologic and 15 with benign findings human papilloma virus (HPV typization was done using the Hybrid Capture method. Results. Patients with ASC-H finding had significantly more pathologic biopsies compared with patients with ASC-US finding (57.84: 20.72. Conclusion. Colposcopy was exhibited somewhat higher sensitivity compared to HPV typization (94.7 : 90, but lower sensitivity (79.27 : 86.6. The usage of HPV typization in the triage of patients with ASC cytologic smear induces statistically significant reduction of unnecessary percentage of cervical biopsies.

  11. Antidepressant-Like Effects of Central BDNF Administration in Mice of Antidepressant Sensitive Catalepsy (ASC) Strain.

    Science.gov (United States)

    Tikhonova, Maria; Kulikov, Alexander V

    2012-08-31

    Although numerous data evidence the implication of brain-derived neurotrophic factor (BDNF) in the pathophysiology of depression, the potential for BDNF to correct genetically defined depressive-like states is poorly studied. This study was aimed to reveal antidepressant-like effects of BDNF (300 ng, 2×, i.c.v.) on behavior and mRNA expression of genes associated with depression-like state in the brain in mice of antidepressant sensitive catalepsy (ASC) strain characterized by high hereditary predisposition to catalepsy and depressive-like features. Behavioral tests were held on the 7th-16th days after the first (4th-13th after the second) BDNF injection. Results showed that BDNF normalized impaired sexual motivation in the ASC males, and this BDNF effect differed, with advantageous effects, from that of widely used antidepressants. The anticataleptic effect of two BDNF injections was enhanced compared with a single administration. A tendency to decrease the immobility duration in tail-suspension test was observed in BDNF-treated ASC mice. The effects on catalepsy and sexual motivation were specific since BDNF did not alter locomotor and exploratory activity or social interest in the ASC mice. Along with behavioral antidepressant-like effects on the ASC mice, BDNF increased hippocampal mRNA levels of Bdnf and Creb1 (cAMP response element-binding protein gene). BDNF also augmented mRNA levels of Arc gene encoding Arc (Activity-regulated cytoskeleton-associated) protein involved in BDNF-induced processes of neuronal and synaptic plasticity in hippocampus and prefrontal cortex. The data suggest that: [1] BDNF is effective in the treatment of some genetically defined behavioral disturbances; [2] BDNF influences sexually-motivated behavior; [3] Arc mRNA levels may serve as a molecular marker of BDNF physiological activity associated with its long-lasting behavioral effects; [4] ASC mouse strain can be used as a suitable model to study mechanisms of BDNF effects on

  12. [Software for illustrating a cost-quality balance carried out by clinical laboratory practice].

    Science.gov (United States)

    Nishibori, Masahiro; Asayama, Hitoshi; Kimura, Satoshi; Takagi, Yasushi; Hagihara, Michio; Fujiwara, Mutsunori; Yoneyama, Akiko; Watanabe, Takashi

    2010-09-01

    We have no proper reference indicating the quality of clinical laboratory practice, which should clearly illustrates that better medical tests require more expenses. Japanese Society of Laboratory Medicine was concerned about recent difficult medical economy and issued a committee report proposing a guideline to evaluate the good laboratory practice. According to the guideline, we developed software that illustrate a cost-quality balance carried out by clinical laboratory practice. We encountered a number of controversial problems, for example, how to measure and weight each quality-related factor, how to calculate costs of a laboratory test and how to consider characteristics of a clinical laboratory. Consequently we finished only prototype software within the given period and the budget. In this paper, software implementation of the guideline and the above-mentioned problems are summarized. Aiming to stimulate these discussions, the operative software will be put on the Society's homepage for trial

  13. A pioneering application of NQA-1 quality assurance standards in the development of software

    International Nuclear Information System (INIS)

    Weisbin, A.N.

    1988-01-01

    One reason for systematically applying quality assurance to computer software is the extensive use of results from computer programs to characterize potential sited for nuclear waste repositories leading ultimately to important policy making decisions. Because data from these programs characterize the likely radioactivity profile for many hundreds of years, experimental validation is not feasible. The Sensitivity and Uncertainty Analysis Methods Development Project (SUAMDP) was developed to formulate and utilize efficient and comprehensive methods for determining sensitivities of calculated results with respect to changes in all input parameters. The computerized methodology was embodied in the Gradient Enhanced Software System (GRESS). Due to the fact that GRESS was to be used in the site characterization for waste storage, stringent NQA-1 requirements were imposed by the sponsor. A working relationship between the Oak Ridge National Laboratory (ORNL) Quality Department and the research scientists developing GRESS was essential in achieving understanding and acceptance of the quality assurance requirements as applied to the SUAMDP. The relationship resulted in the SUAMDP becoming the first software project at ORNL to develop a comprehensive NQA-1 Quality Assurance Plan; this plan now serves as a model for software quality assurance at ORNL. This paper describes the evolution of this plan and its impact on the application of quality assurance procedures to software. 2 refs

  14. ASC Trilab L2 Codesign Milestone 2015

    Energy Technology Data Exchange (ETDEWEB)

    Trott, Christian Robert [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Hammond, Simon David [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Dinge, Dennis [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Lin, Paul T. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Vaughan, Courtenay T. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Cook, Jeanine [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Edwards, Harold C. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Rajan, Mahesh [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Hoekstra, Robert J. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-09-01

    For the FY15 ASC L2 Trilab Codesign milestone Sandia National Laboratories performed two main studies. The first study investigated three topics (performance, cross-platform portability and programmer productivity) when using OpenMP directives and the RAJA and Kokkos programming models available from LLNL and SNL respectively. The focus of this first study was the LULESH mini-application developed and maintained by LLNL. In the coming sections of the report the reader will find performance comparisons (and a demonstration of portability) for a variety of mini-application implementations produced during this study with varying levels of optimization. Of note is that the implementations utilized including optimizations across a number of programming models to help ensure claims that Kokkos can provide native-class application performance are valid. The second study performed during FY15 is a performance assessment of the MiniAero mini-application developed by Sandia. This mini-application was developed by the SIERRA Thermal-Fluid team at Sandia for the purposes of learning the Kokkos programming model and so is available in only a single implementation. For this report we studied its performance and scaling on a number of machines with the intent of providing insight into potential performance issues that may be experienced when similar algorithms are deployed on the forthcoming Trinity ASC ATS platform.

  15. Guidance and Control Software Project Data - Volume 4: Configuration Management and Quality Assurance Documents

    Science.gov (United States)

    Hayhurst, Kelly J. (Editor)

    2008-01-01

    The Guidance and Control Software (GCS) project was the last in a series of software reliability studies conducted at Langley Research Center between 1977 and 1994. The technical results of the GCS project were recorded after the experiment was completed. Some of the support documentation produced as part of the experiment, however, is serving an unexpected role far beyond its original project context. Some of the software used as part of the GCS project was developed to conform to the RTCA/DO-178B software standard, "Software Considerations in Airborne Systems and Equipment Certification," used in the civil aviation industry. That standard requires extensive documentation throughout the software development life cycle, including plans, software requirements, design and source code, verification cases and results, and configuration management and quality control data. The project documentation that includes this information is open for public scrutiny without the legal or safety implications associated with comparable data from an avionics manufacturer. This public availability has afforded an opportunity to use the GCS project documents for DO-178B training. This report provides a brief overview of the GCS project, describes the 4-volume set of documents and the role they are playing in training, and includes configuration management and quality assurance documents from the GCS project. Volume 4 contains six appendices: A. Software Accomplishment Summary for the Guidance and Control Software Project; B. Software Configuration Index for the Guidance and Control Software Project; C. Configuration Management Records for the Guidance and Control Software Project; D. Software Quality Assurance Records for the Guidance and Control Software Project; E. Problem Report for the Pluto Implementation of the Guidance and Control Software Project; and F. Support Documentation Change Reports for the Guidance and Control Software Project.

  16. NLRP3 and ASC suppress lupus-like autoimmunity by driving the immunosuppressive effects of TGF-β receptor signalling.

    Science.gov (United States)

    Lech, Maciej; Lorenz, Georg; Kulkarni, Onkar P; Grosser, Marian O O; Stigrot, Nora; Darisipudi, Murthy N; Günthner, Roman; Wintergerst, Maximilian W M; Anz, David; Susanti, Heni Eka; Anders, Hans-Joachim

    2015-12-01

    The NLRP3/ASC inflammasome drives host defence and autoinflammatory disorders by activating caspase-1 to trigger the secretion of mature interleukin (IL)-1β/IL-18, but its potential role in autoimmunity is speculative. We generated and phenotyped Nlrp3-deficient, Asc-deficient, Il-1r-deficient and Il-18-deficient C57BL/6-lpr/lpr mice, the latter being a mild model of spontaneous lupus-like autoimmunity. While lack of IL-1R or IL-18 did not affect the C57BL/6-lpr/lpr phenotype, lack of NLRP3 or ASC triggered massive lymphoproliferation, lung T cell infiltrates and severe proliferative lupus nephritis within 6 months, which were all absent in age-matched C57BL/6-lpr/lpr controls. Lack of NLRP3 or ASC increased dendritic cell and macrophage activation, the expression of numerous proinflammatory mediators, lymphocyte necrosis and the expansion of most T cell and B cell subsets. In contrast, plasma cells and autoantibody production were hardly affected. This unexpected immunosuppressive effect of NLRP3 and ASC may relate to their known role in SMAD2/3 phosphorylation during tumour growth factor (TGF)-β receptor signalling, for example, Nlrp3-deficiency and Asc-deficiency significantly suppressed the expression of numerous TGF-β target genes in C57BL/6-lpr/lpr mice and partially recapitulated the known autoimmune phenotype of Tgf-β1-deficient mice. These data identify a novel non-canonical immunoregulatory function of NLRP3 and ASC in autoimmunity. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/

  17. EARLY IDENTIFICATION OF SOFTWARE DEFECTS USING OCL PREDICATES TO IMPROVE SOFTWARE QUALITY

    Directory of Open Access Journals (Sweden)

    A. JALILA

    2015-03-01

    Full Text Available Formal specification-based testing has been used widely to assess potential faults or prove their absence in a given system at the earliest. This research work has proposed an automated fault-based testing framework to test the specification of the system using Object Constraint Language (OCL. Accordingly, the possible faults in OCL specification of the system has been anticipated by mutating its method based on OCL predicate-based fault classes. Then, test cases are generated using Genetic Algorithm with simulated annealing technique. In this paper, a novel OCL-predicate based fitness function is defined to evaluate the generated test data. Finally, this paper presents the experimental results, which indicate that the proposed methodology provides more test coverage with the reduced test suite and test run. This results in cost-effective software development so as to improve software quality.

  18. Measuring the impact of computer resource quality on the software development process and product

    Science.gov (United States)

    Mcgarry, Frank; Valett, Jon; Hall, Dana

    1985-01-01

    The availability and quality of computer resources during the software development process was speculated to have measurable, significant impact on the efficiency of the development process and the quality of the resulting product. Environment components such as the types of tools, machine responsiveness, and quantity of direct access storage may play a major role in the effort to produce the product and in its subsequent quality as measured by factors such as reliability and ease of maintenance. During the past six years, the NASA Goddard Space Flight Center has conducted experiments with software projects in an attempt to better understand the impact of software development methodologies, environments, and general technologies on the software process and product. Data was extracted and examined from nearly 50 software development projects. All were related to support of satellite flight dynamics ground-based computations. The relationship between computer resources and the software development process and product as exemplified by the subject NASA data was examined. Based upon the results, a number of computer resource-related implications are provided.

  19. ASC Tri-lab Co-design Level 2 Milestone Report 2015

    Energy Technology Data Exchange (ETDEWEB)

    Hornung, Rich [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Jones, Holger [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Keasler, Jeff [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Neely, Rob [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Pearce, Olga [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Hammond, Si [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Trott, Christian [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Lin, Paul [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Vaughan, Courtenay [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Cook, Jeanine [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Hoekstra, Rob [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Bergen, Ben [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Payne, Josh [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Womeldorff, Geoff [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2015-09-23

    In 2015, the three Department of Energy (DOE) National Laboratories that make up the Advanced Sci- enti c Computing (ASC) Program (Sandia, Lawrence Livermore, and Los Alamos) collaboratively explored performance portability programming environments in the context of several ASC co-design proxy applica- tions as part of a tri-lab L2 milestone executed by the co-design teams at each laboratory. The programming environments that were studied included Kokkos (developed at Sandia), RAJA (LLNL), and Legion (Stan- ford University). The proxy apps studied included: miniAero, LULESH, CoMD, Kripke, and SNAP. These programming models and proxy-apps are described herein. Each lab focused on a particular combination of abstractions and proxy apps, with the goal of assessing performance portability using those. Performance portability was determined by: a) the ability to run a single application source code on multiple advanced architectures, b) comparing runtime performance between \

  20. Evaluation of Software Quality to Improve Application Performance Using Mc Call Model

    Directory of Open Access Journals (Sweden)

    Inda D Lestantri

    2018-04-01

    Full Text Available The existence of software should have more value to improve the performance of the organization in addition to having the primary function to automate. Before being implemented in an operational environment, software must pass the test gradually to ensure that the software is functioning properly, meeting user needs and providing convenience for users to use it. This test is performed on a web-based application, by taking a test case in an e-SAP application. E-SAP is an application used to monitor teaching and learning activities used by a university in Jakarta. To measure software quality, testing can be done on users randomly. The user samples selected in this test are users with an age range of 18 years old up to 25 years, background information technology. This test was conducted on 30 respondents. This test is done by using Mc Call model. Model of testing Mc Call consists of 11 dimensions are grouped into 3 categories. This paper describes the testing with reference to the category of product operation, which includes 5 dimensions. The dimensions of testing performed include the dimensions of correctness, usability, efficiency, reliability, and integrity. This paper discusses testing on each dimension to measure software quality as an effort to improve performance. The result of research is e-SAP application has good quality with product operation value equal to 85.09%. This indicates that the e-SAP application has a great quality, so this application deserves to be examined in the next stage on the operational environment.

  1. Quality assurance requirements for the computer software and safety analyses

    International Nuclear Information System (INIS)

    Husarecek, J.

    1992-01-01

    The requirements are given as placed on the development, procurement, maintenance, and application of software for the creation or processing of data during the design, construction, operation, repair, maintenance and safety-related upgrading of nuclear power plants. The verification and validation processes are highlighted, and the requirements put on the software documentation are outlined. The general quality assurance principles applied to safety analyses are characterized. (J.B.). 1 ref

  2. Aeroacoustics research in Europe : the CEAS-ASC report on 2007 highlights

    NARCIS (Netherlands)

    Brouwer, H.H.; Rienstra, S.W.

    2008-01-01

    The Council of European Aerospace Societies (CEAS) Aeroacoustics Specialists Committee (ASC) supports and promotes the interests of the scientific and industrial aeroacoustics community on a European scale and European aeronautics activities internationally. In this context, "aeroacoustics"

  3. A Discussion of the Software Quality Assurance Role

    Science.gov (United States)

    Kandt, Ronald Kirk

    2010-01-01

    The basic idea underlying this paper is that the conventional understanding of the role of a Software Quality Assurance (SQA) engineer is unduly limited. This is because few have asked who the customers of a SQA engineer are. Once you do this, you can better define what tasks a SQA engineer should perform, as well as identify the knowledge and skills that such a person should have. The consequence of doing this is that a SQA engineer can provide greater value to his or her customers. It is the position of this paper that a SQA engineer providing significant value to his or her customers must not only assume the role of an auditor, but also that of a software and systems engineer. This is because software engineers and their managers particularly value contributions that directly impact products and their development. These ideas are summarized as lessons learned, based on my experience at Jet Propulsion Laboratory (JPL).

  4. The influence of software filtering in digital mammography image quality

    Science.gov (United States)

    Michail, C.; Spyropoulou, V.; Kalyvas, N.; Valais, I.; Dimitropoulos, N.; Fountos, G.; Kandarakis, I.; Panayiotakis, G.

    2009-05-01

    Breast cancer is one of the most frequently diagnosed cancers among women. Several techniques have been developed to help in the early detection of breast cancer such as conventional and digital x-ray mammography, positron and single-photon emission mammography, etc. A key advantage in digital mammography is that images can be manipulated as simple computer image files. Thus non-dedicated commercially available image manipulation software can be employed to process and store the images. The image processing tools of the Photoshop (CS 2) software usually incorporate digital filters which may be used to reduce image noise, enhance contrast and increase spatial resolution. However, improving an image quality parameter may result in degradation of another. The aim of this work was to investigate the influence of three sharpening filters, named hereafter sharpen, sharpen more and sharpen edges on image resolution and noise. Image resolution was assessed by means of the Modulation Transfer Function (MTF).In conclusion it was found that the correct use of commercial non-dedicated software on digital mammograms may improve some aspects of image quality.

  5. The influence of software filtering in digital mammography image quality

    International Nuclear Information System (INIS)

    Michail, C; Spyropoulou, V; Valais, I; Panayiotakis, G; Kalyvas, N; Fountos, G; Kandarakis, I; Dimitropoulos, N

    2009-01-01

    Breast cancer is one of the most frequently diagnosed cancers among women. Several techniques have been developed to help in the early detection of breast cancer such as conventional and digital x-ray mammography, positron and single-photon emission mammography, etc. A key advantage in digital mammography is that images can be manipulated as simple computer image files. Thus non-dedicated commercially available image manipulation software can be employed to process and store the images. The image processing tools of the Photoshop (CS 2) software usually incorporate digital filters which may be used to reduce image noise, enhance contrast and increase spatial resolution. However, improving an image quality parameter may result in degradation of another. The aim of this work was to investigate the influence of three sharpening filters, named hereafter sharpen, sharpen more and sharpen edges on image resolution and noise. Image resolution was assessed by means of the Modulation Transfer Function (MTF).In conclusion it was found that the correct use of commercial non-dedicated software on digital mammograms may improve some aspects of image quality.

  6. Interface-based software testing

    Directory of Open Access Journals (Sweden)

    Aziz Ahmad Rais

    2016-10-01

    Full Text Available Software quality is determined by assessing the characteristics that specify how it should work, which are verified through testing. If it were possible to touch, see, or measure software, it would be easier to analyze and prove its quality. Unfortunately, software is an intangible asset, which makes testing complex. This is especially true when software quality is not a question of particular functions that can be tested through a graphical user interface. The primary objective of software architecture is to design quality of software through modeling and visualization. There are many methods and standards that define how to control and manage quality. However, many IT software development projects still fail due to the difficulties involved in measuring, controlling, and managing software quality. Software quality failure factors are numerous. Examples include beginning to test software too late in the development process, or failing properly to understand, or design, the software architecture and the software component structure. The goal of this article is to provide an interface-based software testing technique that better measures software quality, automates software quality testing, encourages early testing, and increases the software’s overall testability

  7. Characterization of a novel variant of amino acid transport system asc in erythrocytes from Przewalski's horse (Equus przewalskii).

    Science.gov (United States)

    Fincham, D A; Ellory, J C; Young, J D

    1992-08-01

    In thoroughbred horses, red blood cell amino acid transport activity is Na(+)-independent and controlled by three codominant genetic alleles (h, l, s), coding for high-affinity system asc1 (L-alanine apparent Km for influx at 37 degrees C congruent to 0.35 mM), low-affinity system asc2 (L-alanine Km congruent to 14 mM), and transport deficiency, respectively. The present study investigated amino acid transport mechanisms in red cells from four wild species: Przewalski's horse (Equus przewalskii), Hartmann's zebra (Zebra hartmannae), Grevy's zebra (Zebra grevyi), and onager (Equus hemonius). Red blood cell samples from different Przewalski's horses exhibited uniformly high rates of L-alanine uptake, mediated by a high-affinity asc1-type transport system. Mean apparent Km and Vmax values (+/- SE) for L-alanine influx at 37 degrees C in red cells from 10 individual animals were 0.373 +/- 0.068 mM and 2.27 +/- 0.11 mmol (L cells.h), respectively. As in thoroughbreds, the Przewalski's horse transporter interacted with dibasic as well as neutral amino acids. However, the Przewalski asc1 isoform transported L-lysine with a substantially (6.4-fold) higher apparent affinity than its thoroughbred counterpart (Km for influx 1.4 mM at 37 degrees C) and was also less prone to trans-stimulation effects. The novel high apparent affinity of the Przewalski's horse transporter for L-lysine provides additional key evidence of functional and possible structural similarities between asc and the classical Na(+)-dependent system ASC and between these systems and the Na(+)-independent dibasic amino acid transport system y+. Unlike Przewalski's horse, zebra red cells were polymorphic with respect to L-alanine transport activity, showing high-affinity or low-affinity saturable mechanisms of L-alanine uptake. Onager red cells transported this amino acid with intermediate affinity (apparent Km for influx 3.0 mM at 37 degrees C). Radiation inactivation analysis was used to estimate the target

  8. Software Quality Assurance and Verification for the MPACT Library Generation Process

    Energy Technology Data Exchange (ETDEWEB)

    Liu, Yuxuan [Univ. of Michigan, Ann Arbor, MI (United States); Williams, Mark L. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Wiarda, Dorothea [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Clarno, Kevin T. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Kim, Kang Seog [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Celik, Cihangir [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2017-05-01

    This report fulfills the requirements for the Consortium for the Advanced Simulation of Light-Water Reactors (CASL) milestone L2:RTM.P14.02, “SQA and Verification for MPACT Library Generation,” by documenting the current status of the software quality, verification, and acceptance testing of nuclear data libraries for MPACT. It provides a brief overview of the library generation process, from general-purpose evaluated nuclear data files (ENDF/B) to a problem-dependent cross section library for modeling of light-water reactors (LWRs). The software quality assurance (SQA) programs associated with each of the software used to generate the nuclear data libraries are discussed; specific tests within the SCALE/AMPX and VERA/XSTools repositories are described. The methods and associated tests to verify the quality of the library during the generation process are described in detail. The library generation process has been automated to a degree to (1) ensure that it can be run without user intervention and (2) to ensure that the library can be reproduced. Finally, the acceptance testing process that will be performed by representatives from the Radiation Transport Methods (RTM) Focus Area prior to the production library’s release is described in detail.

  9. Factors Affecting Innovation Within Aeronautical Systems Center (ASC) Organizations - An inductive Study

    National Research Council Canada - National Science Library

    Feil, Eric

    2003-01-01

    .... This thesis analyzed data collected during the 2002 Chief of Staff of the Air Force Organizational Climate Survey to identify factors that affect innovation within Aeronautical Systems Center (ASC) organizations...

  10. Software Configuration Management: The Quality Weakness

    International Nuclear Information System (INIS)

    Arrojo, E.; Garcia, P.

    1998-01-01

    At the moment it is very difficult to din any process in the industry where software is not involved. We trust software does minimize the possibility of process failures. In parallel, the quality and safety requirements of our processes have been improved to satisfactory levels. Let's look around us. Every day, thousands of calculations are carried out by our engineers using computer programs. Hundreds of processes are controlled automatically. Safety marging, limits, operation controls..., are derived from them. The tools begin to control our processes but, Who does control the tool? Once they have been installed and once they are running, are they always reliable? NO If you think that your current system are satisfactory, we propose you a game in this report. It is just a test. Which is your score?. Then we revise the concept of Configuration Management and we describe an ideal machine; the ''Perpetuum Mobile'' of the Configuration. We describe some rules to implement and improvement and we comment on the operative experience in ENUSA. (Author)

  11. Profiling Autism Symptomatology: An Exploration of the Q-ASC Parental Report Scale in Capturing Sex Differences in Autism

    Science.gov (United States)

    Ormond, Sarah; Brownlow, Charlotte; Garnett, Michelle Sarah; Rynkiewicz, Agnieszka; Attwood, Tony

    2018-01-01

    The Questionnaire for Autism Spectrum Conditions (Q-ASC) was developed by Attwood et al. (2011) to identify gender-sensitive profiles of autism symptomatology; prioritise and adjust the direction of clinical interventions; and support positive psychosocial outcomes and prognosis into adulthood. The current research piloted the Q-ASC with parents…

  12. Aeroacoustics research in Europe : the CEAS-ASC report on 1997 highlights

    NARCIS (Netherlands)

    Rienstra, S.W.

    1998-01-01

    This paper is a report on the highlights of aeroacoustics research and development in Europe in 1997, compiled from information provided in the CEAS Aeroacoustics Specialists Committee (ASC). The Confederation of European Aerospace Societies (CEAS) comprises the national Aerospace Societies of

  13. Lightweight and Continuous Architectural Software Quality Assurance using the aSQA Technique

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak; Hansen, Klaus Marius; Lindstrøm, Bo

    2010-01-01

    In this paper, we present a novel technique for assessing and prioritizing architectural quality in large-scale software development projects. The technique can be applied with relatively little effort by software architects and thus suited for agile development in which quality attributes can...... be assessed and prioritized, e.g., within each development sprint. We outline the processes and metrics embodied in the technique, and report initial experiences on the benefits and liabilities. In conclusion, the technique is considered valuable and a viable tool, and has benefits in an architectural...

  14. Human adipose stem cell and ASC-derived cardiac progenitor cellular therapy improves outcomes in a murine model of myocardial infarction

    Directory of Open Access Journals (Sweden)

    Davy PMC

    2015-10-01

    Full Text Available Philip MC Davy,1 Kevin D Lye,2,3 Juanita Mathews,1 Jesse B Owens,1 Alice Y Chow,1 Livingston Wong,2 Stefan Moisyadi,1 Richard C Allsopp1 1Institute for Biogenesis Research, 2John A. Burns School of Medicine, University of Hawaii at Mānoa, 3Tissue Genesis, Inc., Honolulu, HI, USA Background: Adipose tissue is an abundant and potent source of adult stem cells for transplant therapy. In this study, we present our findings on the potential application of adipose-derived stem cells (ASCs as well as induced cardiac-like progenitors (iCPs derived from ASCs for the treatment of myocardial infarction. Methods and results: Human bone marrow (BM-derived stem cells, ASCs, and iCPs generated from ASCs using three defined cardiac lineage transcription factors were assessed in an immune-compromised mouse myocardial infarction model. Analysis of iCP prior to transplant confirmed changes in gene and protein expression consistent with a cardiac phenotype. Endpoint analysis was performed 1 month posttransplant. Significantly increased endpoint fractional shortening, as well as reduction in the infarct area at risk, was observed in recipients of iCPs as compared to the other recipient cohorts. Both recipients of iCPs and ASCs presented higher myocardial capillary densities than either recipients of BM-derived stem cells or the control cohort. Furthermore, mice receiving iCPs had a significantly higher cardiac retention of transplanted cells than all other groups. Conclusion: Overall, iCPs generated from ASCs outperform BM-derived stem cells and ASCs in facilitating recovery from induced myocardial infarction in mice. Keywords: adipose stem cells, myocardial infarction, cellular reprogramming, cellular therapy, piggyBac, induced cardiac-like progenitors

  15. Software as quality product

    International Nuclear Information System (INIS)

    Enders, A.

    1975-01-01

    In many discussions on the reliability of computer systems, software is presented as the weak link in the chain. The contribution attempts to identify the reasons for this situation as seen from the software development. The concepts correctness and reliability of programmes are explained as they are understood in the specialist discussion of today. Measures and methods are discussed which are particularly relevant as far as the obtaining of fault-free and reliable programmes is concerned. Conclusions are drawn for the user of software so that he is in the position to judge himself what can be justly expected frm the product software compared to other products. (orig./LH) [de

  16. UrQt: an efficient software for the Unsupervised Quality trimming of NGS data.

    Science.gov (United States)

    Modolo, Laurent; Lerat, Emmanuelle

    2015-04-29

    Quality control is a necessary step of any Next Generation Sequencing analysis. Although customary, this step still requires manual interventions to empirically choose tuning parameters according to various quality statistics. Moreover, current quality control procedures that provide a "good quality" data set, are not optimal and discard many informative nucleotides. To address these drawbacks, we present a new quality control method, implemented in UrQt software, for Unsupervised Quality trimming of Next Generation Sequencing reads. Our trimming procedure relies on a well-defined probabilistic framework to detect the best segmentation between two segments of unreliable nucleotides, framing a segment of informative nucleotides. Our software only requires one user-friendly parameter to define the minimal quality threshold (phred score) to consider a nucleotide to be informative, which is independent of both the experiment and the quality of the data. This procedure is implemented in C++ in an efficient and parallelized software with a low memory footprint. We tested the performances of UrQt compared to the best-known trimming programs, on seven RNA and DNA sequencing experiments and demonstrated its optimality in the resulting tradeoff between the number of trimmed nucleotides and the quality objective. By finding the best segmentation to delimit a segment of good quality nucleotides, UrQt greatly increases the number of reads and of nucleotides that can be retained for a given quality objective. UrQt source files, binary executables for different operating systems and documentation are freely available (under the GPLv3) at the following address: https://lbbe.univ-lyon1.fr/-UrQt-.html .

  17. IEEE [Institute of Electrical and Electronics Engineers] standards and nuclear software quality engineering

    International Nuclear Information System (INIS)

    Daughtrey, T.

    1988-01-01

    Significant new nuclear-specific software standards have recently been adopted under the sponsorship of the American Nuclear Society and the American Society of Mechanical Engineers. The interest of the US Nuclear Regulatory Commission has also been expressed through their issuance of NUREG/CR-4640. These efforts all indicate a growing awareness of the need for thorough, referenceable expressions of the way to build in and evaluate quality in nuclear software. A broader professional perspective can be seen in the growing number of software engineering standards sponsored by the Institute of Electrical and Electronics Engineers (IEEE) Computer Society. This family of standards represents a systematic effort to capture professional consensus on quality practices throughout the software development life cycle. The only omission-the implementation phase-is treated by accepted American National Standards Institute or de facto standards for programming languages

  18. ARCHITECTURE SOFTWARE SOLUTION TO SUPPORT AND DOCUMENT MANAGEMENT QUALITY SYSTEM

    Directory of Open Access Journals (Sweden)

    Milan Eric

    2010-12-01

    Full Text Available One of the basis of a series of standards JUS ISO 9000 is quality system documentation. An architecture of the quality system documentation depends on the complexity of business system. An establishment of an efficient management documentation of system of quality is of a great importance for the business system, as well as in the phase of introducing the quality system and in further stages of its improvement. The study describes the architecture and capability of software solutions to support and manage the quality system documentation in accordance with the requirements of standards ISO 9001:2001, ISO 14001:2005 HACCP etc.

  19. Software Tools for Electrical Quality Assurance in the LHC

    CERN Document Server

    Bednarek, Mateusz

    2011-01-01

    There are over 1600 superconducting magnet circuits in the LHC machine. Many of them consist of a large number of components electrically connected in series. This enhances the sensitivity of the whole circuits to electrical faults of individual components. Furthermore, circuits are equipped with a large number of instrumentation wires, which are exposed to accidental damage or swapping. In order to ensure safe operation, an Electrical Quality Assurance (ELQA) campaign is needed after each thermal cycle. Due to the complexity of the circuits, as well as their distant geographical distribution (tunnel of 27km circumference divided in 8 sectors), suitable software and hardware platforms had to be developed. The software combines an Oracle database, LabView data acquisition applications and PHP-based web follow-up tools. This paper describes the software used for the ELQA of the LHC.

  20. Software tools for electrical quality assurance in the LHC

    International Nuclear Information System (INIS)

    Bednarek, M.; Ludwin, J.

    2012-01-01

    There are over 1600 superconducting magnet circuits in the LHC machine. Many of them consist of a large number of components electrically connected in series. This enhances the sensitivity of the whole circuits to electrical faults of individual components. Furthermore, circuits are equipped with a large number of instrumentation wires, which are exposed to accidental damage or swapping. In order to ensure safe operation, an Electrical Quality Assurance (ELQA) campaign is needed after each thermal cycle. Due to the complexity of the circuits, as well as their distant geographical distribution (tunnel of 27 km circumference divided in 8 sectors), suitable software and hardware platforms had to be developed. The software combines an Oracle database, LabView data acquisition applications and PHP-based web follow-up tools. This paper describes the software used for the ELQA of the LHC. (authors)

  1. Cost Analysis of Poor Quality Using a Software Simulation

    Directory of Open Access Journals (Sweden)

    Jana Fabianová

    2017-02-01

    Full Text Available The issues of quality, cost of poor quality and factors affecting quality are crucial to maintaining a competitiveness regarding to business activities. Use of software applications and computer simulation enables more effective quality management. Simulation tools offer incorporating the variability of more variables in experiments and evaluating their common impact on the final output. The article presents a case study focused on the possibility of using computer simulation Monte Carlo in the field of quality management. Two approaches for determining the cost of poor quality are introduced here. One from retrospective scope of view, where the cost of poor quality and production process are calculated based on historical data. The second approach uses the probabilistic characteristics of the input variables by means of simulation, and reflects as a perspective view of the costs of poor quality. Simulation output in the form of a tornado and sensitivity charts complement the risk analysis.

  2. An empirical evaluation of software quality assurance practices and challenges in a developing country: a comparison of Nigeria and Turkey.

    Science.gov (United States)

    Sowunmi, Olaperi Yeside; Misra, Sanjay; Fernandez-Sanz, Luis; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    The importance of quality assurance in the software development process cannot be overemphasized because its adoption results in high reliability and easy maintenance of the software system and other software products. Software quality assurance includes different activities such as quality control, quality management, quality standards, quality planning, process standardization and improvement amongst others. The aim of this work is to further investigate the software quality assurance practices of practitioners in Nigeria. While our previous work covered areas on quality planning, adherence to standardized processes and the inherent challenges, this work has been extended to include quality control, software process improvement and international quality standard organization membership. It also makes comparison based on a similar study carried out in Turkey. The goal is to generate more robust findings that can properly support decision making by the software community. The qualitative research approach, specifically, the use of questionnaire research instruments was applied to acquire data from software practitioners. In addition to the previous results, it was observed that quality assurance practices are quite neglected and this can be the cause of low patronage. Moreover, software practitioners are neither aware of international standards organizations or the required process improvement techniques; as such their claimed standards are not aligned to those of accredited bodies, and are only limited to their local experience and knowledge, which makes it questionable. The comparison with Turkey also yielded similar findings, making the results typical of developing countries. The research instrument used was tested for internal consistency using the Cronbach's alpha, and it was proved reliable. For the software industry in developing countries to grow strong and be a viable source of external revenue, software assurance practices have to be taken seriously

  3. Identifying strengths and weaknesses of Quality Management Unit University of Sumatera Utara software using SCAMPI C

    Science.gov (United States)

    Gunawan, D.; Amalia, A.; Rahmat, R. F.; Muchtar, M. A.; Siregar, I.

    2018-02-01

    Identification of software maturity level is a technique to determine the quality of the software. By identifying the software maturity level, the weaknesses of the software can be observed. As a result, the recommendations might be a reference for future software maintenance and development. This paper discusses the software Capability Level (CL) with case studies on Quality Management Unit (Unit Manajemen Mutu) University of Sumatera Utara (UMM-USU). This research utilized Standard CMMI Appraisal Method for Process Improvement class C (SCAMPI C) model with continuous representation. This model focuses on activities for developing quality products and services. The observation is done in three process areas, such as Project Planning (PP), Project Monitoring and Control (PMC), and Requirements Management (REQM). According to the measurement of software capability level for UMM-USU software, turns out that the capability level for the observed process area is in the range of CL1 and CL2. Planning Project (PP) is the only process area which reaches capability level 2, meanwhile, PMC and REQM are still in CL 1 or in performed level. This research reveals several weaknesses of existing UMM-USU software. Therefore, this study proposes several recommendations for UMM-USU to improve capability level for observed process areas.

  4. Application of software quality assurance methods in validation and maintenance of reactor analysis computer codes

    International Nuclear Information System (INIS)

    Reznik, L.

    1994-01-01

    Various computer codes employed at Israel Electricity Company for preliminary reactor design analysis and fuel cycle scoping calculations have been often subject to program source modifications. Although most changes were due to computer or operating system compatibility problems, a number of significant modifications were due to model improvement and enhancements of algorithm efficiency and accuracy. With growing acceptance of software quality assurance requirements and methods, a program of implementing extensive testing of modified software has been adopted within the regular maintenance activities. In this work survey has been performed of various software quality assurance methods of software testing which belong mainly to the two major categories of implementation ('white box') and specification-based ('black box') testing. The results of this survey exhibits a clear preference of specification-based testing. In particular the equivalence class partitioning method and the boundary value method have been selected as especially suitable functional methods for testing reactor analysis codes.A separate study of software quality assurance methods and techniques has been performed in this work objective to establish appropriate pre-test software specification methods. Two methods of software analysis and specification have been selected as the most suitable for this purpose: The method of data flow diagrams has been shown to be particularly valuable for performing the functional/procedural software specification while the entities - relationship diagrams has been approved to be efficient for specifying software data/information domain. Feasibility of these two methods has been analyzed in particular for software uncertainty analysis and overall code accuracy estimation. (author). 14 refs

  5. Identification of Patient Safety Risks Associated with Electronic Health Records: A Software Quality Perspective.

    Science.gov (United States)

    Virginio, Luiz A; Ricarte, Ivan Luiz Marques

    2015-01-01

    Although Electronic Health Records (EHR) can offer benefits to the health care process, there is a growing body of evidence that these systems can also incur risks to patient safety when developed or used improperly. This work is a literature review to identify these risks from a software quality perspective. Therefore, the risks were classified based on the ISO/IEC 25010 software quality model. The risks identified were related mainly to the characteristics of "functional suitability" (i.e., software bugs) and "usability" (i.e., interface prone to user error). This work elucidates the fact that EHR quality problems can adversely affect patient safety, resulting in errors such as incorrect patient identification, incorrect calculation of medication dosages, and lack of access to patient data. Therefore, the risks presented here provide the basis for developers and EHR regulating bodies to pay attention to the quality aspects of these systems that can result in patient harm.

  6. Repair of Achilles tendon defect with autologous ASCs engineered tendon in a rabbit model.

    Science.gov (United States)

    Deng, Dan; Wang, Wenbo; Wang, Bin; Zhang, Peihua; Zhou, Guangdong; Zhang, Wen Jie; Cao, Yilin; Liu, Wei

    2014-10-01

    Adipose derived stem cells (ASCs) are an important cell source for tissue regeneration and have been demonstrated the potential of tenogenic differentiation in vitro. This study explored the feasibility of using ASCs for engineered tendon repair in vivo in a rabbit Achilles tendon model. Total 30 rabbits were involved in this study. A composite tendon scaffold composed of an inner part of polyglycolic acid (PGA) unwoven fibers and an outer part of a net knitted with PGA/PLA (polylactic acid) fibers was used to provide mechanical strength. Autologous ASCs were harvested from nuchal subcutaneous adipose tissues and in vitro expanded. The expanded ASCs were harvested and resuspended in culture medium and evenly seeded onto the scaffold in the experimental group, whereas cell-free scaffolds served as the control group. The constructs of both groups were cultured inside a bioreactor under dynamic stretch for 5 weeks. In each of 30 rabbits, a 2 cm defect was created on right side of Achilles tendon followed by the transplantation of a 3 cm cell-seeded scaffold in the experimental group of 15 rabbits, or by the transplantation of a 3 cm cell-free scaffold in the control group of 15 rabbits. Animals were sacrificed at 12, 21 and 45 weeks post-surgery for gross view, histology, and mechanical analysis. The results showed that short term in vitro culture enabled ASCs to produce matrix on the PGA fibers and the constructs showed tensile strength around 50 MPa in both groups (p > 0.05). With the increase of implantation time, cell-seeded constructs gradually form neo-tendon and became more mature at 45 weeks with histological structure similar to that of native tendon and with the presence of bipolar pattern and D-periodic structure of formed collagen fibrils. Additionally, both collagen fibril diameters and tensile strength increased continuously with significant difference among different time points (p tendon tissue with fibril structure observable only at 45 weeks

  7. Improving Software Quality and Management Through Use of Service Level Agreements

    National Research Council Canada - National Science Library

    Gaines, Leonard T

    2005-01-01

    .... SLAs are typically used in outsourcing contracts for post-production support. We propose that SLAs be used in software acquisition to support quality and process control throughout the lifecycle...

  8. 18 CFR Appendix 1 to Part 301 - ASC Utility Filing Template

    Science.gov (United States)

    2010-04-01

    ... 18 Conservation of Power and Water Resources 1 2010-04-01 2010-04-01 false ASC Utility Filing Template 1 Appendix 1 to Part 301 Conservation of Power and Water Resources FEDERAL ENERGY REGULATORY COMMISSION, DEPARTMENT OF ENERGY REGULATIONS FOR FEDERAL POWER MARKETING ADMINISTRATIONS AVERAGE SYSTEM COST...

  9. Geologic software for nuclear waste repository studies: A quality assurance program

    International Nuclear Information System (INIS)

    Figuli, S.; English, S.L.

    1987-04-01

    This paper discusses a Quality Assurance (QA) program that Kent State University (KSU) has implemented for the development of geologic software. The software being developed at KSU will be used in the site characterization of nuclear waste repositories and must meet the requirements of federal regulations. This QA program addresses the development of models that will be used in the evaluation of the long-term climatic stability of three sites in the western US

  10. PLANNING QUALITY ASSURANCE PROCESSES IN A LARGE SCALE GEOGRAPHICALLY SPREAD HYBRID SOFTWARE DEVELOPMENT PROJECT

    Directory of Open Access Journals (Sweden)

    Святослав Аркадійович МУРАВЕЦЬКИЙ

    2016-02-01

    Full Text Available There have been discussed key points of operational activates in a large scale geographically spread software development projects. A look taken at required QA processes structure in such project. There have been given up to date methods of integration quality assurance processes into software development processes. There have been reviewed existing groups of software development methodologies. Such as sequential, agile and based on RPINCE2. There have been given a condensed overview of quality assurance processes in each group. There have been given a review of common challenges that sequential and agile models are having in case of large geographically spread hybrid software development project. Recommendations were given in order to tackle those challenges.  The conclusions about the best methodology choice and appliance to the particular project have been made.

  11. A Case of Engineering Quality for Mobile Healthcare Applications Using Augmented Personal Software Process Improvement

    Directory of Open Access Journals (Sweden)

    Shahbaz Ahmed Khan Ghayyur

    2016-01-01

    Full Text Available Mobile healthcare systems are currently considered as key research areas in the domain of software engineering. The adoption of modern technologies, for mobile healthcare systems, is a quick option for industry professionals. Software architecture is a key feature that contributes towards a software product, solution, or services. Software architecture helps in better communication, documentation of design decisions, risks identification, basis for reusability, scalability, scheduling, and reduced maintenance cost and lastly it helps to avoid software failures. Hence, in order to solve the abovementioned issues in mobile healthcare, the software architecture is integrated with personal software process. Personal software process has been applied successfully but it is unable to address the issues related to architectural design and evaluation capabilities. Hence, a new technique architecture augmented personal process is presented in order to enhance the quality of the mobile healthcare systems through the use of architectural design with integration of personal software process. The proposed process was validated by case studies. It was found that the proposed process helped in reducing the overall costs and effort. Moreover, an improved architectural design helped in development of high quality mobile healthcare system.

  12. A Comparison of ASCE and FAO56 Reference Evapotranspiration at Different Subdaily Timescales: a Numerical Study

    Directory of Open Access Journals (Sweden)

    Farzin Parchami-Araghi

    2017-01-01

    Full Text Available Introduction: Subdaily estimates of reference evapotranspiration (ETo are needed in many applications such as dynamic agro-hydrological modeling. The ASCE and FAO56 Penman–Monteith models (ASCE-PM and FAO56-PM, respectively has received favorable acceptance and application over much of the world, including the United States, for establishing a reference evapotranspiration (ETo index as a function of weather parameters. In the past several years various studies have evaluated ASCE-PM and FAO56-PM models for calculating the commonest hourly or 15-min ETo either by comparing them with lysimetric measurements or by comparison with one another (2, 3, 5, 9, 10, 11, 16, 17, 19. In this study, sub-daily ET o estimates made by the ASCE-PM and FAO56-PM models at different timescales (1-360 min were compared through conduction of a computational experiment, using a daily to sub-daily disaggregation framework developed by Parchami-Araghi et al. (14. Materials and Methods: Daily and sub-daily weather data at different timescales (1-360 min were generated via a daily-to-sub-daily weather data disaggregation framework developed by Parchami-Araghi et al. (14, using long-term (59 years daily weather data obtained from Abadan synoptic weather station. Daily/sub-daily net long wave radiation (Rnl was estimated through 6 different approaches, including using two different criteria for identifying the daytime/nighttime periods : 1 the standard criteria implemented in both ASCE-PM and FAO56-PM models and 2 criterion of actual time of sunset and sunrise in combination with 1 estimation of clear-sky radiation (Rso based on the standard approach implemented in both ASCE-PM and FAO56-PM models (1st and 2nd Rnl estimation approaches, respectively, 2 integral of the Rso estimates derived via a physically based solar radiation model developed by Yang et al. (25, YNG model, for one-second time-steps (3rd and 4th Rnl estimation approaches, respectively, and 3 integral of

  13. Evaluation of features to support safety and quality in general practice clinical software

    Science.gov (United States)

    2011-01-01

    Background Electronic prescribing is now the norm in many countries. We wished to find out if clinical software systems used by general practitioners in Australia include features (functional capabilities and other characteristics) that facilitate improved patient safety and care, with a focus on quality use of medicines. Methods Seven clinical software systems used in general practice were evaluated. Fifty software features that were previously rated as likely to have a high impact on safety and/or quality of care in general practice were tested and are reported here. Results The range of results for the implementation of 50 features across the 7 clinical software systems was as follows: 17-31 features (34-62%) were fully implemented, 9-13 (18-26%) partially implemented, and 9-20 (18-40%) not implemented. Key findings included: Access to evidence based drug and therapeutic information was limited. Decision support for prescribing was available but varied markedly between systems. During prescribing there was potential for medicine mis-selection in some systems, and linking a medicine with its indication was optional. The definition of 'current medicines' versus 'past medicines' was not always clear. There were limited resources for patients, and some medicines lists for patients were suboptimal. Results were provided to the software vendors, who were keen to improve their systems. Conclusions The clinical systems tested lack some of the features expected to support patient safety and quality of care. Standards and certification for clinical software would ensure that safety features are present and that there is a minimum level of clinical functionality that clinicians could expect to find in any system.

  14. Fuzzy system for risk analysis in software projects through the attributes of quality standards iso 25000

    Directory of Open Access Journals (Sweden)

    Chau Sen Shia

    2014-02-01

    Full Text Available With the growth in demand for products and services in the IT area, companies encounter difficulties in establishing a metric or measure of quality of services to address measurably qualitative values in their planning. In this work fuzzy logic, standard SQuaRE (measurement of the quality of software products, Likertscale, GQM method (Goal-Question-Metric -indicator of quality of Software and the project risk analysis model of Boehm were used to assess the quality of services and decision-making, according to your demand and requests for software development. With the aim of improving the quality in the provision of services, the application is used to integrate the team and follow the life cycle of a project from its initial phase, and to assist in the comparison with the proposed schedule during the requirements elicitation.

  15. Specialized software for optimization of the quality control of the mammography units

    International Nuclear Information System (INIS)

    Stoeva, M.; Vassileva, J.

    2004-01-01

    Quality control is essential to ensure the equipment used is reliable and consistent in order to maintain radiation does as low as reasonably achievable whilst optimizing image quality and performance in mammography. The effectiveness of mammographic screening is highly dependent on the consistent production of high quality diagnostic images. Mammography is highly dependent on the equipment status, which requires an effective Quality Control (QC) program to provide tools for continuous assessment of the equipment performance and also data storage and analysis of the protocols' data. The objective of this paper is to present the specialized software for Quality Control of the Mammography Units, as tool providing additional functionality for optimizations of the Mammography QC data storage and management. The PC program was developed according to the requirements stated in the European protocol for Quality Control of the Mammography Screening and the data collected as a result of its application in several Bulgarian hospitals. The Structured Analysis method was used in order to perform a case, which resulted in the development of the specialized software with a database module, providing the following functionality: Data Storage, Preliminary Data Processing and Post-Processing, Manual Data Entry, Data Import from XLS format, Data Export to XLS format, Printing, Data Filters, Automated Calculation, Automated Graphical Representation, Archiving The development of specialized QC software with a database for mammography units facilitates the process of QC data storage and handling and minimizes the errors. The electronic format for data storage is especially useful in case of long-term storage and periodical data analysis/access. The integrated data processing functionality and the automated import/export features based on standard platform increase the compatibility of the data. (authors)

  16. A concept of software testing for SMART MMIS software

    International Nuclear Information System (INIS)

    Seo, Yong Seok; Seong, Seung Hwan; Park, Keun Ok; Hur, Sub; Kim, Dong Hoon

    2001-01-01

    In order to achieve high quality of SMART MMIS software, the well-constructed software testing concept shall be required. This paper established software testing concept which is to be applied to SMART MMIS software, in terms of software testing organization, documentation. procedure, and methods. The software testing methods are classified into source code static analysis and dynamic testing. The software dynamic testing methods are discussed with two aspects: white-box and black-box testing. As software testing concept introduced in this paper is applied to the SMART MMIS software. the high quality of the software will be produced. In the future, software failure data will be collected through the construction of SMART MMIS prototyping facility which the software testing concept of this paper is applied to

  17. SAPHIRE 8 Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Curtis Smith

    2010-02-01

    This Quality Assurance (QA) Plan documents the QA activities that will be managed by the INL related to JCN N6423. The NRC developed the SAPHIRE computer code for performing probabilistic risk assessments (PRAs) using a personal computer (PC) at the Idaho National Laboratory (INL) under Job Code Number (JCN) L1429. SAPHIRE started out as a feasibility study for a PRA code to be run on a desktop personal PC and evolved through several phases into a state-of-the-art PRA code. The developmental activity of SAPHIRE was the result of two concurrent important events: The tremendous expansion of PC software and hardware capability of the 90s and the onset of a risk-informed regulation era.

  18. Software: our quest for excellence. Honoring 50 years of software history, progress, and process

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-06-01

    The Software Quality Forum was established by the Software Quality Assurance (SQA) Subcommittee, which serves as a technical advisory group on software engineering and quality initiatives and issues for DOE`s quality managers. The forum serves as an opportunity for all those involved in implementing SQA programs to meet and share ideas and concerns. Participation from managers, quality engineers, and software professionals provides an ideal environment for identifying and discussing issues and concerns. The interaction provided by the forum contributes to the realization of a shared goal--high quality software product. Topics include: testing, software measurement, software surety, software reliability, SQA practices, assessments, software process improvement, certification and licensing of software professionals, CASE tools, software project management, inspections, and management`s role in ensuring SQA. The bulk of this document consists of vugraphs. Selected papers have been indexed separately for inclusion in the Energy Science and Technology Database.

  19. The Impact of Organization, Project and Governance Variables on Software Quality and Project Success

    OpenAIRE

    Abbas, Noura; Gravell, Andy; Wills, Gary

    2010-01-01

    In this paper we present a statistically tested evidence about how quality and success rate are correlated with variables reflecting the organization and aspects of its project’s governance, namely retrospectives and metrics. The results presented in this paper are based on the Agile Projects Governance Survey that collected 129 responses. This paper discuss the deep analysis of this survey, and the main findings suggest that when applying agile software development, the quality of software i...

  20. Development and case study of a science-based software platform to support policy making on air quality.

    Science.gov (United States)

    Zhu, Yun; Lao, Yanwen; Jang, Carey; Lin, Chen-Jen; Xing, Jia; Wang, Shuxiao; Fu, Joshua S; Deng, Shuang; Xie, Junping; Long, Shicheng

    2015-01-01

    This article describes the development and implementations of a novel software platform that supports real-time, science-based policy making on air quality through a user-friendly interface. The software, RSM-VAT, uses a response surface modeling (RSM) methodology and serves as a visualization and analysis tool (VAT) for three-dimensional air quality data obtained by atmospheric models. The software features a number of powerful and intuitive data visualization functions for illustrating the complex nonlinear relationship between emission reductions and air quality benefits. The case study of contiguous U.S. demonstrates that the enhanced RSM-VAT is capable of reproducing the air quality model results with Normalized Mean Bias quality policy making in near real time. Copyright © 2014. Published by Elsevier B.V.

  1. Medicare and Medicaid programs: Hospital Outpatient Prospective Payment and Ambulatory Surgical Center Payment Systems and Quality Reporting Programs; electronic reporting pilot; Inpatient Rehabilitation Facilities Quality Reporting Program; revision to Quality Improvement Organization regulations. Final rule with comment period.

    Science.gov (United States)

    2012-11-15

    This final rule with comment period revises the Medicare hospital outpatient prospective payment system (OPPS) and the Medicare ambulatory surgical center (ASC) payment system for CY 2013 to implement applicable statutory requirements and changes arising from our continuing experience with these systems. In this final rule with comment period, we describe the changes to the amounts and factors used to determine the payment rates for Medicare services paid under the OPPS and those paid under the ASC payment system. In addition, this final rule with comment period updates and refines the requirements for the Hospital Outpatient Quality Reporting (OQR) Program, the ASC Quality Reporting (ASCQR) Program, and the Inpatient Rehabilitation Facility (IRF) Quality Reporting Program. We are continuing the electronic reporting pilot for the Electronic Health Record (EHR) Incentive Program, and revising the various regulations governing Quality Improvement Organizations (QIOs), including the secure transmittal of electronic medical information, beneficiary complaint resolution and notification processes, and technical changes. The technical changes to the QIO regulations reflect CMS' commitment to the general principles of the President's Executive Order on Regulatory Reform, Executive Order 13563 (January 18, 2011).

  2. Androgen receptor (AR) degradation enhancer ASC-J9® in an FDA-approved formulated solution suppresses castration resistant prostate cancer cell growth.

    Science.gov (United States)

    Cheng, Max A; Chou, Fu-Ju; Wang, Keliang; Yang, Rachel; Ding, Jie; Zhang, Qiaoxia; Li, Gonghui; Yeh, Shuyuan; Xu, Defeng; Chang, Chawnshang

    2018-03-28

    ASC-J9 ® is a recently-developed androgen receptor (AR)-degradation enhancer that effectively suppresses castration resistant prostate cancer (PCa) cell proliferation and invasion. The optimal half maximum inhibitory concentrations (IC 50 ) of ASC-J9 ® at various PCa cell confluences (20%, 50%, and 100%) were assessed via both short-term MTT growth assays and long-term clonogenic proliferation assays. Our results indicate that the IC 50 values for ASC-J9 ® increased with increasing cell confluency. The IC 50 values were significantly decreased in PCa AR-positive cells compared to PCa AR-negative cells or in normal prostate cells. This suggests that ASC-J9 ® may function mainly via targeting the AR-positive PCa cells with limited unwanted side-effects to suppress the surrounding normal prostate cells. Mechanism dissection indicated that ASC-J9 ® might function via altering the apoptosis signals to suppress the PCa AR-negative PC-3 cells. Preclinical studies using multiple in vitro PCa cell lines and an in vivo mouse model with xenografted castration-resistant PCa CWR22Rv1 cells demonstrated that ASC-J9 ® has similar AR degradation effects when dissolved in FDA-approved solvents, including DMSO, PEG-400:Tween-80 (95:5), DMA:Labrasol:Tween-80 (10:45:45), and DMA:Labrasol:Tween-20 (10:45:45). Together, results from preclinical studies suggest a potential new therapy with AR-degradation enhancer ASC-J9 ® may potentially be ready to be used in human clinical trials in order to better suppress PCa at later castration resistant stages. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Attributes Of Quality Scenarios/Scenario Sets Used In Software Requirements Elicitation

    National Research Council Canada - National Science Library

    Braun, Kimberly

    1997-01-01

    .... This thesis examines scenarios used in software requirements elicitation. Many different definitions, formats, and ideas exist on scenarios, but no thorough work has been done on what makes a good, quality scenario and scenario set...

  4. QUALITY SERVICES EVALUATION MODEL BASED ON DEDICATED SOFTWARE TOOL

    Directory of Open Access Journals (Sweden)

    ANDREEA CRISTINA IONICĂ

    2012-10-01

    Full Text Available In this paper we introduced a new model, called Service Quality (SQ, which combines QFD and SERVQUAL methods. This model takes from the SERVQUAL method the five dimensions of requirements and three of characteristics and from the QFD method the application methodology. The originality of the SQ model consists in computing a global index that reflects the customers’ requirements accomplishment level by the quality characteristics. In order to prove the viability of the SQ model, there was developed a software tool that was applied for the evaluation of a health care services provider.

  5. Manual on quality assurance for computer software related to the safety of nuclear power plants

    International Nuclear Information System (INIS)

    1988-01-01

    The objective of the Manual is to provide guidance in the assurance of quality of specification, design, maintenance and use of computer software related to items and activities important to safety (hereinafter referred to as safety related) in nuclear power plants. This guidance is consistent with, and supplements, the requirements and recommendations of Quality Assurance for Safety in Nuclear Power Plants: A Code of Practice, 50-C-QA, and related Safety Guides on quality assurance for nuclear power plants. Annex A identifies the IAEA documents referenced in the Manual. The Manual is intended to be of use to all those who, in any way, are involved with software for safety related applications for nuclear power plants, including auditors who may be called upon to audit management systems and product software. Figs

  6. A study of HPV typing for the management of HPV-positive ASC-US cervical cytologic results.

    Science.gov (United States)

    Schiffman, Mark; Vaughan, Laurence M; Raine-Bennett, Tina R; Castle, Philip E; Katki, Hormuzd A; Gage, Julia C; Fetterman, Barbara; Befano, Brian; Wentzensen, Nicolas

    2015-09-01

    In US cervical screening, immediate colposcopy is recommended for women with HPV-positive ASC-US (equivocal) cytology. We evaluated whether partial typing by Onclarity™ (BD) might identify HPV-positive women with low enough CIN3+ risk to permit 1-year follow-up instead. The NCI-Kaiser Permanente Northern California Persistence and Progression cohort includes a subset of 13,890 women aged 21+ with HC2 (Qiagen)-positive ASC-US at enrollment; current median follow-up is 3.0years. Using stratified random sampling, we typed 2079 archived enrollment specimens including 329 women subsequently diagnosed with CIN3+, 563 with CIN2, and 1187 with ASC-US was 5.2%; this establishes the "benchmark" risk for colposcopic referral. Hierarchically, 3-year cumulative risks for each typing channel were 16.0% for HPV16, 7.4% for HPV18, 7.0% for HPV31, 7.1% for grouped HPV33/58, 4.3% for HPV52, 3.9% for HPV45, 2.7% for HPV51, 1.6% for HPV39/68/35, and 1.3% for HPV59/56/66. ASC-US linked to HPV16, HPV18, HPV31, or HPV33/58 warrants immediate colposcopy. Optimal management of women with HPV52 or HPV45 is uncertain. Risk of women with only HPV51, HPV39/68/35, or HPV59/56/66 might be low enough to recommend 1-year retesting permitting viral clearance. This strategy would defer colposcopy for 40% of women with HPV-positive ASC-US, half of whom would be cotest-negative at 1-year return. Approximately 10% of those with CIN3 diagnosable at enrollment would be delayed 1year instead. Cost-effectiveness analyses are needed. Published by Elsevier Inc.

  7. Improving the quality of care of patients with rheumatic disease using patient-centric electronic redesign software.

    Science.gov (United States)

    Newman, Eric D; Lerch, Virginia; Billet, Jon; Berger, Andrea; Kirchner, H Lester

    2015-04-01

    Electronic health records (EHRs) are not optimized for chronic disease management. To improve the quality of care for patients with rheumatic disease, we developed electronic data capture, aggregation, display, and documentation software. The software integrated and reassembled information from the patient (via a touchscreen questionnaire), nurse, physician, and EHR into a series of actionable views. Core functions included trends over time, rheumatology-related demographics, and documentation for patient and provider. Quality measures collected included patient-reported outcomes, disease activity, and function. The software was tested and implemented in 3 rheumatology departments, and integrated into routine care delivery. Post-implementation evaluation measured adoption, efficiency, productivity, and patient perception. Over 2 years, 6,725 patients completed 19,786 touchscreen questionnaires. The software was adopted for use by 86% of patients and rheumatologists. Chart review and documentation time trended downward, and productivity increased by 26%. Patient satisfaction, activation, and adherence remained unchanged, although pre-implementation values were high. A strong correlation was seen between use of the software and disease control (weighted Pearson's correlation coefficient 0.5927, P = 0.0095), and a relative increase in patients with low disease activity of 3% per quarter was noted. We describe innovative software that aggregates, stores, and displays information vital to improving the quality of care for patients with chronic rheumatic disease. The software was well-adopted by patients and providers. Post-implementation, significant improvements in quality of care, efficiency of care, and productivity were demonstrated. Copyright © 2015 by the American College of Rheumatology.

  8. IEEE Std 730 Software Quality Assurance: Supporting CMMI-DEV v1.3, Product and Process Quality Assurance

    Science.gov (United States)

    2011-05-27

    frameworks 4 CMMI-DEV IEEE / ISO / IEC 15288 / 12207 Quality Assurance ©2011 Walz IEEE Life Cycle Processes & Artifacts • Systems Life Cycle Processes...TAG to ISO TC 176 Quality Management • Quality: ASQ, work experience • Software: three books, consulting, work experience • Systems: Telecom & DoD...and IEEE 730 SQA need to align. The P730 IEEE standards working group has expanded the scope of the SQA process standard to align with IS 12207

  9. Low-frequency, low-magnitude vibrations (LFLM enhances chondrogenic differentiation potential of human adipose derived mesenchymal stromal stem cells (hASCs

    Directory of Open Access Journals (Sweden)

    Krzysztof Marycz

    2016-02-01

    Full Text Available The aim of this study was to evaluate if low-frequency, low-magnitude vibrations (LFLM could enhance chondrogenic differentiation potential of human adipose derived mesenchymal stem cells (hASCs with simultaneous inhibition of their adipogenic properties for biomedical purposes. We developed a prototype device that induces low-magnitude (0.3 g low-frequency vibrations with the following frequencies: 25, 35 and 45 Hz. Afterwards, we used human adipose derived mesenchymal stem cell (hASCS, to investigate their cellular response to the mechanical signals. We have also evaluated hASCs morphological and proliferative activity changes in response to each frequency. Induction of chondrogenesis in hASCs, under the influence of a 35 Hz signal leads to most effective and stable cartilaginous tissue formation through highest secretion of Bone Morphogenetic Protein 2 (BMP-2, and Collagen type II, with low concentration of Collagen type I. These results correlated well with appropriate gene expression level. Simultaneously, we observed significant up-regulation of α3, α4, β1 and β3 integrins in chondroblast progenitor cells treated with 35 Hz vibrations, as well as Sox-9. Interestingly, we noticed that application of 35 Hz frequencies significantly inhibited adipogenesis of hASCs. The obtained results suggest that application of LFLM vibrations together with stem cell therapy might be a promising tool in cartilage regeneration.

  10. QCScreen: a software tool for data quality control in LC-HRMS based metabolomics.

    Science.gov (United States)

    Simader, Alexandra Maria; Kluger, Bernhard; Neumann, Nora Katharina Nicole; Bueschl, Christoph; Lemmens, Marc; Lirk, Gerald; Krska, Rudolf; Schuhmacher, Rainer

    2015-10-24

    Metabolomics experiments often comprise large numbers of biological samples resulting in huge amounts of data. This data needs to be inspected for plausibility before data evaluation to detect putative sources of error e.g. retention time or mass accuracy shifts. Especially in liquid chromatography-high resolution mass spectrometry (LC-HRMS) based metabolomics research, proper quality control checks (e.g. for precision, signal drifts or offsets) are crucial prerequisites to achieve reliable and comparable results within and across experimental measurement sequences. Software tools can support this process. The software tool QCScreen was developed to offer a quick and easy data quality check of LC-HRMS derived data. It allows a flexible investigation and comparison of basic quality-related parameters within user-defined target features and the possibility to automatically evaluate multiple sample types within or across different measurement sequences in a short time. It offers a user-friendly interface that allows an easy selection of processing steps and parameter settings. The generated results include a coloured overview plot of data quality across all analysed samples and targets and, in addition, detailed illustrations of the stability and precision of the chromatographic separation, the mass accuracy and the detector sensitivity. The use of QCScreen is demonstrated with experimental data from metabolomics experiments using selected standard compounds in pure solvent. The application of the software identified problematic features, samples and analytical parameters and suggested which data files or compounds required closer manual inspection. QCScreen is an open source software tool which provides a useful basis for assessing the suitability of LC-HRMS data prior to time consuming, detailed data processing and subsequent statistical analysis. It accepts the generic mzXML format and thus can be used with many different LC-HRMS platforms to process both multiple

  11. Interferência do ácido ascórbico na dosagem glicêmica - doi: 10.5102/ucs.v6i2.722

    OpenAIRE

    Aline Cardoso Barbosa; Tania Cristina Andrade

    2009-01-01

    Sabe-se que exames laboratoriais estão sujeitos a fatores interferentes. A literatura científica cita interferência do ácido ascórbico em dosagens que envolvam oxirredução, como a glicêmica. Devido a isto, buscou-se avaliar o efeito do ácido ascórbico sobre a dosagem glicêmica em dois tipos de soro controle submetidos a diferentes concentrações de ácido ascórbico, variando de 2,5 a 100 mg/dL. Utilizou-se um soro controle com valores normais de glicemia e outro com valores patológicos. Observo...

  12. Prevalence of cervical intraepithelial neoplasia grades II/III and cervical cancer in patients with cytological diagnosis of atypical squamous cells when high-grade intraepithelial lesions (ASC-H) cannot be ruled out.

    Science.gov (United States)

    Cytryn, Andréa; Russomano, Fábio Bastos; Camargo, Maria José de; Zardo, Lucília Maria Gama; Horta, Nilza Maria Sobral Rebelo; Fonseca, Rachel de Carvalho Silveira de Paula; Tristão, Maria Aparecida; Monteiro, Aparecida Cristina Sampaio

    2009-09-01

    The latest update of the Bethesda System divided the category of atypical squamous cells of undetermined significance (ASCUS) into ASC-US (undetermined significance) and ASC-H (high-grade intraepithelial lesion cannot be ruled out). The aims here were to measure the prevalence of pre-invasive lesions (cervical intraepithelial neoplasia, CIN II/III) and cervical cancer among patients referred to Instituto Fernandes Figueira (IFF) with ASC-H cytology, and compare them with ASC-US cases. Cross-sectional study with retrospective data collection, at the IFF Cervical Pathology outpatient clinic. ASCUS cases referred to IFF from November 1997 to September 2007 were reviewed according to the 2001 Bethesda System to reach cytological consensus. The resulting ASC-H and ASC-US cases, along with new cases, were analyzed relative to the outcome of interest. The histological diagnosis (or cytocolposcopic follow-up in cases without such diagnosis) was taken as the gold standard. The prevalence of CIN II/III in cases with ASC-H cytology was 19.29% (95% confidence interval, CI, 9.05-29.55%) and the risk of these lesions was greater among patients with ASC-H than with ASC-US cytology (prevalence ratio, PR, 10.42; 95% CI, 2.39-45.47; P = 0.0000764). Pre-invasive lesions were more frequently found in patients under 50 years of age with ASC-H cytology (PR, 2.67; 95% CI, 0.38-18.83); P = 0.2786998). There were no uterine cervical cancer cases. The prevalence of CIN II/III in patients with ASC-H cytology was significantly higher than with ASC-US, and division into ASC diagnostic subcategories had good capacity for discriminating the presence of pre-invasive lesions.

  13. Unisys' experience in software quality and productivity management of an existing system

    Science.gov (United States)

    Munson, John B.

    1988-01-01

    A summary of Quality Improvement techniques, implementation, and results in the maintenance, management, and modification of large software systems for the Space Shuttle Program's ground-based systems is provided.

  14. A tool to include gamma analysis software into a quality assurance program.

    Science.gov (United States)

    Agnew, Christina E; McGarry, Conor K

    2016-03-01

    To provide a tool to enable gamma analysis software algorithms to be included in a quality assurance (QA) program. Four image sets were created comprising two geometric images to independently test the distance to agreement (DTA) and dose difference (DD) elements of the gamma algorithm, a clinical step and shoot IMRT field and a clinical VMAT arc. The images were analysed using global and local gamma analysis with 2 in-house and 8 commercially available software encompassing 15 software versions. The effect of image resolution on gamma pass rates was also investigated. All but one software accurately calculated the gamma passing rate for the geometric images. Variation in global gamma passing rates of 1% at 3%/3mm and over 2% at 1%/1mm was measured between software and software versions with analysis of appropriately sampled images. This study provides a suite of test images and the gamma pass rates achieved for a selection of commercially available software. This image suite will enable validation of gamma analysis software within a QA program and provide a frame of reference by which to compare results reported in the literature from various manufacturers and software versions. Copyright © 2015. Published by Elsevier Ireland Ltd.

  15. The Astrobiology in Secondary Classrooms (ASC) curriculum: focusing upon diverse students and teachers.

    Science.gov (United States)

    Arino de la Rubia, Leigh S

    2012-09-01

    The Minority Institution Astrobiology Collaborative (MIAC) began working with the NASA Goddard Center for Astrobiology in 2003 to develop curriculum materials for high school chemistry and Earth science classes based on astrobiology concepts. The Astrobiology in Secondary Classrooms (ASC) modules emphasize interdisciplinary connections in astronomy, biology, chemistry, geoscience, physics, mathematics, and ethics through hands-on activities that address national educational standards. Field-testing of the Astrobiology in Secondary Classrooms materials occurred over three years in eight U.S. locations, each with populations that are underrepresented in the career fields of science, technology, engineering, and mathematics. Analysis of the educational research upon the high school students participating in the ASC project showed statistically significant increases in students' perceived knowledge and science reasoning. The curriculum is in its final stages, preparing for review to become a NASA educational product.

  16. Specific developed phantoms and software to assess radiological equipment image quality

    International Nuclear Information System (INIS)

    Verdu, G.; Rodenas, F.

    2011-01-01

    The use of radiographic phantoms specifically designed to evaluate the operation of the radiographic equipment lets the study of the image quality obtained by this equipment in an objective way. In digital radiographic equipment, the analysis of the image quality can be automatized because the acquisition of the image is possible in different technologies that are, computerized radiography or phosphor plate and direct radiography or detector. In this work we have shown an application to assess automatically the constancy quality image in the image chain of the radiographic equipment. This application is integrated by designed radiographic phantoms which are adapted to conventional, dental equipment and specific developed software for the automatic evaluation of the phantom image quality. The software is based on digital image processing techniques that let the automatic detection of the different phantom tests by edge detector, morphological operators, threshold histogram techniques, etc. The utility developed is enough sensitive to the radiographic equipment of operating conditions of voltage (kV) and charge (mAs). It is a friendly user programme connected with a data base of the hospital or clinic where it has been used. After the phantom image processing the user can obtain an inform with a resume of the imaging system state with accepting and constancy results. (author)

  17. Specific developed phantoms and software to assess radiological equipment image quality

    Energy Technology Data Exchange (ETDEWEB)

    Verdu, G., E-mail: gverdu@iqn.upv.es [Universidad Politecnica de Valencia (Spain). Dept. de Ingenieria Quimica y Nuclear; Mayo, P., E-mail: p.mayo@titaniast.com [TITANIA Servicios Teconologicos, Valencia (Spain); Rodenas, F., E-mail: frodenas@mat.upv.es [Universidad Politecnica de Valencia (Spain). Dept. de Matematica Aplicada; Campayo, J.M., E-mail: j.campayo@lainsa.com [Logistica y Acondicionamientos Industriales S.A.U (LAINSA), Valencia (Spain)

    2011-07-01

    The use of radiographic phantoms specifically designed to evaluate the operation of the radiographic equipment lets the study of the image quality obtained by this equipment in an objective way. In digital radiographic equipment, the analysis of the image quality can be automatized because the acquisition of the image is possible in different technologies that are, computerized radiography or phosphor plate and direct radiography or detector. In this work we have shown an application to assess automatically the constancy quality image in the image chain of the radiographic equipment. This application is integrated by designed radiographic phantoms which are adapted to conventional, dental equipment and specific developed software for the automatic evaluation of the phantom image quality. The software is based on digital image processing techniques that let the automatic detection of the different phantom tests by edge detector, morphological operators, threshold histogram techniques, etc. The utility developed is enough sensitive to the radiographic equipment of operating conditions of voltage (kV) and charge (mAs). It is a friendly user programme connected with a data base of the hospital or clinic where it has been used. After the phantom image processing the user can obtain an inform with a resume of the imaging system state with accepting and constancy results. (author)

  18. The Role and Quality of Software Safety in the NASA Constellation Program

    Science.gov (United States)

    Layman, Lucas; Basili, Victor R.; Zelkowitz, Marvin V.

    2010-01-01

    In this study, we examine software safety risk in the early design phase of the NASA Constellation spaceflight program. Obtaining an accurate, program-wide picture of software safety risk is difficult across multiple, independently-developing systems. We leverage one source of safety information, hazard analysis, to provide NASA quality assurance managers with information regarding the ongoing state of software safety across the program. The goal of this research is two-fold: 1) to quantify the relative importance of software with respect to system safety; and 2) to quantify the level of risk presented by software in the hazard analysis. We examined 154 hazard reports created during the preliminary design phase of three major flight hardware systems within the Constellation program. To quantify the importance of software, we collected metrics based on the number of software-related causes and controls of hazardous conditions. To quantify the level of risk presented by software, we created a metric scheme to measure the specificity of these software causes. We found that from 49-70% of hazardous conditions in the three systems could be caused by software or software was involved in the prevention of the hazardous condition. We also found that 12-17% of the 2013 hazard causes involved software, and that 23-29% of all causes had a software control. Furthermore, 10-12% of all controls were software-based. There is potential for inaccuracy in these counts, however, as software causes are not consistently scoped, and the presence of software in a cause or control is not always clear. The application of our software specificity metrics also identified risks in the hazard reporting process. In particular, we found a number of traceability risks in the hazard reports may impede verification of software and system safety.

  19. Prevalence of cervical intraepithelial neoplasia grades II/III and cervical cancer in patients with cytological diagnosis of atypical squamous cells when high-grade intraepithelial lesions (ASC-H cannot be ruled out

    Directory of Open Access Journals (Sweden)

    Andréa Cytryn

    Full Text Available CONTEXT AND OBJECTIVE: The latest update of the Bethesda System divided the category of atypical squamous cells of undetermined significance (ASCUS into ASC-US (undetermined significance and ASC-H (high-grade intraepithelial lesion cannot be ruled out. The aims here were to measure the prevalence of pre-invasive lesions (cervical intraepithelial neoplasia, CIN II/III and cervical cancer among patients referred to Instituto Fernandes Figueira (IFF with ASC-H cytology, and compare them with ASC-US cases. DESIGN AND SETTING: Cross-sectional study with retrospective data collection, at the IFF Cervical Pathology outpatient clinic. METHODS: ASCUS cases referred to IFF from November 1997 to September 2007 were reviewed according to the 2001 Bethesda System to reach cytological consensus. The resulting ASC-H and ASC-US cases, along with new cases, were analyzed relative to the outcome of interest. The histological diagnosis (or cytocolposcopic follow-up in cases without such diagnosis was taken as the gold standard. RESULTS: The prevalence of CIN II/III in cases with ASC-H cytology was 19.29% (95% confidence interval, CI, 9.05-29.55% and the risk of these lesions was greater among patients with ASC-H than with ASC-US cytology (prevalence ratio, PR, 10.42; 95% CI, 2.39-45.47; P = 0.0000764. Pre-invasive lesions were more frequently found in patients under 50 years of age with ASC-H cytology (PR, 2.67; 95% CI, 0.38-18.83; P = 0.2786998. There were no uterine cervical cancer cases. CONCLUSION: The prevalence of CIN II/III in patients with ASC-H cytology was significantly higher than with ASC-US, and division into ASC diagnostic subcategories had good capacity for discriminating the presence of pre-invasive lesions.

  20. Application of newly developed Fluoro-QC software for image quality evaluation in cardiac X-ray systems.

    Science.gov (United States)

    Oliveira, M; Lopez, G; Geambastiani, P; Ubeda, C

    2018-05-01

    A quality assurance (QA) program is a valuable tool for the continuous production of optimal quality images. The aim of this paper is to assess a newly developed automatic computer software for image quality (IR) evaluation in fluoroscopy X-ray systems. Test object images were acquired using one fluoroscopy system, Siemens Axiom Artis model (Siemens AG, Medical Solutions Erlangen, Germany). The software was developed as an ImageJ plugin. Two image quality parameters were assessed: high-contrast spatial resolution (HCSR) and signal-to-noise ratio (SNR). The time between manual and automatic image quality assessment procedures were compared. The paired t-test was used to assess the data. p Values of less than 0.05 were considered significant. The Fluoro-QC software generated faster IQ evaluation results (mean = 0.31 ± 0.08 min) than manual procedure (mean = 4.68 ± 0.09 min). The mean difference between techniques was 4.36 min. Discrepancies were identified in the region of interest (ROI) areas drawn manually with evidence of user dependence. The new software presented the results of two tests (HCSR = 3.06, SNR = 5.17) and also collected information from the DICOM header. Significant differences were not identified between manual and automatic measures of SNR (p value = 0.22) and HCRS (p value = 0.46). The Fluoro-QC software is a feasible, fast and free to use method for evaluating imaging quality parameters on fluoroscopy systems. Copyright © 2017 The College of Radiographers. Published by Elsevier Ltd. All rights reserved.

  1. NEMA NU-1 2007 based and independent quality control software for gamma cameras and SPECT

    International Nuclear Information System (INIS)

    Vickery, A; Joergensen, T; De Nijs, R

    2011-01-01

    A thorough quality assurance of gamma and SPECT cameras requires a careful handling of the measured quality control (QC) data. Most gamma camera manufacturers provide the users with camera specific QC Software. This QC software is indeed a useful tool for the following of day-to-day performance of a single camera. However, when it comes to objective performance comparison of different gamma cameras and a deeper understanding of the calculated numbers, the use of camera specific QC software without access to the source code is rather avoided. Calculations and definitions might differ, and manufacturer independent standardized results are preferred. Based upon the NEMA Standards Publication NU 1-2007, we have developed a suite of easy-to-use data handling software for processing acquired QC data providing the user with instructive images and text files with the results.

  2. Digitized hand-wrist radiographs: comparison of subjective and software-derived image quality at various compression ratios.

    Science.gov (United States)

    McCord, Layne K; Scarfe, William C; Naylor, Rachel H; Scheetz, James P; Silveira, Anibal; Gillespie, Kevin R

    2007-05-01

    The objectives of this study were to compare the effect of JPEG 2000 compression of hand-wrist radiographs on observer image quality qualitative assessment and to compare with a software-derived quantitative image quality index. Fifteen hand-wrist radiographs were digitized and saved as TIFF and JPEG 2000 images at 4 levels of compression (20:1, 40:1, 60:1, and 80:1). The images, including rereads, were viewed by 13 orthodontic residents who determined the image quality rating on a scale of 1 to 5. A quantitative analysis was also performed by using a readily available software based on the human visual system (Image Quality Measure Computer Program, version 6.2, Mitre, Bedford, Mass). ANOVA was used to determine the optimal compression level (P quality. When we used quantitative indexes, the JPEG 2000 images had lower quality at all compression ratios compared with the original TIFF images. There was excellent correlation (R2 >0.92) between qualitative and quantitative indexes. Image Quality Measure indexes are more sensitive than subjective image quality assessments in quantifying image degradation with compression. There is potential for this software-based quantitative method in determining the optimal compression ratio for any image without the use of subjective raters.

  3. A survey of quality assurance practices in biomedical open source software projects.

    Science.gov (United States)

    Koru, Günes; El Emam, Khaled; Neisa, Angelica; Umarji, Medha

    2007-05-07

    Open source (OS) software is continuously gaining recognition and use in the biomedical domain, for example, in health informatics and bioinformatics. Given the mission critical nature of applications in this domain and their potential impact on patient safety, it is important to understand to what degree and how effectively biomedical OS developers perform standard quality assurance (QA) activities such as peer reviews and testing. This would allow the users of biomedical OS software to better understand the quality risks, if any, and the developers to identify process improvement opportunities to produce higher quality software. A survey of developers working on biomedical OS projects was conducted to examine the QA activities that are performed. We took a descriptive approach to summarize the implementation of QA activities and then examined some of the factors that may be related to the implementation of such practices. Our descriptive results show that 63% (95% CI, 54-72) of projects did not include peer reviews in their development process, while 82% (95% CI, 75-89) did include testing. Approximately 74% (95% CI, 67-81) of developers did not have a background in computing, 80% (95% CI, 74-87) were paid for their contributions to the project, and 52% (95% CI, 43-60) had PhDs. A multivariate logistic regression model to predict the implementation of peer reviews was not significant (likelihood ratio test = 16.86, 9 df, P = .051) and neither was a model to predict the implementation of testing (likelihood ratio test = 3.34, 9 df, P = .95). Less attention is paid to peer review than testing. However, the former is a complementary, and necessary, QA practice rather than an alternative. Therefore, one can argue that there are quality risks, at least at this point in time, in transitioning biomedical OS software into any critical settings that may have operational, financial, or safety implications. Developers of biomedical OS applications should invest more effort

  4. Report of AAPM Task Group 162: Software for planar image quality metrology.

    Science.gov (United States)

    Samei, Ehsan; Ikejimba, Lynda C; Harrawood, Brian P; Rong, John; Cunningham, Ian A; Flynn, Michael J

    2018-02-01

    The AAPM Task Group 162 aimed to provide a standardized approach for the assessment of image quality in planar imaging systems. This report offers a description of the approach as well as the details of the resultant software bundle to measure detective quantum efficiency (DQE) as well as its basis components and derivatives. The methodology and the associated software include the characterization of the noise power spectrum (NPS) from planar images acquired under specific acquisition conditions, modulation transfer function (MTF) using an edge test object, the DQE, and effective DQE (eDQE). First, a methodological framework is provided to highlight the theoretical basis of the work. Then, a step-by-step guide is included to assist in proper execution of each component of the code. Lastly, an evaluation of the method is included to validate its accuracy against model-based and experimental data. The code was built using a Macintosh OSX operating system. The software package contains all the source codes to permit an experienced user to build the suite on a Linux or other *nix type system. The package further includes manuals and sample images and scripts to demonstrate use of the software for new users. The results of the code are in close alignment with theoretical expectations and published results of experimental data. The methodology and the software package offered in AAPM TG162 can be used as baseline for characterization of inherent image quality attributes of planar imaging systems. © 2017 American Association of Physicists in Medicine.

  5. Interface-based software testing

    OpenAIRE

    Aziz Ahmad Rais

    2016-01-01

    Software quality is determined by assessing the characteristics that specify how it should work, which are verified through testing. If it were possible to touch, see, or measure software, it would be easier to analyze and prove its quality. Unfortunately, software is an intangible asset, which makes testing complex. This is especially true when software quality is not a question of particular functions that can be tested through a graphical user interface. The primary objective of softwar...

  6. Improving Code Quality of the Compact Muon Solenoid Electromagnetic Calorimeter Control Software to Increase System Maintainability

    CERN Multimedia

    Holme, Oliver; Dissertori, Günther; Djambazov, Lubomir; Lustermann, Werner; Zelepoukine, Serguei

    2013-01-01

    The Detector Control System (DCS) software of the Electromagnetic Calorimeter (ECAL) of the Compact Muon Solenoid (CMS) experiment at CERN is designed primarily to enable safe and efficient operation of the detector during Large Hadron Collider (LHC) data-taking periods. Through a manual analysis of the code and the adoption of ConQAT [1], a software quality assessment toolkit, the CMS ECAL DCS team has made significant progress in reducing complexity and improving code quality, with observable results in terms of a reduction in the effort dedicated to software maintenance. This paper explains the methodology followed, including the motivation to adopt ConQAT, the specific details of how this toolkit was used and the outcomes that have been achieved. [1] ConQAT, Continuous Quality Assessment Toolkit; https://www.conqat.org/

  7. Daily quality assurance software for a satellite radiometer system

    Science.gov (United States)

    Keegstra, P. B.; Smoot, G. F.; Bennett, C. L.; Aymon, J.; Backus, C.; Deamici, G.; Hinshaw, G.; Jackson, P. D.; Kogut, A.; Lineweaver, C.

    1992-01-01

    Six Differential Microwave Radiometers (DMR) on COBE (Cosmic Background Explorer) measure the large-angular-scale isotropy of the cosmic microwave background (CMB) at 31.5, 53, and 90 GHz. Quality assurance software analyzes the daily telemetry from the spacecraft to ensure that the instrument is operating correctly and that the data are not corrupted. Quality assurance for DMR poses challenging requirements. The data are differential, so a single bad point can affect a large region of the sky, yet the CMB isotropy requires lengthy integration times (greater than 1 year) to limit potential CMB anisotropies. Celestial sources (with the exception of the moon) are not, in general, visible in the raw differential data. A 'quicklook' software system was developed that, in addition to basic plotting and limit-checking, implements a collection of data tests as well as long-term trending. Some of the key capabilities include the following: (1) stability analysis showing how well the data RMS averages down with increased data; (2) a Fourier analysis and autocorrelation routine to plot the power spectrum and confirm the presence of the 3 mK 'cosmic' dipole signal; (3) binning of the data against basic spacecraft quantities such as orbit angle; (4) long-term trending; and (5) dipole fits to confirm the spacecraft attitude azimuth angle.

  8. SOFTWARE QUALITY ASSURANCE FOR EMERGENCY RESPONSE CONSEQUENCE ASSESSMENT MODELS AT DOE'S SAVANNAH RIVER SITE

    International Nuclear Information System (INIS)

    Hunter, C

    2007-01-01

    The Savannah River National Laboratory's (SRNL) Atmospheric Technologies Group develops, maintains, and operates computer-based software applications for use in emergency response consequence assessment at DOE's Savannah River Site. These applications range from straightforward, stand-alone Gaussian dispersion models run with simple meteorological input to complex computational software systems with supporting scripts that simulate highly dynamic atmospheric processes. A software quality assurance program has been developed to ensure appropriate lifecycle management of these software applications. This program was designed to meet fully the overall structure and intent of SRNL's institutional software QA programs, yet remain sufficiently practical to achieve the necessary level of control in a cost-effective manner. A general overview of this program is described

  9. Maintaining Quality and Confidence in Open-Source, Evolving Software: Lessons Learned with PFLOTRAN

    Science.gov (United States)

    Frederick, J. M.; Hammond, G. E.

    2017-12-01

    Software evolution in an open-source framework poses a major challenge to a geoscientific simulator, but when properly managed, the pay-off can be enormous for both the developers and the community at large. Developers must juggle implementing new scientific process models, adopting increasingly efficient numerical methods and programming paradigms, changing funding sources (or total lack of funding), while also ensuring that legacy code remains functional and reported bugs are fixed in a timely manner. With robust software engineering and a plan for long-term maintenance, a simulator can evolve over time incorporating and leveraging many advances in the computational and domain sciences. In this positive light, what practices in software engineering and code maintenance can be employed within open-source development to maximize the positive aspects of software evolution and community contributions while minimizing its negative side effects? This presentation will discusses steps taken in the development of PFLOTRAN (www.pflotran.org), an open source, massively parallel subsurface simulator for multiphase, multicomponent, and multiscale reactive flow and transport processes in porous media. As PFLOTRAN's user base and development team continues to grow, it has become increasingly important to implement strategies which ensure sustainable software development while maintaining software quality and community confidence. In this presentation, we will share our experiences and "lessons learned" within the context of our open-source development framework and community engagement efforts. Topics discussed will include how we've leveraged both standard software engineering principles, such as coding standards, version control, and automated testing, as well unique advantages of object-oriented design in process model coupling, to ensure software quality and confidence. We will also be prepared to discuss the major challenges faced by most open-source software teams, such

  10. Model reliability and software quality assurance in simulation of nuclear fuel waste management systems

    International Nuclear Information System (INIS)

    Oeren, T.I.; Elzas, M.S.; Sheng, G.; Wageningen Agricultural Univ., Netherlands; McMaster Univ., Hamilton, Ontario)

    1985-01-01

    As is the case with all scientific simulation studies, computerized simulation of nuclear fuel waste management systems can introduce and hide various types of errors. Frameworks to clarify issues of model reliability and software quality assurance are offered. Potential problems with reference to the main areas of concern for reliability and quality are discussed; e.g., experimental issues, decomposition, scope, fidelity, verification, requirements, testing, correctness, robustness are treated with reference to the experience gained in the past. A list comprising over 80 most common computerization errors is provided. Software tools and techniques used to detect and to correct computerization errors are discussed

  11. Engineering high quality medical software

    CERN Document Server

    Coronato, Antonio

    2018-01-01

    This book focuses on high-confidence medical software in the growing field of e-health, telecare services and health technology. It covers the development of methodologies and engineering tasks together with standards and regulations for medical software.

  12. Software for improving the quality of project management, a case study: international manufacture of electrical equipment

    Science.gov (United States)

    Preradović, D. M.; Mićić, Lj S.; Barz, C.

    2017-05-01

    Production conditions in today’s world require software support at every stage of production and development of new products, for quality assurance and compliance with ISO standards. In addition to ISO standards such as usual metrics of quality, companies today are focused on other optional standards, such as CMMI (Capability Maturity Model Integrated) or prescribing they own standards. However, while there is intensive progress being made in the PM (project management), there is still a significant number of projects, at the global level, that are failures. These have failed to achieve their goals, within budget or timeframe. This paper focuses on checking the role of software tools through the rate of success in projects implemented in the case of internationally manufactured electrical equipment. The results of this research show the level of contribution of the project management software used to manage and develop new products to improve PM processes and PM functions, and how selection of the software tools affects the quality of PM processes and successfully completed projects.

  13. Differential regulation of caspase-1 activation, pyroptosis, and autophagy via Ipaf and ASC in Shigella-infected macrophages.

    Directory of Open Access Journals (Sweden)

    Toshihiko Suzuki

    2007-08-01

    Full Text Available Shigella infection, the cause of bacillary dysentery, induces caspase-1 activation and cell death in macrophages, but the precise mechanisms of this activation remain poorly understood. We demonstrate here that caspase-1 activation and IL-1beta processing induced by Shigella are mediated through Ipaf, a cytosolic pattern-recognition receptor of the nucleotide-binding oligomerization domain (NOD-like receptor (NLR family, and the adaptor protein apoptosis-associated speck-like protein containing a C-terminal caspase recruitment domain (ASC. We also show that Ipaf was critical for pyroptosis, a specialized form of caspase-1-dependent cell death induced in macrophages by bacterial infection, whereas ASC was dispensable. Unlike that observed in Salmonella and Legionella, caspase-1 activation induced by Shigella infection was independent of flagellin. Notably, infection of macrophages with Shigella induced autophagy, which was dramatically increased by the absence of caspase-1 or Ipaf, but not ASC. Autophagy induced by Shigella required an intact bacterial type III secretion system but not VirG protein, a bacterial factor required for autophagy in epithelial-infected cells. Treatment of macrophages with 3-methyladenine, an inhibitor of autophagy, enhanced pyroptosis induced by Shigella infection, suggesting that autophagy protects infected macrophages from pyroptosis. Thus, Ipaf plays a critical role in caspase-1 activation induced by Shigella independently of flagellin. Furthermore, the absence of Ipaf or caspase-1, but not ASC, regulates pyroptosis and the induction of autophagy in Shigella-infected macrophages, providing a novel function for NLR proteins in bacterial-host interactions.

  14. Taking advantage of ground data systems attributes to achieve quality results in testing software

    Science.gov (United States)

    Sigman, Clayton B.; Koslosky, John T.; Hageman, Barbara H.

    1994-01-01

    During the software development life cycle process, basic testing starts with the development team. At the end of the development process, an acceptance test is performed for the user to ensure that the deliverable is acceptable. Ideally, the delivery is an operational product with zero defects. However, the goal of zero defects is normally not achieved but is successful to various degrees. With the emphasis on building low cost ground support systems while maintaining a quality product, a key element in the test process is simulator capability. This paper reviews the Transportable Payload Operations Control Center (TPOCC) Advanced Spacecraft Simulator (TASS) test tool that is used in the acceptance test process for unmanned satellite operations control centers. The TASS is designed to support the development, test and operational environments of the Goddard Space Flight Center (GSFC) operations control centers. The TASS uses the same basic architecture as the operations control center. This architecture is characterized by its use of distributed processing, industry standards, commercial off-the-shelf (COTS) hardware and software components, and reusable software. The TASS uses much of the same TPOCC architecture and reusable software that the operations control center developer uses. The TASS also makes use of reusable simulator software in the mission specific versions of the TASS. Very little new software needs to be developed, mainly mission specific telemetry communication and command processing software. By taking advantage of the ground data system attributes, successful software reuse for operational systems provides the opportunity to extend the reuse concept into the test area. Consistency in test approach is a major step in achieving quality results.

  15. Inter-comparison and Quality Assurance of acquisition and processing software for MUGA studies in Cuba

    International Nuclear Information System (INIS)

    Lopez, A.; Ponce, F.; Peix, A.; Gonzalez, J.; Perez, M.; Diaz, M.

    2002-01-01

    With the purpose of create the bases for quality control and quality assurance of the acquisition and processing program of gated cardiac blood-pool (MUGA) studies, we used the VENSTRA cardiac function phantom in 7 cameras (4 SOPHA- DSX-1000, 2 GE- IMAGAMMA-2001 and 1 SIEMENS- HERMES) and made 3 acquisition for each Global Left Ventricular Ejection Fraction (LVEF 30%, 60% and 80%) and for each Heart Rate (HR 40, 80 and 160 heart beat/min). The planar resolution and the planar uniformity were proper in all the equipment. Differences less than 5% were found between the acquisition and processing program. To evaluate the processing program without the acquisition parameter's influence, we used one group of these image like software phantom and test the semi-automatic software in all cameras. The semi-automatic protocol showed difference less than 3% between software. The automatic processing software of gated cardiac studies were checked with the COST-B2 software phantom; the difference between the Left Ventricle Ejection Fraction calculated by these software was less than 5% and the regional wall motion analysis was complete coincident in the 93% of the cases. The use of VENSTRA and COST- B2 phantom confirm the correct functioning of the acquisition and the LVEF calculus software of MUGA studies in the 83% of cuban nuclear medicine centers

  16. Estabilidade de ácido ascórbico em sucos de frutas frescos sob diferentes formas de armazenamento

    OpenAIRE

    Cunha, Kelly Damasceno; Silva, Priscila Ribeiro da; Costa, Ana Lígia Faria e Silva da Fonseca; Teodoro, Anderson Junger; Koblitz, Maria Gabriela Bello

    2014-01-01

    O ácido ascórbico é uma vitamina hidrossolúvel de importância nutricional há muito estabelecida, por sua atuação como cofator em diversos processos fisiológicos e como antioxidante. O ser humano depende da ingestão diária desse micronutriente, cujas principais fontes são as frutas e hortaliças. Por ser um nutriente menos estável, o ácido ascórbico sofre perdas no processamento e no armazenamento, influenciadas por diversos fatores, como pH, temperatura, presença de íons, etc. A literatura apr...

  17. Measurement and Management of the Level of Quality Control Process in SoC (System on Chip Embedded Software Development

    Directory of Open Access Journals (Sweden)

    Ki-Won Song

    2012-04-01

    Full Text Available This paper presents the process of measuring the level of quality control process to ensure the quality of delivered software package during the development cycle. The success of the project requires three pre-requisites and they constrain one another. Quality is the most important factor for successful project completion. In other words, quality should not be sacrificed for the sake of meeting cost budget or delivering within schedule. Also, cost caused by any quality issues such as defect resolution increases exponentially once the product is out of the door. Having said that, we also have to consider the schedule side of constraints for the successful project. In other words, we have no time to do a quality job and we have to compete with other competitors to ship the product to the market earlier than them. So, the quality measurement and management concept is introduced to meet the agile software development environment in conjunction with performance strategies to execute within organization. Obviously, there are many key performance indexes derivable from the actual data associated with quality control activities and it is desirable to create a quality process to integrally represent overall level of quality control activities performed while developing the software deliverables. With the quality process, it is possible to evaluate whether enough quality control activities are performed for the project officially and secure the quality of the software deliverables before it is delivered to the customers.

  18. A conceptual persistent healthcare quality improvement process for software development management.

    Science.gov (United States)

    Lin, Jen-Chiun; Su, Mei-Ju; Cheng, Po-Hsun; Weng, Yung-Chien; Chen, Sao-Jie; Lai, Jin-Shin; Lai, Feipei

    2007-01-01

    This paper illustrates a sustained conceptual service quality improvement process for the management of software development within a healthcare enterprise. Our proposed process is revised from Niland's healthcare quality information system (HQIS). This process includes functions to survey the satisfaction of system functions, describe the operation bylaws on-line, and provide on-demand training. To achieve these goals, we integrate five information systems in National Taiwan University Hospital, including healthcare information systems, health quality information system, requirement management system, executive information system, and digital learning system, to form a full Deming cycle. A preliminary user satisfaction survey showed that our outpatient information system scored an average of 71.31 in 2006.

  19. Tipo de respuesta según el genotipado del virus del papiloma humano según cobas 4800 en las lesiones asc-us HPV positivas

    OpenAIRE

    Kanjou Augé, Nadwa

    2016-01-01

    Objetivos: Evaluar el riesgo de patología subyacente en las pacientes ASC-US HPV positivas de nuestra área de influencia analizadas por el método COBAS 4800. Se estudió el riesgo de CIN2+ en el momento del diagnóstico de las mujeres ASC-US HPV positivas según el genotipo del virus del papiloma: HPV16, HPV18 ( ambos incluyendo coinfecciones) u otros HPV de alto riesgo (HR-HPV). El objetivo es detectar aquellas mujeres con citologías ASC-US HPV positivo con un mayor riesgo de progresión y por l...

  20. Model of software quality

    OpenAIRE

    Valencia Ayala, Luz Estela; Villa Sánchez, Paula Andréa; Ocampo S., Carlos Alberto

    2009-01-01

    En un mercado globalizado donde las empresas deben innovar y mejorar continuamente para crecer y ser más competitivas, es necesario tener acceso a certificaciones de calidad internacionales que les den un respaldo y puedan mantenerse en este mercado. Las certificaciones de calidad en la industria del software ayudan a las empresas a ser más productivas disminuyendo costos y tiempo en sus desarrollos. Las empresas de desarrollo de software de nuestro país en su mayoría son micro y pequeñas...

  1. Uso do ácido ascórbico no controle do escurecimento do pericarpo de lichia Use of ascorbic acid in the control of browning in the pericarp of lychees

    Directory of Open Access Journals (Sweden)

    Danieele Fabíola Pereira da Silva

    2010-06-01

    Full Text Available Um dos maiores problemas na pós-colheita da lichia é o escurecimento do pericarpo, o qual tem sido atribuído à degradação da antocianina. O objetivo do trabalho foi avaliar o efeito de diferentes doses de ácido ascórbico na prevenção do escurecimento do pericarpo e na manutenção da qualidade pós-colheita de lichia. Frutos de lichieira 'Bengal' com o pericarpo completamente vermelho foram imersos em solução com diferentes doses de ácido ascórbico (0; 5; 10; 15 e 30 mM, por 5 minutos. Após secagem à temperatura ambiente, foram acondicionados em bandejas de poliestireno, armazenados em câmara fria a 5 ± 1,2°C e 90 ± 5% de UR e avaliados a cada 4 dias, durante 12 dias. Observou-se que a perda de massa fresca foi maior nos frutos não tratados com ácido ascórbico. Independentemente da dose, o ácido ascórbico teve pouco efeito na retenção da cor vermelha do pericarpo de lichia. A atividade das enzimas polifenoloxidase e peroxidase no pericarpo foi maior com as menores doses de ácido ascórbico (0; 5 e 10 mM, entretanto observou-se escurecimento a partir do quarto dia no pericarpo dos frutos tratados com este ácido. O ácido ascórbico também não foi eficiente na manutenção da qualidade interna dos frutos.One of the biggest problems in postharvest of lychees is the pericarp browning, which has been attributed to anthocyanins degradation. The purpose of this work was to evaluate the effect of different doses of ascorbic acid to prevent pericarp browning and on the maintenance of postharvest quality of lychees. Fruits of 'Bengal' lychee with completely red pericarp were immersed in solution with different doses of ascorbic acid (0, 5, 10, 15 and 30 mM for 5 minutes. After dried, the fruits were packed in polystyrene trays, stored in cold chamber at 5 ± 1,2°C e 90 ± 5% de RH and evaluated every other 4 days for 12 days. It was observed that the loss of fresh weight was higher in non-treated with ascorbic acid

  2. Development of the quality assessment model of EHR software in family medicine practices: research based on user satisfaction

    Directory of Open Access Journals (Sweden)

    Damir Kralj

    2015-09-01

    Full Text Available Background Family medicine practices (FMPs make the basis for the Croatian health care system. Use of electronic health record (EHR software is mandatory and it plays an important role in running these practices, but important functional features still remain uneven and largely left to the will of the software developers.Objective The objective of this study was to develop a novel and comprehensive model for functional evaluation of the EHR software in FMPs, based on current world standards, models and projects, as well as on actual user satisfaction and requirements.Methods Based on previous theoretical and experimental research in this area, we made the initial framework model consisting of six basic categories as a base for online survey questionnaire. Family doctors assessed perceived software quality by using a five-point Likert-type scale. Using exploratory factor analysis and appropriate statistical methods over the collected data, the final optimal structure of the novel model was formed. Special attention was focused on the validity and quality of the novel model.Results The online survey collected a total of 384 cases. The obtained results indicate both the quality of the assessed software and the quality in use of the novel model. The intense ergonomic orientation of the novel measurement model was particularly emphasised.Conclusions The resulting novel model is multiple validated, comprehensive and universal. It could be used to assess the user-perceived quality of almost all forms of the ambulatory EHR software and therefore useful to all stakeholders in this area of the health care informatisation. 

  3. Development of the quality assessment model of EHR software in family medicine practices: research based on user satisfaction.

    Science.gov (United States)

    Kralj, Damir; Kern, Josipa; Tonkovic, Stanko; Koncar, Miroslav

    2015-09-09

    Family medicine practices (FMPs) make the basis for the Croatian health care system. Use of electronic health record (EHR) software is mandatory and it plays an important role in running these practices, but important functional features still remain uneven and largely left to the will of the software developers. The objective of this study was to develop a novel and comprehensive model for functional evaluation of the EHR software in FMPs, based on current world standards, models and projects, as well as on actual user satisfaction and requirements. Based on previous theoretical and experimental research in this area, we made the initial framework model consisting of six basic categories as a base for online survey questionnaire. Family doctors assessed perceived software quality by using a five-point Likert-type scale. Using exploratory factor analysis and appropriate statistical methods over the collected data, the final optimal structure of the novel model was formed. Special attention was focused on the validity and quality of the novel model. The online survey collected a total of 384 cases. The obtained results indicate both the quality of the assessed software and the quality in use of the novel model. The intense ergonomic orientation of the novel measurement model was particularly emphasised. The resulting novel model is multiple validated, comprehensive and universal. It could be used to assess the user-perceived quality of almost all forms of the ambulatory EHR software and therefore useful to all stakeholders in this area of the health care informatisation.

  4. Increasing quality and managing complexity in neuroinformatics software development with continuous integration

    Directory of Open Access Journals (Sweden)

    Yury V. Zaytsev

    2013-01-01

    Full Text Available High quality neuroscience research requires accurate, reliable and well maintained neuroinformatics applications. As software projects become larger, offering more functionality and developing a denser web of interdependence between their component parts, we need more sophisticated methods to manage their complexity. If complexity is allowed to get out of hand, either the quality of the software or the speed of development suffer, and in many cases both. To address this issue, here we develop a scalable, low-cost and open source solution for continuous integration (CI, a technique which ensures the quality of changes to the code base during the development procedure, rather than relying on a pre-release integration phase. We demonstrate that a CI based workflow, due to rapid feedback about code integration problems and tracking of code health measures, enabled substantial increases in productivity for a major neuroinformatics project and additional benefits for three further projects. Beyond the scope of the current study, we identify multiple areas in which CI can be employed to further increase the quality of neuroinformatics projects by improving development practices and incorporating appropriate development tools. Finally, we discuss what measures can be taken to lower the barrier for developers of neuroinformatics applications to adopt this useful technique.

  5. Increasing quality and managing complexity in neuroinformatics software development with continuous integration.

    Science.gov (United States)

    Zaytsev, Yury V; Morrison, Abigail

    2012-01-01

    High quality neuroscience research requires accurate, reliable and well maintained neuroinformatics applications. As software projects become larger, offering more functionality and developing a denser web of interdependence between their component parts, we need more sophisticated methods to manage their complexity. If complexity is allowed to get out of hand, either the quality of the software or the speed of development suffer, and in many cases both. To address this issue, here we develop a scalable, low-cost and open source solution for continuous integration (CI), a technique which ensures the quality of changes to the code base during the development procedure, rather than relying on a pre-release integration phase. We demonstrate that a CI-based workflow, due to rapid feedback about code integration problems and tracking of code health measures, enabled substantial increases in productivity for a major neuroinformatics project and additional benefits for three further projects. Beyond the scope of the current study, we identify multiple areas in which CI can be employed to further increase the quality of neuroinformatics projects by improving development practices and incorporating appropriate development tools. Finally, we discuss what measures can be taken to lower the barrier for developers of neuroinformatics applications to adopt this useful technique.

  6. Completion Report for Multi-Site Incentive MRT 2779 Implement ASC Tripod Initiative by 30SEP08

    Energy Technology Data Exchange (ETDEWEB)

    East, D; Cerutti, J; Noe, J; Cupps, K; Loncaric, J; Sturtevant, J

    2008-09-22

    This report provides documentation and evidence for the completion of the deployment of the Tripod common operating system (TripodOS, also known as and generally referred to below as TOSS). Background documents for TOSS are provided in Appendices A and B, including the initial TOSS proposal accepted by ASC HQ and Executives in July 2007 and a Governance Model defined by a Tri-Lab working group in September 2007. Appendix C contains a document that clarifies the intent and requirements for the completion criteria associated with MRT 2779. The deployment of TOSS is a Multi-Site Incentive from the ASC FY08-09 Implementation Plan due at the end of Quarter 4 in FY08.

  7. Quality control in urodynamics and the role of software support in the QC procedure.

    Science.gov (United States)

    Hogan, S; Jarvis, P; Gammie, A; Abrams, P

    2011-11-01

    This article aims to identify quality control (QC) best practice, to review published QC audits in order to identify how closely good practice is followed, and to carry out a market survey of the software features that support QC offered by urodynamics machines available in the UK. All UK distributors of urodynamic systems were contacted and asked to provide information on the software features relating to data quality of the products they supply. The results of the market survey show that the features offered by manufacturers differ greatly. Automated features, which can be turned off in most cases, include: cough recognition, detrusor contraction detection, and high pressure alerts. There are currently no systems that assess data quality based on published guidelines. A literature review of current QC guidelines for urodynamics was carried out; QC audits were included in the literature review to see how closely guidelines were being followed. This review highlights the fact that basic QC is not being carried out effectively by urodynamicists. Based on the software features currently available and the results of the literature review there is both the need and capacity for a greater degree of automation in relation to urodynamic data quality and accuracy assessment. Some progress has been made in this area and certain manufacturers have already developed automated cough detection. Copyright © 2011 Wiley Periodicals, Inc.

  8. An approach to software quality assurance for robotic inspection systems

    International Nuclear Information System (INIS)

    Kiebel, G.R.

    1993-10-01

    Software quality assurance (SQA) for robotic systems used in nuclear waste applications is vital to ensure that the systems operate safely and reliably and pose a minimum risk to humans and the environment. This paper describes the SQA approach for the control and data acquisition system for a robotic system being developed for remote surveillance and inspection of underground storage tanks (UST) at the Hanford Site

  9. Evaluation of a software package for automated quality assessment of contrast detail images-comparison with subjective visual assessment

    International Nuclear Information System (INIS)

    Pascoal, A; Lawinski, C P; Honey, I; Blake, P

    2005-01-01

    Contrast detail analysis is commonly used to assess image quality (IQ) associated with diagnostic imaging systems. Applications include routine assessment of equipment performance and optimization studies. Most frequently, the evaluation of contrast detail images involves human observers visually detecting the threshold contrast detail combinations in the image. However, the subjective nature of human perception and the variations in the decision threshold pose limits to the minimum image quality variations detectable with reliability. Objective methods of assessment of image quality such as automated scoring have the potential to overcome the above limitations. A software package (CDRAD analyser) developed for automated scoring of images produced with the CDRAD test object was evaluated. Its performance to assess absolute and relative IQ was compared with that of an average observer. Results show that the software does not mimic the absolute performance of the average observer. The software proved more sensitive and was able to detect smaller low-contrast variations. The observer's performance was superior to the software's in the detection of smaller details. Both scoring methods showed frequent agreement in the detection of image quality variations resulting from changes in kVp and KERMA detector , which indicates the potential to use the software CDRAD analyser for assessment of relative IQ

  10. Quality control in dual head γ-cameras: comparison between methods and software s used for image analysis

    International Nuclear Information System (INIS)

    Nayl E, A.; Fornasier, M. R.; De Denaro, M.; Sulieman, A.; Alkhorayef, M.; Bradley, D.

    2017-10-01

    Patient radiation dose and image quality are the main issues in nuclear medicine (Nm) procedures. Currently, many protocols are used for image acquisition and analysis of quality control (Qc) tests. National Electrical Manufacturers Association (Nema) methods and protocols are widely accepted method used for providing accurate description, measurement and reporting of γ-camera performance parameters. However, no standard software is available for image analysis. The aim os this study was to compare between the vendor Qc software analysis and three software from different developers downloaded free from internet; NMQC, Nm Tool kit and ImageJ-Nm Tool kit software. The three software are used for image analysis of some Qc tests for γ-cameras based on Nema protocols including non-uniformity evaluation. Ten non-uniformity Qc images were taken from dual head γ-camera (Siemens Symbia) installed in Trieste general hospital (Italy), and analyzed. Excel analysis was used as baseline calculation of the non-uniformity test according Nema procedures. The results of the non-uniformity analysis showed good agreement between the three independent software and excel calculation (the average differences were 0.3%, 2.9%, 1.3% and 1.6% for UFOV integral, UFOV differential, CFOV integral and CFOV differential respectively), while significant difference was detected on the analysis of the company Qc software with compare to the excel analysis (the average differences were 14.6%, 20.7%, 25.7% and 31.9% for UFOV integral, UFOV differential, CFOV integral and CFOV differential respectively). NMQC software was the best in comparison with the excel calculations. The variation in the results is due to different pixel sizes used for analysis in the three software and the γ-camera Qc software. Therefore, is important to perform the tests by the vendor Qc software as well as by independent analysis to understand the differences between the values. Moreover, the medical physicist should know

  11. Quality control in dual head γ-cameras: comparison between methods and software s used for image analysis

    Energy Technology Data Exchange (ETDEWEB)

    Nayl E, A. [Sudan Atomic Energy Commission, Radiation Safety Institute, Khartoum (Sudan); Fornasier, M. R.; De Denaro, M. [Azienda Sanitaria Universitaria Integrata di Trieste, Medical Physics Department, Via Giovanni Sai 7, 34128 Trieste (Italy); Sulieman, A. [Prince Sattam bin Abdulaziz University, College of Applied Medical Sciences, Radiology and Medical Imaging Department, P. O. Box 422, 11942 Al-Kharj (Saudi Arabia); Alkhorayef, M.; Bradley, D., E-mail: abdwsh10@hotmail.com [University of Surrey, Department of Physics, GU2-7XH Guildford, Surrey (United Kingdom)

    2017-10-15

    Patient radiation dose and image quality are the main issues in nuclear medicine (Nm) procedures. Currently, many protocols are used for image acquisition and analysis of quality control (Qc) tests. National Electrical Manufacturers Association (Nema) methods and protocols are widely accepted method used for providing accurate description, measurement and reporting of γ-camera performance parameters. However, no standard software is available for image analysis. The aim os this study was to compare between the vendor Qc software analysis and three software from different developers downloaded free from internet; NMQC, Nm Tool kit and ImageJ-Nm Tool kit software. The three software are used for image analysis of some Qc tests for γ-cameras based on Nema protocols including non-uniformity evaluation. Ten non-uniformity Qc images were taken from dual head γ-camera (Siemens Symbia) installed in Trieste general hospital (Italy), and analyzed. Excel analysis was used as baseline calculation of the non-uniformity test according Nema procedures. The results of the non-uniformity analysis showed good agreement between the three independent software and excel calculation (the average differences were 0.3%, 2.9%, 1.3% and 1.6% for UFOV integral, UFOV differential, CFOV integral and CFOV differential respectively), while significant difference was detected on the analysis of the company Qc software with compare to the excel analysis (the average differences were 14.6%, 20.7%, 25.7% and 31.9% for UFOV integral, UFOV differential, CFOV integral and CFOV differential respectively). NMQC software was the best in comparison with the excel calculations. The variation in the results is due to different pixel sizes used for analysis in the three software and the γ-camera Qc software. Therefore, is important to perform the tests by the vendor Qc software as well as by independent analysis to understand the differences between the values. Moreover, the medical physicist should know

  12. Improvement of Computer Software Quality through Software Automated Tools.

    Science.gov (United States)

    1986-08-30

    information that are returned from the tools to the human user, and the forms in which these outputs are presented. Page 2 of 4 STAGE OF DEVELOPMENT: What... AUTOMIATED SOFTWARE TOOL MONITORING SYSTEM APPENDIX 2 2-1 INTRODUCTION This document and Automated Software Tool Monitoring Program (Appendix 1) are...t Output Output features provide links from the tool to both the human user and the target machine (where applicable). They describe the types

  13. Use of a quality improvement tool, the prioritization matrix, to identify and prioritize triage software algorithm enhancement.

    Science.gov (United States)

    North, Frederick; Varkey, Prathiba; Caraballo, Pedro; Vsetecka, Darlene; Bartel, Greg

    2007-10-11

    Complex decision support software can require significant effort in maintenance and enhancement. A quality improvement tool, the prioritization matrix, was successfully used to guide software enhancement of algorithms in a symptom assessment call center.

  14. Practical experience with software tools to assess and improve the quality of existing nuclear analysis and safety codes

    International Nuclear Information System (INIS)

    Marshall, N.H.; Marwil, E.S.; Matthews, S.D.; Stacey, B.J.

    1990-01-01

    Within the constraints of schedule and budget, software tools and techniques were applied to existing FORTRAN codes determining software quality metrics and improving the code quality. Specifically discussed are INEL experiences in applying pretty printers, cross-reference analyzers, and computer aided software engineering (CASE) tools and techniques. These have provided management with measures of the risk potential for individual program modules so that rational decisions can be made on resource allocation. Selected program modules have been modified to reduce the complexity, achieve higher functional independence, and improve the code vectorization. (orig.)

  15. The six critical attributes of the next generation of quality management software systems.

    Science.gov (United States)

    Clark, Kathleen

    2011-07-01

    Driven by both the need to meet regulatory requirements and a genuine desire to drive improved quality, quality management systems encompassing standard operating procedure, corrective and preventative actions and related processes have existed for many years, both in paper and electronic form. The impact of quality management systems on 'actual' quality, however, is often reported as far less than desired. A quality management software system that moves beyond formal forms-driven processes to include a true closed loop design, manage disparate processes across the enterprise, provide support for collaborative processes and deliver insight into the overall state of control has the potential to close the gap between simply accomplishing regulatory compliance and delivering measurable improvements in quality and efficiency.

  16. Addressing Software Engineering Issues in Real-Time Software ...

    African Journals Online (AJOL)

    Addressing Software Engineering Issues in Real-Time Software ... systems, manufacturing process, process control, military, space exploration, and ... but also physical properties such as timeliness, Quality of Service and reliability.

  17. Software Quality Assurance and Controls Standard

    Science.gov (United States)

    2010-04-27

    dassurance a wor pro uc s an processes comply with predefined provisions and plans. • According to International Standard (IS) 12207 – of the 44...from document (plan) focus to process focus – Alignment with framework standard IS 12207 software life cycle (SLC) processes with exact...Books and P blications IEEE Software and Systems Engineering curriculum ABET u Certified Software Development Professional Standards ISO /IEC

  18. Quality control of the software in the JT-60 computer control system

    International Nuclear Information System (INIS)

    Isaji, Nobuaki; Kurihara, Kenichi; Kimura, Toyoaki

    1990-07-01

    The JT-60 Control System should be improved corresponding to the experimental requirements. In order to keep the integrity of the system even in the modification the concept of quality control (QC) was introduced in the software development. What we have done for QC activity are (1) to establish standard procedures of the software development, (2) to develop support tools for grasping the present status of the program structure, and (3) to develop a document system, and a source program management system. This paper reports these QC activities and their problems for the JT-60 control system. (author)

  19. The Data Quality Monitoring Software for the CMS experiment at the LHC

    CERN Document Server

    AUTHOR|(CDS)2071602

    2016-01-01

    The Data Quality Monitoring (DQM) Software is a central tool in the CMS experiment. Its flexibility allows for integration in several key environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release-Validation, to constantly validate the functionalities and the performance of the reconstruction software; in Monte Carlo productions.Since the end of data taking at a center of mass energy of 8 TeV, the environment in which the DQM lives has undergone fundamental changes. In turn, the DQM system has made significant upgrades in many areas to respond to not only the changes in infrastructure, but also the growing specialized needs of the collaboration with an emphasis on more sophisticated methods for evaluating dataquality, as well as advancing the DQM system to provide quality assessments of various Monte Carlo simulations versus data distributions, monitoring changes in physical effects due to modifications of algorithms or framework, and enabling reg...

  20. The effect of magnetic stimulation on the osteogenic and chondrogenic differentiation of human stem cells derived from the adipose tissue (hASCs)

    Science.gov (United States)

    Lima, João; Gonçalves, Ana I.; Rodrigues, Márcia T.; Reis, Rui L.; Gomes, Manuela E.

    2015-11-01

    The use of magnetic nanoparticles (MNPs) towards the musculoskeletal tissues has been the focus of many studies, regarding MNPs ability to promote and direct cellular stimulation and orient tissue responses. This is thought to be mainly achieved by mechano-responsive pathways, which can induce changes in cell behavior, including the processes of proliferation and differentiation, in response to external mechanical stimuli. Thus, the application of MNP-based strategies in tissue engineering may hold potential to propose novel solutions for cell therapy on bone and cartilage strategies to accomplish tissue regeneration. The present work aims at studying the influence of MNPs on the osteogenic and chondrogenic differentiation of human adipose derived stem cells (hASCs). MNPs were incorporated in hASCs and cultured in medium supplemented for osteogenic and chondrogenic differentiation. Cultures were maintained up to 28 days with/without an external magnetic stimulus provided by a magnetic bioreactor, to determine if the MNPs alone could affect the osteogenic or chondrogenic phenotype of the hASCs. Results indicate that the incorporation of MNPs does not negatively affect the viability nor the proliferation of hASCs. Furthermore, Alizarin Red staining evidences an enhancement in extracellular (ECM) mineralization under the influence of an external magnetic field. Although not as evident as for osteogenic differentiation, Toluidine blue and Safranin-O stainings also suggest the presence of a cartilage-like ECM with glycosaminoglycans and proteoglycans under the magnetic stimulus provided. Thus, MNPs incorporated in hASCs under the influence of an external magnetic field have the potential to induce differentiation towards the osteogenic and chondrogenic lineages.

  1. Software quality assurance and information management, October 1986 to October 1992

    International Nuclear Information System (INIS)

    Hill, I.E.

    1993-01-01

    This report describes the work carried out by Cedar Design Systems Limited under contract PECD 7/9/384. The brief for the contract was initially to provide advice on Software Quality Assurance (SQA) as part of the CEC PACOMA project. This was later extended to include further SQA and information management tasks specific to the HMIP Radioactive Waste Disposal Assessments Research Programme. (Author)

  2. SOFTWARE PROCESS IMPROVEMENT: AWARENESS, USE, AND BENEFITS IN CANADIAN SOFTWARE DEVELOPMENT FIRMS

    OpenAIRE

    CHEVERS, DELROY

    2017-01-01

    ABSTRACT Since 1982, the software development community has been concerned with the delivery of quality systems. Software process improvement (SPI) is an initiative to avoid the delivery of low quality systems. However, the awareness and adoption of SPI is low. Thus, this study examines the rate of awareness, use, and benefits of SPI initiatives in Canadian software development firms. Using SPSS as the analytical tool, this study found that 59% of Canadian software development firms are aware...

  3. Global Software Engineering: A Software Process Approach

    Science.gov (United States)

    Richardson, Ita; Casey, Valentine; Burton, John; McCaffery, Fergal

    Our research has shown that many companies are struggling with the successful implementation of global software engineering, due to temporal, cultural and geographical distance, which causes a range of factors to come into play. For example, cultural, project managementproject management and communication difficulties continually cause problems for software engineers and project managers. While the implementation of efficient software processes can be used to improve the quality of the software product, published software process models do not cater explicitly for the recent growth in global software engineering. Our thesis is that global software engineering factors should be included in software process models to ensure their continued usefulness in global organisations. Based on extensive global software engineering research, we have developed a software process, Global Teaming, which includes specific practices and sub-practices. The purpose is to ensure that requirements for successful global software engineering are stipulated so that organisations can ensure successful implementation of global software engineering.

  4. Contribuição dos modelos de qualidade e maturidade na melhoria dos processos de software Contribution of quality and maturity models to software process improvement

    Directory of Open Access Journals (Sweden)

    Antonio Carlos Tonini

    2008-01-01

    Full Text Available Grande parte das empresas desenvolvedoras de software criou seu próprio processo de trabalho. Devido à rápida expansão do mercado de software, a concorrência ocorre muito mais em custo do que em diferenciação. Para obter vantagem competitiva, as empresas devem atualizar-se continuamente na tecnologia, buscar a maturidade nos processos e eliminar a ineficiência operacional. Isso requer um envolvimento das pessoas, dos processos e da organização como um todo. O artigo discute a implementação de melhorias nos processos de software segundo os principais modelos de qualidade e de maturidade. Com base em um Estudo de Casos Múltiplos, verifica-se que a melhoria dos processos de software requer que a melhoria ocorra primeiramente entre cada um dos desenvolvedores e, a seguir, envolva os grupos de desenvolvimento e por fim, a organização como um todo. A pesquisa conclui que os modelos de qualidade e maturidade servem como orientadores do processo de melhoria.Many software development companies have developed their own work method. Due to the fast software market growth, the competition focuses more on cost than on differentiation. To achieve competitive advantage, software developer organizations must continually update their technology, reach high level process maturity and eliminate all the operational inefficiency. These procedures involve people, processes and the whole organization. The aim of the paper is to discuss software process improvement implementation according to the most important quality and maturity models. Based on a Multiple Case Study, it is verified that the software process improvement needs firstly individual improvement and, later, it involves the developer teams and the whole organization. The research concludes that the quality and maturity models must be used as improvement process drivers.

  5. Optimizing the ASC WAN: evaluating network performance tools for comparing transport protocols.

    Energy Technology Data Exchange (ETDEWEB)

    Lydick, Christopher L.

    2007-07-01

    The Advanced Simulation & Computing Wide Area Network (ASC WAN), which is a high delay-bandwidth network connection between US Department of Energy National Laboratories, is constantly being examined and evaluated for efficiency. One of the current transport-layer protocols which is used, TCP, was developed for traffic demands which are different from that on the ASC WAN. The Stream Control Transport Protocol (SCTP), on the other hand, has shown characteristics which make it more appealing to networks such as these. Most important, before considering a replacement for TCP on any network, a testing tool that performs well against certain criteria needs to be found. In order to try to find such a tool, two popular networking tools (Netperf v.2.4.3 & v.2.4.6 (OpenSS7 STREAMS), and Iperf v.2.0.6) were tested. These tools implement both TCP and SCTP and were evaluated using four metrics: (1) How effectively can the tool reach a throughput near the bandwidth? (2) How much of the CPU does the tool utilize during operation? (3) Is the tool freely and widely available? And, (4) Is the tool actively developed? Following the analysis of those tools, this paper goes further into explaining some recommendations and ideas for future work.

  6. Development of Software to Model AXAF-I Image Quality

    Science.gov (United States)

    Ahmad, Anees; Hawkins, Lamar

    1996-01-01

    This draft final report describes the work performed under the delivery order number 145 from May 1995 through August 1996. The scope of work included a number of software development tasks for the performance modeling of AXAF-I. A number of new capabilities and functions have been added to the GT software, which is the command mode version of the GRAZTRACE software, originally developed by MSFC. A structural data interface has been developed for the EAL (old SPAR) finite element analysis FEA program, which is being used by MSFC Structural Analysis group for the analysis of AXAF-I. This interface utility can read the structural deformation file from the EAL and other finite element analysis programs such as NASTRAN and COSMOS/M, and convert the data to a suitable format that can be used for the deformation ray-tracing to predict the image quality for a distorted mirror. There is a provision in this utility to expand the data from finite element models assuming 180 degrees symmetry. This utility has been used to predict image characteristics for the AXAF-I HRMA, when subjected to gravity effects in the horizontal x-ray ground test configuration. The development of the metrology data processing interface software has also been completed. It can read the HDOS FITS format surface map files, manipulate and filter the metrology data, and produce a deformation file, which can be used by GT for ray tracing for the mirror surface figure errors. This utility has been used to determine the optimum alignment (axial spacing and clocking) for the four pairs of AXAF-I mirrors. Based on this optimized alignment, the geometric images and effective focal lengths for the as built mirrors were predicted to cross check the results obtained by Kodak.

  7. Software Engineering Guidebook

    Science.gov (United States)

    Connell, John; Wenneson, Greg

    1993-01-01

    The Software Engineering Guidebook describes SEPG (Software Engineering Process Group) supported processes and techniques for engineering quality software in NASA environments. Three process models are supported: structured, object-oriented, and evolutionary rapid-prototyping. The guidebook covers software life-cycles, engineering, assurance, and configuration management. The guidebook is written for managers and engineers who manage, develop, enhance, and/or maintain software under the Computer Software Services Contract.

  8. Safety critical software development qualification

    International Nuclear Information System (INIS)

    Marron, J. E.

    2006-01-01

    With the increasing use of digital systems in control applications, customers must acquire appropriate expectations for software development and quality assurance procedures. Purchasers and users of digital systems need to understand the benefits to the supplier of effective quality systems. These systems consist not only of procedures but tools that enable automation. Without the use of automation, quality can not be assured. A software and systems quality program starts with the documents you are very familiar with. But these documents must define more than the final system. They must address specific development environment characteristics and testing capabilities. Starting with the RFP, some of the items that should be introduced are Software Configuration Management, regression testing and defect tracking. The digital system customer is in the best position to enforce the use of software and systems quality programs by including them in project requirements as early as the Purchase Order. The customer's understanding of the full scope and implementation of a software quality program is essential to achieving the quality necessary in nuclear projects, and, incidentally, completing those projects on schedule. (authors)

  9. Predicting Software Suitability Using a Bayesian Belief Network

    Science.gov (United States)

    Beaver, Justin M.; Schiavone, Guy A.; Berrios, Joseph S.

    2005-01-01

    The ability to reliably predict the end quality of software under development presents a significant advantage for a development team. It provides an opportunity to address high risk components earlier in the development life cycle, when their impact is minimized. This research proposes a model that captures the evolution of the quality of a software product, and provides reliable forecasts of the end quality of the software being developed in terms of product suitability. Development team skill, software process maturity, and software problem complexity are hypothesized as driving factors of software product quality. The cause-effect relationships between these factors and the elements of software suitability are modeled using Bayesian Belief Networks, a machine learning method. This research presents a Bayesian Network for software quality, and the techniques used to quantify the factors that influence and represent software quality. The developed model is found to be effective in predicting the end product quality of small-scale software development efforts.

  10. Software testing concepts and operations

    CERN Document Server

    Mili, Ali

    2015-01-01

    Explores and identifies the main issues, concepts, principles and evolution of software testing, including software quality engineering and testing concepts, test data generation, test deployment analysis, and software test managementThis book examines the principles, concepts, and processes that are fundamental to the software testing function. This book is divided into five broad parts. Part I introduces software testing in the broader context of software engineering and explores the qualities that testing aims to achieve or ascertain, as well as the lifecycle of software testing. Part II c

  11. System for inspection and quality assurance of software - A knowledge-based experiment with code understanding

    International Nuclear Information System (INIS)

    Das, B.K.

    1989-01-01

    This paper describes a knowledge-based prototype that inspects and quality-assures software components. The prototype model, which offers a singular representation of these components, is used to automate both the mechanical and nonmechanical activities in the quality assurance (QA) process. It is shown that the prototype, in addition to automating the QA process, provides a novel approach to understanding code. These approaches are compared with recent approaches to code understanding. The paper also presents the results of an experiment with several classes of nonsyntactic bugs. It is argued that a structured environment, as facilitated by this unique architecture, along with software development standards used in the QA process, is essential for meaningful analysis of code. 8 refs

  12. Report on the working conference on requirements engineering: foundation for software quality (REFSQ'09)

    NARCIS (Netherlands)

    Glinz, Martin; Heymans, Patrick; Persson, Anne; Sindre, Guttorm; Aurum, Aybüke; Madhavji, Nazim; Madhavji, N.; Paech, Barbara; Regev, Gil; Wieringa, Roelf J.

    This report summarizes the presentations and discussions at REFSQ’09, the 15th International Working Conference on Requirements Engineering: Foundation for Software Quality which was held on June 8-9, 2009 in Amsterdam, The Netherlands.

  13. Space Shuttle Program Primary Avionics Software System (PASS) Success Legacy - Quality and Reliability Date

    Science.gov (United States)

    Orr, James K.; Peltier, Daryl

    2010-01-01

    Thsi slide presentation reviews the avionics software system on board the space shuttle, with particular emphasis on the quality and reliability. The Primary Avionics Software System (PASS) provides automatic and fly-by-wire control of critical shuttle systems which executes in redundant computers. Charts given show the number of space shuttle flights vs time, PASS's development history, and other charts that point to the reliability of the system's development. The reliability of the system is also compared to predicted reliability.

  14. On Quality and Measures in Software Engineering

    Science.gov (United States)

    Bucur, Ion I.

    2006-01-01

    Complexity measures are mainly used to estimate vital information about reliability and maintainability of software systems from regular analysis of the source code. Such measures also provide constant feedback during a software project to assist the control of the development procedure. There exist several models to classify a software product's…

  15. The deoxyhypusine synthase mutant dys1-1 reveals the association of eIF5A and Asc1 with cell wall integrity.

    Directory of Open Access Journals (Sweden)

    Fabio Carrilho Galvão

    Full Text Available The putative eukaryotic translation initiation factor 5A (eIF5A is a highly conserved protein among archaea and eukaryotes that has recently been implicated in the elongation step of translation. eIF5A undergoes an essential and conserved posttranslational modification at a specific lysine to generate the residue hypusine. The enzymes deoxyhypusine synthase (Dys1 and deoxyhypusine hydroxylase (Lia1 catalyze this two-step modification process. Although several Saccharomyces cerevisiae eIF5A mutants have importantly contributed to the study of eIF5A function, no conditional mutant of Dys1 has been described so far. In this study, we generated and characterized the dys1-1 mutant, which showed a strong depletion of mutated Dys1 protein, resulting in more than 2-fold decrease in hypusine levels relative to the wild type. The dys1-1 mutant demonstrated a defect in total protein synthesis, a defect in polysome profile indicative of a translation elongation defect and a reduced association of eIF5A with polysomes. The growth phenotype of dys1-1 mutant is severe, growing only in the presence of 1 M sorbitol, an osmotic stabilizer. Although this phenotype is characteristic of Pkc1 cell wall integrity mutants, the sorbitol requirement from dys1-1 is not associated with cell lysis. We observed that the dys1-1 genetically interacts with the sole yeast protein kinase C (Pkc1 and Asc1, a component of the 40S ribosomal subunit. The dys1-1 mutant was synthetically lethal in combination with asc1Δ and overexpression of TIF51A (eIF5A or DYS1 is toxic for an asc1Δ strain. Moreover, eIF5A is more associated with translating ribosomes in the absence of Asc1 in the cell. Finally, analysis of the sensitivity to cell wall-perturbing compounds revealed a more similar behavior of the dys1-1 and asc1Δ mutants in comparison with the pkc1Δ mutant. These data suggest a correlated role for eIF5A and Asc1 in coordinating the translational control of a subset of m

  16. The effect of magnetic stimulation on the osteogenic and chondrogenic differentiation of human stem cells derived from the adipose tissue (hASCs)

    Energy Technology Data Exchange (ETDEWEB)

    Lima, João; Gonçalves, Ana I.; Rodrigues, Márcia T.; Reis, Rui L. [3Bs Research Group–Biomaterials, Biodegradables and Biomimetics, University of Minho, Guimarães (Portugal); ICVS/3Bs–PT Government Associate Laboratory, Braga/Guimarães (Portugal); Gomes, Manuela E., E-mail: megomes@dep.uminho.pt [3Bs Research Group–Biomaterials, Biodegradables and Biomimetics, University of Minho, Guimarães (Portugal); ICVS/3Bs–PT Government Associate Laboratory, Braga/Guimarães (Portugal)

    2015-11-01

    The use of magnetic nanoparticles (MNPs) towards the musculoskeletal tissues has been the focus of many studies, regarding MNPs ability to promote and direct cellular stimulation and orient tissue responses. This is thought to be mainly achieved by mechano-responsive pathways, which can induce changes in cell behavior, including the processes of proliferation and differentiation, in response to external mechanical stimuli. Thus, the application of MNP-based strategies in tissue engineering may hold potential to propose novel solutions for cell therapy on bone and cartilage strategies to accomplish tissue regeneration. The present work aims at studying the influence of MNPs on the osteogenic and chondrogenic differentiation of human adipose derived stem cells (hASCs). MNPs were incorporated in hASCs and cultured in medium supplemented for osteogenic and chondrogenic differentiation. Cultures were maintained up to 28 days with/without an external magnetic stimulus provided by a magnetic bioreactor, to determine if the MNPs alone could affect the osteogenic or chondrogenic phenotype of the hASCs. Results indicate that the incorporation of MNPs does not negatively affect the viability nor the proliferation of hASCs. Furthermore, Alizarin Red staining evidences an enhancement in extracellular (ECM) mineralization under the influence of an external magnetic field. Although not as evident as for osteogenic differentiation, Toluidine blue and Safranin-O stainings also suggest the presence of a cartilage-like ECM with glycosaminoglycans and proteoglycans under the magnetic stimulus provided. Thus, MNPs incorporated in hASCs under the influence of an external magnetic field have the potential to induce differentiation towards the osteogenic and chondrogenic lineages. - Highlights: • Cellular viability was not negatively influenced by the nanoparticles. • Chondrogenic medium influences more the synthesis of cartilage-like ECM than MNPs. • Synergetic effect among

  17. Genomic presence of gadD1 glutamate decarboxylase correlates with the organization of ascB-dapE internalin cluster in Listeria monocytogenes.

    Science.gov (United States)

    Chen, Jianshun; Fang, Chun; Zheng, Tianlun; Zhu, Ningyu; Bei, Yijiang; Fang, Weihuan

    2012-02-01

    The ability to survive and proliferate in acidic environments is a prerequisite for the infection of Listeria monocytogenes. The glutamate decarboxylase (GAD) system is responsible for acid resistance, and three GAD homologs have been identified in L. monocytogenes: gadD1, gadD2, and gadD3. To examine whether GAD genes are specific to lineage, serovar, or certain subpopulation, we performed a systematic investigation on the prevalence of GAD genes in 164 L. monocytogenes. In contrast to gadD2 and gadD3 conserved in all L. monocytogenes strains, gadD1 was identified in 36.6% (60/164) of L. monocytogenes strains, including all serovar 1/2c and 68.5% (37/54) of serovar 1/2a strains, as well as a small fraction of serovar 1/2b (3.4%, 1/29) and lineage III (13.8%, 4/29) strains. All serovar 4b and lineage IV strains lacked this gene. According to the ascB-dapE structure, L. monocytogenes strains were classified into four subpopulations, carrying inlC2DE, inlGC2DE, inlGHE, or no internalin cluster, respectively. All L. monocytogenes strains with inlGC2DE or inlGHE pattern harbored gadD1, whereas those bearing inlC2DE or no internalin cluster between ascB and dapE lacked gadD1. In addition, other five non-monocytogenes Listeria species lacking ascB-dapE internalin cluster were gadD1-negative. Overall, the presence of gadD1 is not fully dependent on lineages or serovars but correlates with ascB-dapE internalin profiles, suggesting gadD1 might have co-evolved with the ascB-dapE internalin cluster in the primitive L. monocytogenes before divergence of serovars.

  18. The effect of magnetic stimulation on the osteogenic and chondrogenic differentiation of human stem cells derived from the adipose tissue (hASCs)

    International Nuclear Information System (INIS)

    Lima, João; Gonçalves, Ana I.; Rodrigues, Márcia T.; Reis, Rui L.; Gomes, Manuela E.

    2015-01-01

    The use of magnetic nanoparticles (MNPs) towards the musculoskeletal tissues has been the focus of many studies, regarding MNPs ability to promote and direct cellular stimulation and orient tissue responses. This is thought to be mainly achieved by mechano-responsive pathways, which can induce changes in cell behavior, including the processes of proliferation and differentiation, in response to external mechanical stimuli. Thus, the application of MNP-based strategies in tissue engineering may hold potential to propose novel solutions for cell therapy on bone and cartilage strategies to accomplish tissue regeneration. The present work aims at studying the influence of MNPs on the osteogenic and chondrogenic differentiation of human adipose derived stem cells (hASCs). MNPs were incorporated in hASCs and cultured in medium supplemented for osteogenic and chondrogenic differentiation. Cultures were maintained up to 28 days with/without an external magnetic stimulus provided by a magnetic bioreactor, to determine if the MNPs alone could affect the osteogenic or chondrogenic phenotype of the hASCs. Results indicate that the incorporation of MNPs does not negatively affect the viability nor the proliferation of hASCs. Furthermore, Alizarin Red staining evidences an enhancement in extracellular (ECM) mineralization under the influence of an external magnetic field. Although not as evident as for osteogenic differentiation, Toluidine blue and Safranin-O stainings also suggest the presence of a cartilage-like ECM with glycosaminoglycans and proteoglycans under the magnetic stimulus provided. Thus, MNPs incorporated in hASCs under the influence of an external magnetic field have the potential to induce differentiation towards the osteogenic and chondrogenic lineages. - Highlights: • Cellular viability was not negatively influenced by the nanoparticles. • Chondrogenic medium influences more the synthesis of cartilage-like ECM than MNPs. • Synergetic effect among

  19. A measurement system for large, complex software programs

    Science.gov (United States)

    Rone, Kyle Y.; Olson, Kitty M.; Davis, Nathan E.

    1994-01-01

    This paper describes measurement systems required to forecast, measure, and control activities for large, complex software development and support programs. Initial software cost and quality analysis provides the foundation for meaningful management decisions as a project evolves. In modeling the cost and quality of software systems, the relationship between the functionality, quality, cost, and schedule of the product must be considered. This explicit relationship is dictated by the criticality of the software being developed. This balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and developers with respect to the processes being employed.

  20. Application of a newly developed software program for image quality assessment in cone-beam computed tomography.

    Science.gov (United States)

    de Oliveira, Marcus Vinicius Linhares; Santos, António Carvalho; Paulo, Graciano; Campos, Paulo Sergio Flores; Santos, Joana

    2017-06-01

    The purpose of this study was to apply a newly developed free software program, at low cost and with minimal time, to evaluate the quality of dental and maxillofacial cone-beam computed tomography (CBCT) images. A polymethyl methacrylate (PMMA) phantom, CQP-IFBA, was scanned in 3 CBCT units with 7 protocols. A macro program was developed, using the free software ImageJ, to automatically evaluate the image quality parameters. The image quality evaluation was based on 8 parameters: uniformity, the signal-to-noise ratio (SNR), noise, the contrast-to-noise ratio (CNR), spatial resolution, the artifact index, geometric accuracy, and low-contrast resolution. The image uniformity and noise depended on the protocol that was applied. Regarding the CNR, high-density structures were more sensitive to the effect of scanning parameters. There were no significant differences between SNR and CNR in centered and peripheral objects. The geometric accuracy assessment showed that all the distance measurements were lower than the real values. Low-contrast resolution was influenced by the scanning parameters, and the 1-mm rod present in the phantom was not depicted in any of the 3 CBCT units. Smaller voxel sizes presented higher spatial resolution. There were no significant differences among the protocols regarding artifact presence. This software package provided a fast, low-cost, and feasible method for the evaluation of image quality parameters in CBCT.

  1. Application of a newly developed software program for image quality assessment in cone-beam computed tomography

    International Nuclear Information System (INIS)

    De Oliveira, Marcus Vinicius Linhares; Campos, Paulo Sergio Flores; Paulo, Graciano; Santos, Antonio Carvalho; Santos, Joana

    2017-01-01

    The purpose of this study was to apply a newly developed free software program, at low cost and with minimal time, to evaluate the quality of dental and maxillofacial cone-beam computed tomography (CBCT) images. A polymethyl methacrylate (PMMA) phantom, CQP-IFBA, was scanned in 3 CBCT units with 7 protocols. A macro program was developed, using the free software ImageJ, to automatically evaluate the image quality parameters. The image quality evaluation was based on 8 parameters: uniformity, the signal-to-noise ratio (SNR), noise, the contrast-to-noise ratio (CNR), spatial resolution, the artifact index, geometric accuracy, and low-contrast resolution. The image uniformity and noise depended on the protocol that was applied. Regarding the CNR, high-density structures were more sensitive to the effect of scanning parameters. There were no significant differences between SNR and CNR in centered and peripheral objects. The geometric accuracy assessment showed that all the distance measurements were lower than the real values. Low-contrast resolution was influenced by the scanning parameters, and the 1-mm rod present in the phantom was not depicted in any of the 3 CBCT units. Smaller voxel sizes presented higher spatial resolution. There were no significant differences among the protocols regarding artifact presence. This software package provided a fast, low-cost, and feasible method for the evaluation of image quality parameters in CBCT

  2. Application of a newly developed software program for image quality assessment in cone-beam computed tomography

    Energy Technology Data Exchange (ETDEWEB)

    De Oliveira, Marcus Vinicius Linhares; Campos, Paulo Sergio Flores [Federal Institute of Bahia, Salvador (Brazil); Paulo, Graciano; Santos, Antonio Carvalho; Santos, Joana [Coimbra Health School, Polytechnic Institute of Coimbra, Coimbra (Portugal)

    2017-06-15

    The purpose of this study was to apply a newly developed free software program, at low cost and with minimal time, to evaluate the quality of dental and maxillofacial cone-beam computed tomography (CBCT) images. A polymethyl methacrylate (PMMA) phantom, CQP-IFBA, was scanned in 3 CBCT units with 7 protocols. A macro program was developed, using the free software ImageJ, to automatically evaluate the image quality parameters. The image quality evaluation was based on 8 parameters: uniformity, the signal-to-noise ratio (SNR), noise, the contrast-to-noise ratio (CNR), spatial resolution, the artifact index, geometric accuracy, and low-contrast resolution. The image uniformity and noise depended on the protocol that was applied. Regarding the CNR, high-density structures were more sensitive to the effect of scanning parameters. There were no significant differences between SNR and CNR in centered and peripheral objects. The geometric accuracy assessment showed that all the distance measurements were lower than the real values. Low-contrast resolution was influenced by the scanning parameters, and the 1-mm rod present in the phantom was not depicted in any of the 3 CBCT units. Smaller voxel sizes presented higher spatial resolution. There were no significant differences among the protocols regarding artifact presence. This software package provided a fast, low-cost, and feasible method for the evaluation of image quality parameters in CBCT.

  3. Software Validation in ATLAS

    International Nuclear Information System (INIS)

    Hodgkinson, Mark; Seuster, Rolf; Simmons, Brinick; Sherwood, Peter; Rousseau, David

    2012-01-01

    The ATLAS collaboration operates an extensive set of protocols to validate the quality of the offline software in a timely manner. This is essential in order to process the large amounts of data being collected by the ATLAS detector in 2011 without complications on the offline software side. We will discuss a number of different strategies used to validate the ATLAS offline software; running the ATLAS framework software, Athena, in a variety of configurations daily on each nightly build via the ATLAS Nightly System (ATN) and Run Time Tester (RTT) systems; the monitoring of these tests and checking the compilation of the software via distributed teams of rotating shifters; monitoring of and follow up on bug reports by the shifter teams and periodic software cleaning weeks to improve the quality of the offline software further.

  4. Consumo de ácido ascórbico y niveles séricos en hombres adultos fumadores y no fumadores de la CD. de Hermosillo, Sonora, México

    OpenAIRE

    Méndez E, Rosa Olivia; Wyatt, C. Jane; Saavedra, Javier; Ornelas, Alicia

    2002-01-01

    El ácido ascórbico es uno de las antioxidantes más importantes a nivel extracelular, sin embargo su papel preventivo de enfermedades degenerativas se puede comprometer al disminuir sus niveles séricos. Bajos valores de ácido ascórbico sérico se han reportado en hombres fumadores. En el presente estudio se estimó la ingestión de ácido ascórbico en 25 hombres adultos sanos de Hermosillo, Sonora, México, divididos en dos grupos: fumadores y no fumadores; se compararon los niveles séricos de ácid...

  5. An Approach for the Implementation of Software Quality Models Adpoting CERTICS and CMMI-DEV

    Directory of Open Access Journals (Sweden)

    GARCIA, F.W.

    2015-12-01

    Full Text Available This paper proposes a mapping between two product quality and software processes models used in the industry, the CERTICS national model and the CMMI-DEV international model. The stages of mapping are presented step by step, as well as the mapping review, which had the cooperation of one specialist in CERTICS and CMMI-DEV models. It aims to correlate the structures of the two models in order to facilitate and reduce the implementation time and costs, and to stimulate the execution of multi-model implementations in software developers companies.

  6. Parameter-based estimation of CT dose index and image quality using an in-house android™-based software

    International Nuclear Information System (INIS)

    Mubarok, S; Lubis, L E; Pawiro, S A

    2016-01-01

    Compromise between radiation dose and image quality is essential in the use of CT imaging. CT dose index (CTDI) is currently the primary dosimetric formalisms in CT scan, while the low and high contrast resolutions are aspects indicating the image quality. This study was aimed to estimate CTDI vol and image quality measures through a range of exposure parameters variation. CTDI measurements were performed using PMMA (polymethyl methacrylate) phantom of 16 cm diameter, while the image quality test was conducted by using catphan ® 600. CTDI measurements were carried out according to IAEA TRS 457 protocol using axial scan mode, under varied parameters of tube voltage, collimation or slice thickness, and tube current. Image quality test was conducted accordingly under the same exposure parameters with CTDI measurements. An Android™ based software was also result of this study. The software was designed to estimate the value of CTDI vol with maximum difference compared to actual CTDI vol measurement of 8.97%. Image quality can also be estimated through CNR parameter with maximum difference to actual CNR measurement of 21.65%. (paper)

  7. Prediction of cervical intraepithelial neoplasia grade 2+ (CIN2+) using HPV DNA testing after a diagnosis of atypical squamous cell of undetermined significance (ASC-US) in Catalonia, Spain.

    Science.gov (United States)

    Ibáñez, Raquel; Moreno-Crespi, Judit; Sardà, Montserrat; Autonell, Josefina; Fibla, Montserrat; Gutiérrez, Cristina; Lloveras, Belen; Alejo, María; Català, Isabel; Alameda, Francesc; Casas, Miquel; Bosch, F Xavier; de Sanjosé, Silvia

    2012-01-26

    A protocol for cervical cancer screening among sexually active women 25 to 65 years of age was introduced in 2006 in Catalonia, Spain to increase coverage and to recommend a 3-year-interval between screening cytology. In addition, Human Papillomavirus (HPV) was offered as a triage test for women with a diagnosis of atypical squamous cells of undetermined significance (ASC-US). HPV testing was recommended within 3 months of ASC-US diagnosis. According to protocol, HPV negative women were referred to regular screening including a cytological exam every 3 years while HPV positive women were referred to colposcopy and closer follow-up. We evaluated the implementation of the protocol and the prediction of HPV testing as a triage tool for cervical intraepithelial lesions grade two or worse (CIN2+) in women with a cytological diagnosis of ASC-US. During 2007-08 a total of 611 women from five reference laboratories in Catalonia with a novel diagnosis of ASC-US were referred for high risk HPV (hrHPV) triage using high risk Hybrid Capture version 2. Using routine record linkage data, women were followed for 3 years to evaluate hrHPV testing efficacy for predicting CIN2+ cases. Logistic regression analysis was used to estimate the odds ratio for CIN2 +. Among the 611 women diagnosed with ASC-US, 493 (80.7%) had at least one follow-up visit during the study period. hrHPV was detected in 48.3% of the women at study entry (mean age 35.2 years). hrHPV positivity decreased with increasing age from 72.6% among women younger than 25 years to 31.6% in women older than 54 years (p < 0.01). At the end of the 3 years follow-up period, 37 women with a diagnosis of CIN2+ (18 CIN2, 16 CIN3, 2 cancers, and 1 with high squamous intraepithelial lesions--HSIL) were identified and all but one had a hrHPV positive test at study entry. Sensitivity to detect CIN2+ of hrHPV was 97.2% (95%confidence interval (CI) = 85.5-99.9) and specificity was 68.3% (95%CI = 63.1-73.2). The odds ratio for CIN2

  8. The State of Software for Evolutionary Biology.

    Science.gov (United States)

    Darriba, Diego; Flouri, Tomáš; Stamatakis, Alexandros

    2018-05-01

    With Next Generation Sequencing data being routinely used, evolutionary biology is transforming into a computational science. Thus, researchers have to rely on a growing number of increasingly complex software. All widely used core tools in the field have grown considerably, in terms of the number of features as well as lines of code and consequently, also with respect to software complexity. A topic that has received little attention is the software engineering quality of widely used core analysis tools. Software developers appear to rarely assess the quality of their code, and this can have potential negative consequences for end-users. To this end, we assessed the code quality of 16 highly cited and compute-intensive tools mainly written in C/C++ (e.g., MrBayes, MAFFT, SweepFinder, etc.) and JAVA (BEAST) from the broader area of evolutionary biology that are being routinely used in current data analysis pipelines. Because, the software engineering quality of the tools we analyzed is rather unsatisfying, we provide a list of best practices for improving the quality of existing tools and list techniques that can be deployed for developing reliable, high quality scientific software from scratch. Finally, we also discuss journal as well as science policy and, more importantly, funding issues that need to be addressed for improving software engineering quality as well as ensuring support for developing new and maintaining existing software. Our intention is to raise the awareness of the community regarding software engineering quality issues and to emphasize the substantial lack of funding for scientific software development.

  9. Software Architecture Evolution

    Science.gov (United States)

    Barnes, Jeffrey M.

    2013-01-01

    Many software systems eventually undergo changes to their basic architectural structure. Such changes may be prompted by new feature requests, new quality attribute requirements, changing technology, or other reasons. Whatever the causes, architecture evolution is commonplace in real-world software projects. Today's software architects, however,…

  10. ASC-J9 Suppresses Castration-Resistant Prostate Cancer Growth through Degradation of Full-length and Splice Variant Androgen Receptors

    Directory of Open Access Journals (Sweden)

    Shinichi Yamashita

    2012-01-01

    Full Text Available Early studies suggested androgen receptor (AR splice variants might contribute to the progression of prostate cancer (PCa into castration resistance. However, the therapeutic strategy to target these AR splice variants still remains unresolved. Through tissue survey of tumors from the same patients before and after castration resistance, we found that the expression of AR3, a major AR splice variant that lacks the AR ligand-binding domain, was substantially increased after castration resistance development. The currently used antiandrogen, Casodex, showed little growth suppression in CWR22Rv1 cells. Importantly, we found that AR degradation enhancer ASC-J9 could degrade both full-length (fAR and AR3 in CWR22Rv1 cells as well as in C4-2 and C81 cells with addition of AR3. The consequences of such degradation of both fAR and AR3 might then result in the inhibition of AR transcriptional activity and cell growth in vitro. More importantly, suppression of AR3 specifically by short-hairpin AR3 or degradation of AR3 by ASC-J9 resulted in suppression of AR transcriptional activity and cell growth in CWR22Rv1-fARKD (fAR knockdown cells in which DHT failed to induce, suggesting the importance of targeting AR3. Finally, we demonstrated the in vivo therapeutic effects of ASC-J9 by showing the inhibition of PCa growth using the xenografted model of CWR22Rv1 cells orthotopically implanted into castrated nude mice with undetectable serum testosterone. These results suggested that targeting both fAR- and AR3-mediated PCa growth by ASC-J9 may represent the novel therapeutic approach to suppress castration-resistant PCa. Successful clinical trials targeting both fAR and AR3 may help us to battle castration-resistant PCa in the future.

  11. Great software debates

    CERN Document Server

    Davis, A

    2004-01-01

    The industry’s most outspoken and insightful critic explains how the software industry REALLY works. In Great Software Debates, Al Davis, shares what he has learned about the difference between the theory and the realities of business and encourages you to question and think about software engineering in ways that will help you succeed where others fail. In short, provocative essays, Davis fearlessly reveals the truth about process improvement, productivity, software quality, metrics, agile development, requirements documentation, modeling, software marketing and sales, empiricism, start-up financing, software research, requirements triage, software estimation, and entrepreneurship.

  12. Systematic Software Development

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Méndez Fernández, Daniel

    2015-01-01

    The speed of innovation and the global allocation of resources to accelerate development or to reduce cost put pressure on the software industry. In the global competition, especially so-called high-price countries have to present arguments why the higher development cost is justified and what...... makes these countries an attractive host for software companies. Often, high-quality engineering and excellent quality of products, e.g., machinery and equipment, are mentioned. Yet, the question is: Can such arguments be also found for the software industry? We aim at investigating the degree...... of professionalism and systematization of software development to draw a map of strengths and weaknesses. To this end, we conducted as a first step an exploratory survey in Germany, presented in this paper. In this survey, we focused on the perceived importance of the two general software engineering process areas...

  13. Framework for Small-Scale Experiments in Software Engineering: Guidance and Control Software Project: Software Engineering Case Study

    Science.gov (United States)

    Hayhurst, Kelly J.

    1998-01-01

    Software is becoming increasingly significant in today's critical avionics systems. To achieve safe, reliable software, government regulatory agencies such as the Federal Aviation Administration (FAA) and the Department of Defense mandate the use of certain software development methods. However, little scientific evidence exists to show a correlation between software development methods and product quality. Given this lack of evidence, a series of experiments has been conducted to understand why and how software fails. The Guidance and Control Software (GCS) project is the latest in this series. The GCS project is a case study of the Requirements and Technical Concepts for Aviation RTCA/DO-178B guidelines, Software Considerations in Airborne Systems and Equipment Certification. All civil transport airframe and equipment vendors are expected to comply with these guidelines in building systems to be certified by the FAA for use in commercial aircraft. For the case study, two implementations of a guidance and control application were developed to comply with the DO-178B guidelines for Level A (critical) software. The development included the requirements, design, coding, verification, configuration management, and quality assurance processes. This paper discusses the details of the GCS project and presents the results of the case study.

  14. A prolog implementation of pattern search to optimize software quality assurance

    OpenAIRE

    Buzzard, Raymond Karl

    1990-01-01

    Approved for public release, distribution is unlimited Quality Assurance (QA) is a critical factor in the development of successful software systems. Through the use of various QA tools, project managers can ensure that a desired level of performance and reliability is built into the system. However, these tools are not without cost. Project managers must weight all QA costs and benefits for each development environment before weigh all QA costs and benefits for each development environmen...

  15. DIDACTIC PRINCIPLES AND PSYCHOLOGICAL CHARACTERISTICS IN DEFINITION OF QUALITY OF SOFTWARE TOOLS FOR EDUCATIONAL PURPOSE IN THE GENERAL EDUCATIONAL ENVIRONMENT OF UKRAINE

    Directory of Open Access Journals (Sweden)

    Maryna V. Pirko

    2011-02-01

    Full Text Available The fundamental feature of economy of postindustrial society is the knowledge that represents the basic source of competitive advantage. In the article the circle of didactic, psychological indicators in researches of problems of achievement of a high degree of quality of education and educational services is considered and described. The attention is paid to pedagogical requirements of the given period which are a standard substantiation in orientations for quality estimation of software tools for educational purpose of the general educational environment in Ukraine. The scheme of internal model of maintenance of quality of software tools for educational purpose is considered, the aspects integrated by internal model of quality of software for educational purpose are listed. The article describes the directions of researches in the conditions of formation of the global international educational environment and uniform information space of  education system taking into account the growth of availability of educational services. It is specified the main principles in the organization of pedagogical software tools.

  16. Non-intrusive Instance Level Software Composition

    NARCIS (Netherlands)

    Hatun, Kardelen

    2014-01-01

    A software system is comprised of parts, which interact through shared interfaces. Certain qualities of integration, such as loose-coupling, requiring minimal changes to the software and fine-grained localisation of dependencies, have impact on the overall software quality. Current general-purpose

  17. SU-E-T-103: Development and Implementation of Web Based Quality Control Software

    International Nuclear Information System (INIS)

    Studinski, R; Taylor, R; Angers, C; La Russa, D; Clark, B

    2014-01-01

    Purpose: Historically many radiation medicine programs have maintained their Quality Control (QC) test results in paper records or Microsoft Excel worksheets. Both these approaches represent significant logistical challenges, and are not predisposed to data review and approval. It has been our group's aim to develop and implement web based software designed not just to record and store QC data in a centralized database, but to provide scheduling and data review tools to help manage a radiation therapy clinics Equipment Quality control program. Methods: The software was written in the Python programming language using the Django web framework. In order to promote collaboration and validation from other centres the code was made open source and is freely available to the public via an online source code repository. The code was written to provide a common user interface for data entry, formalize the review and approval process, and offer automated data trending and process control analysis of test results. Results: As of February 2014, our installation of QAtrack+ has 180 tests defined in its database and has collected ∼22 000 test results, all of which have been reviewed and approved by a physicist via QATrack+'s review tools. These results include records for quality control of Elekta accelerators, CT simulators, our brachytherapy programme, TomoTherapy and Cyberknife units. Currently at least 5 other centres are known to be running QAtrack+ clinically, forming the start of an international user community. Conclusion: QAtrack+ has proven to be an effective tool for collecting radiation therapy QC data, allowing for rapid review and trending of data for a wide variety of treatment units. As free and open source software, all source code, documentation and a bug tracker are available to the public at https://bitbucket.org/tohccmedphys/qatrackplus/

  18. SU-E-T-103: Development and Implementation of Web Based Quality Control Software

    Energy Technology Data Exchange (ETDEWEB)

    Studinski, R; Taylor, R; Angers, C; La Russa, D; Clark, B [The Ottawa Hospital Regional Cancer Ctr., Ottawa, ON (Canada)

    2014-06-01

    Purpose: Historically many radiation medicine programs have maintained their Quality Control (QC) test results in paper records or Microsoft Excel worksheets. Both these approaches represent significant logistical challenges, and are not predisposed to data review and approval. It has been our group's aim to develop and implement web based software designed not just to record and store QC data in a centralized database, but to provide scheduling and data review tools to help manage a radiation therapy clinics Equipment Quality control program. Methods: The software was written in the Python programming language using the Django web framework. In order to promote collaboration and validation from other centres the code was made open source and is freely available to the public via an online source code repository. The code was written to provide a common user interface for data entry, formalize the review and approval process, and offer automated data trending and process control analysis of test results. Results: As of February 2014, our installation of QAtrack+ has 180 tests defined in its database and has collected ∼22 000 test results, all of which have been reviewed and approved by a physicist via QATrack+'s review tools. These results include records for quality control of Elekta accelerators, CT simulators, our brachytherapy programme, TomoTherapy and Cyberknife units. Currently at least 5 other centres are known to be running QAtrack+ clinically, forming the start of an international user community. Conclusion: QAtrack+ has proven to be an effective tool for collecting radiation therapy QC data, allowing for rapid review and trending of data for a wide variety of treatment units. As free and open source software, all source code, documentation and a bug tracker are available to the public at https://bitbucket.org/tohccmedphys/qatrackplus/.

  19. Next generation software process improvement

    OpenAIRE

    Turnas, Daniel

    2003-01-01

    Approved for public release; distribution is unlimited Software is often developed under a process that can at best be described as ad hoc. While it is possible to develop quality software under an ad hoc process, formal processes can be developed to help increase the overall quality of the software under development. The application of these processes allows for an organization to mature. The software maturity level, and process improvement, of an organization can be measured with the Cap...

  20. Estimation of Remained defects in a Safety-Critical Software using Bayesian Belief Network of Software Development Life Cycle

    International Nuclear Information System (INIS)

    Lee, Seung Jun; Jung, Wondea Jung

    2015-01-01

    Some researchers recognized Bayesian belief network (BBN) method to be a promising method of quantifying software reliability. Brookhaven National Laboratory (BNL) comprehensively reviewed various quantitative software reliability methods to identify the most promising methods for use in probabilistic safety assessments (PSAs) of digital systems of NPPs against a set of the most desirable characteristics developed therein. BBNs are recognized as a promising way of quantifying software reliability and are useful for integrating many aspects of software engineering and quality assurance. The method explicitly incorporates important factors relevant to reliability, such as the quality of the developer, the development process, problem complexity, testing effort, and the operation environment. In this work, a BBN model was developed to estimate the number of remained defects in a safety-critical software based on the quality evaluation of software development life cycle (SDLC). Even though a number of software reliability evaluation methods exist, none of them can be applicable to the safety-critical software in an NPP because software quality in terms of PDF is required for the PSA

  1. Software quality assurance in the 1996 performance assessment for the Waste Isolation Pilot Plant

    International Nuclear Information System (INIS)

    Froehlich, G.K.; Ogden, H.C.; Byle, K.A.

    2000-01-01

    The US Department of Energy (DOE) Waste Isolation Pilot Plant (WIPP), located in southeast New Mexico, is a deep geologic repository for the permanent disposal of transuranic waste generated by DOE defense-related activities. Sandia National Laboratories (SNL), in its role as scientific advisor to the DOE, is responsible for evaluating the long-term performance of the WIPP. This risk-based Performance Assessment (PA) is accomplished in part through the use of numerous scientific modeling codes, which rely for some of their inputs on data gathered during characterization of the site. The PA is subject to formal requirements set forth in federal regulations. In particular, the components of the calculation fall under the configuration management and software quality assurance aegis of the American Society of Mechanical Engineers (ASME) Nuclear Quality Assurance (NQA) requirements. This paper describes SNL's implementation of the NQA requirements regarding software quality assurance (SQA). The description of the implementation of SQA for a PA calculation addresses not only the interpretation of the NQA requirements, it also discusses roles, deliverables, and the resources necessary for effective implementation. Finally, examples are given which illustrate the effectiveness of SNL's SQA program, followed by a detailed discussion of lessons learned

  2. Software quality assurance in the 1996 performance assessment for the Waste Isolation Pilot Plant

    International Nuclear Information System (INIS)

    Froehlich, Gary K.; Ogden, Harvey C.; Byle, Kathleen A.

    2000-01-01

    The US Department of Energy (DOE) Waste Isolation Pilot Plant (WIPP), located in southeast New Mexico, is a deep geologic repository for the permanent disposal of transuranic waste generated by DOE defense-related activities. Sandia National Laboratories (SNL), in its role as scientific advisor to the DOE, is responsible for evaluating the long-term performance of the WIPP. This risk-based Performance Assessment (PA) is accomplished in part through the use of numerous scientific modeling codes, which rely for some of their inputs on data gathered during characterization of the site. The PA is subject to formal requirements set forth in federal regulations. In particular, the components of the calculation fall under the configuration management and software quality assurance aegis of the American Society of Mechanical Engineers(ASME) Nuclear Quality Assurance (NQA) requirements. This paper describes SNL's implementation of the NQA requirements regarding software quality assurance (SQA). The description of the implementation of SQA for a PA calculation addresses not only the interpretation of the NQA requirements, it also discusses roles, deliverables, and the resources necessary for effective implementation. Finally, examples are given which illustrate the effectiveness of SNL's SQA program, followed by a detailed discussion of lessons learned

  3. ParseCNV integrative copy number variation association software with quality tracking.

    Science.gov (United States)

    Glessner, Joseph T; Li, Jin; Hakonarson, Hakon

    2013-03-01

    A number of copy number variation (CNV) calling algorithms exist; however, comprehensive software tools for CNV association studies are lacking. We describe ParseCNV, unique software that takes CNV calls and creates probe-based statistics for CNV occurrence in both case-control design and in family based studies addressing both de novo and inheritance events, which are then summarized based on CNV regions (CNVRs). CNVRs are defined in a dynamic manner to allow for a complex CNV overlap while maintaining precise association region. Using this approach, we avoid failure to converge and non-monotonic curve fitting weaknesses of programs, such as CNVtools and CNVassoc, and although Plink is easy to use, it only provides combined CNV state probe-based statistics, not state-specific CNVRs. Existing CNV association methods do not provide any quality tracking information to filter confident associations, a key issue which is fully addressed by ParseCNV. In addition, uncertainty in CNV calls underlying CNV associations is evaluated to verify significant results, including CNV overlap profiles, genomic context, number of probes supporting the CNV and single-probe intensities. When optimal quality control parameters are followed using ParseCNV, 90% of CNVs validate by polymerase chain reaction, an often problematic stage because of inadequate significant association review. ParseCNV is freely available at http://parsecnv.sourceforge.net.

  4. Modularity analysis of automotive control software

    OpenAIRE

    Dajsuren, Y.; Brand, van den, M.G.J.; Serebrenik, A.

    2013-01-01

    A design language and tool like MATLAB/Simulink is used for the graphical modelling and simulation of automotive control software. As the functionality based on electronics and software systems increases in motor vehicles, it is becoming increasingly important for system/software architects and control engineers in the automotive industry to ensure the quality of the highly complex MATLAB/Simulink control software. For automotive software, modularity is recognized as being a crucial quality a...

  5. Software to implement the IAEA-TECDOC 1517: quality control in mammography

    International Nuclear Information System (INIS)

    Mora Rodriguez, Patricia

    2011-01-01

    Mammography quality control is presented; also, guidelines for establishing a quality control program through tests to be performed, equipment, frequencies and tolerances. The work is intended for radiologists, medical physicists and technologists. Also examples of data collection sheets, minimum equipment list and overview of a quality control program are included. Two projects (RLA/9/057 and RLA/9/067) raised the goal of developing a software that allows a simple and effective way to automate the evidence contained in the TECDOC. The V1 is presented in the IRPA 12 (2008), which was tested in some countries to detect problems in their use and include possible improvements. However, the final revised version for distribution in member countries was get in 2010. The way of work on the project is described and who has worked. Also, a complete program, its importance and the projection of the Universidad de Costa Rica to work with the same. (author) [es

  6. How Does Software Process Improvement Address Global Software Engineering?

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Diebold, Philipp; Münch, Jürgen

    2016-01-01

    For decades, Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. To set up, guide, and carry out SPI projects, and to measure SPI state, impact, and success, a multitude of different SPI approaches and considerable...

  7. Software Process Improvement through the Removal of Project-Level Knowledge Flow Obstacles: The Perceptions of Software Engineers

    Science.gov (United States)

    Mitchell, Susan Marie

    2012-01-01

    Uncontrollable costs, schedule overruns, and poor end product quality continue to plague the software engineering field. Innovations formulated with the expectation to minimize or eliminate cost, schedule, and quality problems have generally fallen into one of three categories: programming paradigms, software tools, and software process…

  8. Prediction of cervical intraepithelial neoplasia grade 2+ (CIN2+ using HPV DNA testing after a diagnosis of atypical squamous cell of undetermined significance (ASC-US in Catalonia, Spain

    Directory of Open Access Journals (Sweden)

    Ibáñez Raquel

    2012-01-01

    Full Text Available Abstract Background A protocol for cervical cancer screening among sexually active women 25 to 65 years of age was introduced in 2006 in Catalonia, Spain to increase coverage and to recommend a 3-year-interval between screening cytology. In addition, Human Papillomavirus (HPV was offered as a triage test for women with a diagnosis of atypical squamous cells of undetermined significance (ASC-US. HPV testing was recommended within 3 months of ASC-US diagnosis. According to protocol, HPV negative women were referred to regular screening including a cytological exam every 3 years while HPV positive women were referred to colposcopy and closer follow-up. We evaluated the implementation of the protocol and the prediction of HPV testing as a triage tool for cervical intraepithelial lesions grade two or worse (CIN2+ in women with a cytological diagnosis of ASC-US. Methods During 2007-08 a total of 611 women from five reference laboratories in Catalonia with a novel diagnosis of ASC-US were referred for high risk HPV (hrHPV triage using high risk Hybrid Capture version 2. Using routine record linkage data, women were followed for 3 years to evaluate hrHPV testing efficacy for predicting CIN2+ cases. Logistic regression analysis was used to estimate the odds ratio for CIN2 +. Results Among the 611 women diagnosed with ASC-US, 493 (80.7% had at least one follow-up visit during the study period. hrHPV was detected in 48.3% of the women at study entry (mean age 35.2 years. hrHPV positivity decreased with increasing age from 72.6% among women younger than 25 years to 31.6% in women older than 54 years (p At the end of the 3 years follow-up period, 37 women with a diagnosis of CIN2+ (18 CIN2, 16 CIN3, 2 cancers, and 1 with high squamous intraepithelial lesions -HSIL were identified and all but one had a hrHPV positive test at study entry. Sensitivity to detect CIN2+ of hrHPV was 97.2% (95%confidence interval (CI = 85.5-99.9 and specificity was 68.3% (95%CI

  9. Software for safety critical applications

    International Nuclear Information System (INIS)

    Kropik, M.; Matejka, K.; Jurickova, M.; Chudy, R.

    2001-01-01

    The contribution gives an overview of the project of the software development for safety critical applications. This project has been carried out since 1997. The principal goal of the project was to establish a research laboratory for the development of the software with the highest requirements for quality and reliability. This laboratory was established at the department, equipped with proper hardware and software to support software development. A research team of predominantly young researchers for software development was created. The activities of the research team started with studying and proposing the software development methodology. In addition, this methodology was applied to the real software development. The verification and validation process followed the software development. The validation system for the integrated hardware and software tests was brought into being and its control software was developed. The quality of the software tools was also observed, and the SOSAT tool was used during these activities. National and international contacts were established and maintained during the project solution.(author)

  10. LDUA software custodian's notebook

    International Nuclear Information System (INIS)

    Aftanas, B.L.

    1998-01-01

    This plan describes the activities to be performed and controls to be applied to the process of specifying, obtaining, and qualifying the control and data acquisition software for the Light Duty Utility Arm (LDUA) System. It serves the purpose of a software quality assurance plan, a verification and validation plan, and a configuration management plan. This plan applies to all software that is an integral part of the LDUA control and data acquisition system, that is, software that is installed in the computers that are part of the LDUA system as it is deployed in the field. This plan applies to the entire development process, including: requirements; design; implementation; and operations and maintenance. This plan does not apply to any software that is not integral with the LDUA system. This plan has-been prepared in accordance with WHC-CM-6-1 Engineering Practices, EP-2.1; WHC-CM-3-10 Software Practices; and WHC-CM-4-2, QR 19.0, Software Quality Assurance Requirements

  11. Antocianinas, ácido ascórbico, polifenoles totales y actividad antioxidante, en la cáscara de camu-camu (Myrciaria dubia (H.B.K McVaugh Antocianinas, ácido ascórbico, polifenóis totais e atividade antioxidante na casca do camu-camu (Myrciaria dubia (H.B.K McVaugh

    Directory of Open Access Journals (Sweden)

    Juan Edson Villanueva-Tiburcio

    2010-05-01

    Full Text Available Este trabajo fue realizado en la UNAS, Tingo María, Perú. Los objetivos fueron evaluar el contenido de antocianinas, ácido ascórbico, y polifenoles totales, en la cáscara fresca y seca de camu-camu (Myrciaria dubia (H.B.K McVaugh en diferentes estados de madurez; evaluar la actividad antioxidante en la cáscara seca, usando diferentes tipos de radicales (DPPH, ABTS+ y Peroxilo y correlacionar el valor de ácido ascórbico y polifenoles totales con la actividad antioxidante. La extracción fue realizada en medio acuoso, y los resultados de las evaluaciones de cada experimento fueron analizados por un diseño completamente al azar (DCA, según la prueba de t-student (p A pesquisa foi realizada na UNAS em Tingo Maria, Peru, teve como objetivos: avaliar o teor de antocianinas, ácido ascórbico e polifenóis totais, na casca fresca e na casca seca do camu-camu (Myrciaria dubia (HBK McVaugh em diferentes tempos de maturação; avaliar a atividade antioxidante na casca seca utilizando diferentes tipos de radicais (DPPH, ABTS+ e Peroxilo e correlacionar o teor de ácido ascórbico e polifenóis totais com a atividade antioxidante. A extraç��o foi realizada em meio aquoso, os resultados das avaliações de cada experiência são analisados no delineamento inteiramente casualizado (DIC, pelo teste de t-student (p < 0,05. O extrato da casca da amostra madura fresca apresentou as concentrações mais elevadas de ácido ascórbico e de antocianinas em relação a meio madura e verde, com 21,95 mg.g-1 de casca e 46,42 mg.L-1 de cianidin-3-glucosídeo, respectivamente, enquanto que o extrato da casca seca meio madura apresentou o maior teor de ácido ascórbico em relação a madura e verde (53,49 mg.g-1 e de polifenóis totais: 7,70 mg Ác. Gálico/g. A maior atividade antioxidante foi encontrada no extrato da casca seca da amostra meio madura com IC50 = 46,20; 20,25 e 8,30 μg.mL-1, em comparação com DPPH radical, ABTS+ e Perox

  12. Tool Use Within NASA Software Quality Assurance

    Science.gov (United States)

    Shigeta, Denise; Port, Dan; Nikora, Allen P.; Wilf, Joel

    2013-01-01

    As space mission software systems become larger and more complex, it is increasingly important for the software assurance effort to have the ability to effectively assess both the artifacts produced during software system development and the development process itself. Conceptually, assurance is a straightforward idea - it is the result of activities carried out by an organization independent of the software developers to better inform project management of potential technical and programmatic risks, and thus increase management's confidence in the decisions they ultimately make. In practice, effective assurance for large, complex systems often entails assessing large, complex software artifacts (e.g., requirements specifications, architectural descriptions) as well as substantial amounts of unstructured information (e.g., anomaly reports resulting from testing activities during development). In such an environment, assurance engineers can benefit greatly from appropriate tool support. In order to do so, an assurance organization will need accurate and timely information on the tool support available for various types of assurance activities. In this paper, we investigate the current use of tool support for assurance organizations within NASA, and describe on-going work at JPL for providing assurance organizations with the information about tools they need to use them effectively.

  13. Design and implementation of software for automated quality control and data analysis for a complex LC/MS/MS assay for urine opiates and metabolites.

    Science.gov (United States)

    Dickerson, Jane A; Schmeling, Michael; Hoofnagle, Andrew N; Hoffman, Noah G

    2013-01-16

    Mass spectrometry provides a powerful platform for performing quantitative, multiplexed assays in the clinical laboratory, but at the cost of increased complexity of analysis and quality assurance calculations compared to other methodologies. Here we describe the design and implementation of a software application that performs quality control calculations for a complex, multiplexed, mass spectrometric analysis of opioids and opioid metabolites. The development and implementation of this application improved our data analysis and quality assurance processes in several ways. First, use of the software significantly improved the procedural consistency for performing quality control calculations. Second, it reduced the amount of time technologists spent preparing and reviewing the data, saving on average over four hours per run, and in some cases improving turnaround time by a day. Third, it provides a mechanism for coupling procedural and software changes with the results of each analysis. We describe several key details of the implementation including the use of version control software and automated unit tests. These generally useful software engineering principles should be considered for any software development project in the clinical lab. Copyright © 2012 Elsevier B.V. All rights reserved.

  14. Quality Assurance Testing of Version 1.3 of U.S. EPA Benchmark Dose Software (Presentation)

    Science.gov (United States)

    EPA benchmark dose software (BMDS) issued to evaluate chemical dose-response data in support of Agency risk assessments, and must therefore be dependable. Quality assurance testing methods developed for BMDS were designed to assess model dependability with respect to curve-fitt...

  15. The National Water-Quality Assessment Program Invertebrate Data Analysis System (IDAS) Software: Version 3 (User's Manual)

    National Research Council Canada - National Science Library

    Cuffney, Thomas F

    2003-01-01

    ... as part of the National Water-Quality Assessment Program and stored in the Biological Transactional Database (Bio-TDB). The IDAS software is a stand-alone program for personal computers that run Microsoft...

  16. Impact on dose and image quality of a software-based scatter correction in mammography.

    Science.gov (United States)

    Monserrat, Teresa; Prieto, Elena; Barbés, Benigno; Pina, Luis; Elizalde, Arlette; Fernández, Belén

    2017-01-01

    Background In 2014, Siemens developed a new software-based scatter correction (Progressive Reconstruction Intelligently Minimizing Exposure [PRIME]), enabling grid-less digital mammography. Purpose To compare doses and image quality between PRIME (grid-less) and standard (with anti-scatter grid) modes. Material and Methods Contrast-to-noise ratio (CNR) was measured for various polymethylmethacrylate (PMMA) thicknesses and dose values provided by the mammograph were recorded. CDMAM phantom images were acquired for various PMMA thicknesses and inverse Image Quality Figure (IQF inv ) was calculated. Values of incident entrance surface air kerma (ESAK) and average glandular dose (AGD) were obtained from the DICOM header for a total of 1088 pairs of clinical cases. Two experienced radiologists compared subjectively the image quality of a total of 149 pairs of clinical cases. Results CNR values were higher and doses were lower in PRIME mode for all thicknesses. IQF inv values in PRIME mode were lower for all thicknesses except for 40 mm of PMMA equivalent, in which IQF inv was slightly greater in PRIME mode. A mean reduction of 10% in ESAK and 12% in AGD in PRIME mode with respect to standard mode was obtained. The clinical image quality in PRIME and standard acquisitions resulted to be similar in most of the cases (84% for the first radiologist and 67% for the second one). Conclusion The use of PRIME software reduces, in average, the dose of radiation to the breast without affecting image quality. This reduction is greater for thinner and denser breasts.

  17. Determinação das constantes cinéticas de degradação do ácido ascórbico em purê de pêssego: efeito da temperatura e concentração

    OpenAIRE

    Toralles,Ricardo Peraça; Vendruscolo,João Luiz; Vendruscolo,Claire Tondo; Del Pino,Francisco Augusto Burkert; Antunes,Pedro Luiz

    2008-01-01

    O ácido ascórbico, vitamina C, é usado extensivamente na indústria de alimentos, não só devido ao seu valor nutricional, mas devido a suas contribuições funcionais na qualidade do produto. Existem muitos estudos sobre a estabilidade cinética do ácido ascórbico em bebidas, mas nenhum estudo foi encontrado sobre as constantes cinéticas de degradação do ácido ascórbico adicionado em purê de pêssego. Neste trabalho, estudou-se a cinética de degradação do ácido ascórbico em purê de pêssego da cult...

  18. Characteristics for Software Optimization Projects

    Directory of Open Access Journals (Sweden)

    Iulian NITESCU

    2008-01-01

    Full Text Available The increasing of the software systems complexity imposes the identification and implementation of some methods and techniques in order to manage it. The software optimization project is a way in which the software complexity is controlled. The software optimization project must face to the organization need to earn profit. The software optimization project is an integrated part of the application cycle because share same resources, depends on other stages and influences next phases. The optimization project has some particularities because it works on an finished product around its quality. The process is quality and performance oriented and it assumes that the product life cycle is almost finished.

  19. STATIC CODE ANALYSIS FOR SOFTWARE QUALITY IMPROVEMENT: A CASE STUDY IN BCI FRAMEWORK DEVELOPMENT

    Directory of Open Access Journals (Sweden)

    Indar Sugiarto

    2008-01-01

    Full Text Available This paper shows how the systematic approach in software testing using static code analysis method can be used for improving the software quality of a BCI framework. The method is best performed during the development phase of framework programs. In the proposed approach, we evaluate several software metrics which are based on the principles of object oriented design. Since such method is depending on the underlying programming language, we describe the method in term of C++ language programming whereas the Qt platform is also currently being used. One of the most important metric is so called software complexity. Applying the software complexity calculation using both McCabe and Halstead method for the BCI framework which consists of two important types of BCI, those are SSVEP and P300, we found that there are two classes in the framework which have very complex and prone to violation of cohesion principle in OOP. The other metrics are fit the criteria of the proposed framework aspects, such as: MPC is less than 20; average complexity is around value of 5; and the maximum depth is below 10 blocks. Such variables are considered very important when further developing the BCI framework in the future.

  20. Quality Model Based on Cots Quality Attributes

    OpenAIRE

    Jawad Alkhateeb; Khaled Musa

    2013-01-01

    The quality of software is essential to corporations in making their commercial software. Good or poorquality to software plays an important role to some systems such as embedded systems, real-time systems,and control systems that play an important aspect in human life. Software products or commercial off theshelf software are usually programmed based on a software quality model. In the software engineeringfield, each quality model contains a set of attributes or characteristics that drives i...

  1. A Framework of the Use of Information in Software Testing

    Science.gov (United States)

    Kaveh, Payman

    2010-01-01

    With the increasing role that software systems play in our daily lives, software quality has become extremely important. Software quality is impacted by the efficiency of the software testing process. There are a growing number of software testing methodologies, models, and initiatives to satisfy the need to improve software quality. The main…

  2. Efeito do ácido ascórbico em massa de pão na presença do ácido tânico - doi: 10.4025/actascitechnol.v32i2.5290

    Directory of Open Access Journals (Sweden)

    Ana Leticia Gomes Saraiva

    2010-07-01

    Full Text Available Pães de farinha de trigo com adição do agente oxidante ácido ascórbico foram elaborados, tentando-se encontrar a concentração de ácido tânico que melhor retivesse o ácido ascórbico, sem afetar significativamente as características reológicas e os principais atributos sensoriais do pão. Os pães foram elaborados com a formulação de pão-padrão. Na primeira etapa, determinou-se a concentração de ácido ascórbico que melhor proporcionasse estabilidade à massa de pão-padrão (300 ppm. Na etapa seguinte, foram utilizadas na formulação da massa-padrão três concentrações diferentes de ácido tânico: 0,10; 0,20 e 0,30%, com a finalidade de se verificar a concentração que melhor retivesse o ácido ascórbico. Os pães foram elaborados com a formulação-padrão. Após o resfriamento dos pães, foi avaliada a retenção de ácido ascórbico, presente no pão de acordo com a concentração previamente determinada. Os resultados indicaram que o pão elaborado com a massa de pão-padrão (300 ppm de ácido ascórbico apresentou concentração de ácido ascórbico de 41,50 mg 100 g-1 de pão.

  3. Observaciones histopatologicas de juveniles penaeus vannamei sometidos a dietas artificiales con diferentes concentraciones de una sal de ácido ascórbico (vitamina c)

    OpenAIRE

    Vera Muñoz, L.

    1995-01-01

    Observaciones histopatologicas de juveniles Penaeus Vannamei sometidos a dietas artificiales con diferentes concentraciones de una sal de ácido ascórbico (vitamina C) Se realizaron análisis histológicos y comprobación mediante métodos histoquímicos de la presencia de melanina, a juveniles Penaeus vannamei sometidos a cinco dietas experimentales con diferentes concentraciones de una sal de L-Ascorbato-2-Fosfato de Mg (APM) usada como fuente de ácido ascórbico (AA).

  4. Clinical significance of HPV DNA cotesting in Korean women with ASCUS or ASC-H.

    Science.gov (United States)

    Lee, Sanghoon; Kim, Jae Won; Hong, Jin Hwa; Song, Jae Yun; Lee, Jae Kwan; Kim, In Sun; Lee, Nak Woo

    2014-12-01

    The purpose of this study was to evaluate the clinical significance of Human papillomavirus (HPV) DNA cotesting in Korean women with abnormal Papanicolaou (Pap) smear results based on colposcopic pathology. A total of 1012 women underwent liquid-based Pap smears and hybrid capture II HPV DNA tests followed by colposcopy at the Korea University Hospital from January 2007 to May 2012. Of these women, 832 women were included in this retrospective study. The mean patient age was 45.4 ± 13.7 years (range:15-80). The distribution of Pap smear results was normal (4.7%), atypical squamous cells of uncertain significance (ASCUS) (42.1%), low-grade squamous intraepithelial lesion (26.8%), ASC-H (7.0%), and high-grade squamous intraepithelial lesion (HSIL) (19.5%). In women with ASCUS, none of the 87 HPV-negative had ≥cervical intraepithelial neoplasia (CIN2) (P age groups: ASCUS and ASC-H furnish healthcare providers with informative data. There is a lower proportion of ≥CIN2 in HPV-negative women and a higher proportion of ≥CIN2 in HPV-positive. When HPV data were further evaluated by age group, the risk of ≥CIN2 was lower in HPV-negative women, especially in women ≥30. © 2014 Wiley Periodicals, Inc.

  5. Effects of Metal Micro and Nano-Particles on hASCs: An In Vitro Model.

    Science.gov (United States)

    Palombella, Silvia; Pirrone, Cristina; Rossi, Federica; Armenia, Ilaria; Cherubino, Mario; Valdatta, Luigi; Raspanti, Mario; Bernardini, Giovanni; Gornati, Rosalba

    2017-08-03

    As the knowledge about the interferences of nanomaterials on human staminal cells are scarce and contradictory, we undertook a comparative multidisciplinary study based on the size effect of zero-valent iron, cobalt, and nickel microparticles (MPs) and nanoparticles (NPs) using human adipose stem cells (hASCs) as a model, and evaluating cytotoxicity, morphology, cellular uptake, and gene expression. Our results suggested that the medium did not influence the cell sensitivity but, surprisingly, the iron microparticles (FeMPs) resulted in being toxic. These data were supported by modifications in mRNA expression of some genes implicated in the inflammatory response. Microscopic analysis confirmed that NPs, mainly internalized by endocytosis, persist in the vesicles without any apparent cell damage. Conversely, MPs are not internalized, and the effects on hASCs have to be ascribed to the release of ions in the culture medium, or to the reduced oxygen and nutrient exchange efficiency due to the presence of MP agglomerating around the cells. Notwithstanding the results depicting a heterogeneous scene that does not allow drawing a general conclusion, this work reiterates the importance of comparative investigations on MPs, NPs, and corresponding ions, and the need to continue the thorough verification of NP and MP innocuousness to ensure unaffected stem cell physiology and differentiation.

  6. Reliability of adaptive multivariate software sensors for sewer water quality monitoring

    DEFF Research Database (Denmark)

    Vezzaro, Luca; Mikkelsen, Peter Steen

    2015-01-01

    This study investigates the use of a multivariate approach, based on Principal Component Analysis PCA), as software sensor for fault detection and reconstruction of missing measurements in on-line monitoring of sewer water quality. The analysis was carried out on a 16-months dataset of five...... commonly available on-line measurements (flow, turbidity, ammonia, conductivity and temperature). The results confirmed the great performance of PCA (up to 10 weeks after parameter estimation) when estimating a measurement from the combination of the remaining four variables, a useful feature in data...

  7. Water Quality Research Program: Development of Unstructured Grid Linkage Methodology and Software for CE-QUAL-ICM

    National Research Council Canada - National Science Library

    Chapman, Raymond

    1997-01-01

    This study was conducted for the purpose of developing a methodology and associated software for linking hydrodynamic output from the RMAlO finite element model to the CE-QUAL-ICM finite volume water quality model...

  8. Increasing the impact of usability work in software development

    DEFF Research Database (Denmark)

    Uldall-Espersen, Tobias; Frøkjær, Erik

    2006-01-01

    Usability, Case Study, Software Engineering, Software Quality, Organizational Impact, Usability Requirement Management, CHI 2007 workshop......Usability, Case Study, Software Engineering, Software Quality, Organizational Impact, Usability Requirement Management, CHI 2007 workshop...

  9. HPCMP CREATE (trademark)-AV Quality Assurance: Best Practices for Validating and Supporting Computation-Based Engineering Software

    Science.gov (United States)

    2015-09-30

    30/2015 Oct 2008-Sep 2015 HPCMP CREATE™- AV Quality Assurance: Best Practices for Validating and Supporting Computation-Based Engineering Software...2) “Does this tool adequately perform any and all advertised capabilities?” This paper will describe how the HPCMP CREATE Air Vehicles ( AV ...discussed and their strengths and weaknesses within the CREATE- AV framework addressed. Work toward the HPCMP CREATE, Quality Assurance, Aviation

  10. 76 FR 54800 - International Business Machines (IBM), Software Group Business Unit, Quality Assurance Group, San...

    Science.gov (United States)

    2011-09-02

    ... DEPARTMENT OF LABOR Employment and Training Administration [TA-W-74,554] International Business Machines (IBM), Software Group Business Unit, Quality Assurance Group, San Jose, California; Notice of Negative Determination on Reconsideration On January 21, 2011, the Department of Labor (Department) issued an Affirmative Determination Regarding...

  11. Development of Automatic Quality Check Software in Mailbox Declaration For Nuclear Fuel Fabrication Plants

    International Nuclear Information System (INIS)

    Kim, Minsu; Shim, Hye Won; Jo, Seong Yeon; Lee, Kwang Yeol; Ban, Myoung Jin

    2014-01-01

    Short Notice Random Inspection (SNRI) is a new IAEA safeguards inspection regime for bulk handing facility, which utilities random inspection through a mailbox system. Its main objective is to verify 100% of the flow components of the safeguarded nuclear material at such a facility. To achieve the SNRI objective, it is required to provide daily mailbox declaration, by a facility's operator, to the IAEA with regard to information, such as the receipt and shipment of nuclear materials. Mailbox declarations are then later compared with accounting records so as to examine the accuracy and consistency of the facility operator's declaration at the time of the SNRI. The IAEA has emphasized the importance of accurate mailbox declarations and recommended that the ROK initiate its own independent quality control system in order to improve and maintain its mailbox declarations as a part of the SSAC activities. In an effort to improve the transparency of operational activities at fuel fabrication plants and to satisfy IAEA recommendation, an automatic quality check software application has been developed to improve mailbox declarations at fabrication plants in Korea. The ROK and the IAEA have recognized the importance of providing good quality mailbox declaration for an effective and efficient SNRI at fuel fabrication plants in Korea. The SRA developed an automatic quality check software program in order to provide an independent QC system of mailbox declaration, as well as to improve the quality of mailbox declaration. Once the automatic QC system is implemented, it will improve the quality of an operator's mailbox declaration by examining data before sending it to the IAEA. The QC system will be applied to fuel fabrication plants in the first half of 2014

  12. Towards an Evaluation Framework for Software Process Improvement

    OpenAIRE

    Cheng, Chow Kian; Permadi, Rahadian Bayu

    2009-01-01

    Software has gained an essential role in our daily life in the last decades. This condition demands high quality software. To produce high quality software many practitioners and researchers put more attention on the software development process. Large investments are poured to improve the software development process. Software Process Improvement (SPI) is a research area which is aimed to address the assessment and improvement issues in the software development process. One of the most impor...

  13. Effects of Metal Micro and Nano-Particles on hASCs: An In Vitro Model

    OpenAIRE

    Palombella, Silvia; Pirrone, Cristina; Rossi, Federica; Armenia, Ilaria; Cherubino, Mario; Valdatta, Luigi; Raspanti, Mario; Bernardini, Giovanni; Gornati, Rosalba

    2017-01-01

    As the knowledge about the interferences of nanomaterials on human staminal cells are scarce and contradictory, we undertook a comparative multidisciplinary study based on the size effect of zero-valent iron, cobalt, and nickel microparticles (MPs) and nanoparticles (NPs) using human adipose stem cells (hASCs) as a model, and evaluating cytotoxicity, morphology, cellular uptake, and gene expression. Our results suggested that the medium did not influence the cell sensitivity but, surprisingly...

  14. Verification of safety critical software

    International Nuclear Information System (INIS)

    Son, Ki Chang; Chun, Chong Son; Lee, Byeong Joo; Lee, Soon Sung; Lee, Byung Chai

    1996-01-01

    To assure quality of safety critical software, software should be developed in accordance with software development procedures and rigorous software verification and validation should be performed. Software verification is the formal act of reviewing, testing of checking, and documenting whether software components comply with the specified requirements for a particular stage of the development phase[1]. New software verification methodology was developed and was applied to the Shutdown System No. 1 and 2 (SDS1,2) for Wolsung 2,3 and 4 nuclear power plants by Korea Atomic Energy Research Institute(KAERI) and Atomic Energy of Canada Limited(AECL) in order to satisfy new regulation requirements of Atomic Energy Control Boars(AECB). Software verification methodology applied to SDS1 for Wolsung 2,3 and 4 project will be described in this paper. Some errors were found by this methodology during the software development for SDS1 and were corrected by software designer. Outputs from Wolsung 2,3 and 4 project have demonstrated that the use of this methodology results in a high quality, cost-effective product. 15 refs., 6 figs. (author)

  15. Waste Management facilities cost information: System Cost Model Software Quality Assurance Plan. Revision 2

    International Nuclear Information System (INIS)

    Peterson, B.L.; Lundeen, A.S.

    1996-02-01

    In May of 1994, Lockheed Idaho Technologies Company (LITCO) in Idaho Falls, Idaho and subcontractors developed the System Cost Model (SCM) application. The SCM estimates life-cycle costs of the entire US Department of Energy (DOE) complex for designing; constructing; operating; and decommissioning treatment, storage, and disposal (TSD) facilities for mixed low-level, low-level, transuranic, and mixed transuranic waste. The SCM uses parametric cost functions to estimate life-cycle costs for various treatment, storage, and disposal modules which reflect planned and existing facilities at DOE installations. In addition, SCM can model new facilities based on capacity needs over the program life cycle. The SCM also provides transportation costs for truck and rail, which include transport of contact-handled, remote-handled, and alpha (transuranic) wastes. The user can provide input data (default data is included in the SCM) including the volume and nature of waste to be managed, the time period over which the waste is to be managed, and the configuration of the waste management complex (i.e., where each installation's generated waste will be treated, stored, and disposed). Then the SCM uses parametric cost equations to estimate the costs of pre-operations (designing), construction costs, operation management, and decommissioning these waste management facilities. For the product to be effective and useful the SCM users must have a high level of confidence in the data generated by the software model. The SCM Software Quality Assurance Plan is part of the overall SCM project management effort to ensure that the SCM is maintained as a quality product and can be relied on to produce viable planning data. This document defines tasks and deliverables to ensure continued product integrity, provide increased confidence in the accuracy of the data generated, and meet the LITCO's quality standards during the software maintenance phase. 8 refs., 1 tab

  16. Waste Management facilities cost information: System Cost Model Software Quality Assurance Plan. Revision 2

    Energy Technology Data Exchange (ETDEWEB)

    Peterson, B.L.; Lundeen, A.S.

    1996-02-01

    In May of 1994, Lockheed Idaho Technologies Company (LITCO) in Idaho Falls, Idaho and subcontractors developed the System Cost Model (SCM) application. The SCM estimates life-cycle costs of the entire US Department of Energy (DOE) complex for designing; constructing; operating; and decommissioning treatment, storage, and disposal (TSD) facilities for mixed low-level, low-level, transuranic, and mixed transuranic waste. The SCM uses parametric cost functions to estimate life-cycle costs for various treatment, storage, and disposal modules which reflect planned and existing facilities at DOE installations. In addition, SCM can model new facilities based on capacity needs over the program life cycle. The SCM also provides transportation costs for truck and rail, which include transport of contact-handled, remote-handled, and alpha (transuranic) wastes. The user can provide input data (default data is included in the SCM) including the volume and nature of waste to be managed, the time period over which the waste is to be managed, and the configuration of the waste management complex (i.e., where each installation`s generated waste will be treated, stored, and disposed). Then the SCM uses parametric cost equations to estimate the costs of pre-operations (designing), construction costs, operation management, and decommissioning these waste management facilities. For the product to be effective and useful the SCM users must have a high level of confidence in the data generated by the software model. The SCM Software Quality Assurance Plan is part of the overall SCM project management effort to ensure that the SCM is maintained as a quality product and can be relied on to produce viable planning data. This document defines tasks and deliverables to ensure continued product integrity, provide increased confidence in the accuracy of the data generated, and meet the LITCO`s quality standards during the software maintenance phase. 8 refs., 1 tab.

  17. Software architecture 2

    CERN Document Server

    Oussalah, Mourad Chabanne

    2014-01-01

    Over the past 20 years, software architectures have significantly contributed to the development of complex and distributed systems. Nowadays, it is recognized that one of the critical problems in the design and development of any complex software system is its architecture, i.e. the organization of its architectural elements. Software Architecture presents the software architecture paradigms based on objects, components, services and models, as well as the various architectural techniques and methods, the analysis of architectural qualities, models of representation of architectural templa

  18. Software architecture 1

    CERN Document Server

    Oussalah , Mourad Chabane

    2014-01-01

    Over the past 20 years, software architectures have significantly contributed to the development of complex and distributed systems. Nowadays, it is recognized that one of the critical problems in the design and development of any complex software system is its architecture, i.e. the organization of its architectural elements. Software Architecture presents the software architecture paradigms based on objects, components, services and models, as well as the various architectural techniques and methods, the analysis of architectural qualities, models of representation of architectural template

  19. Qualifying codes under software quality assurance: Two examples as guidelines for codes that are existing or under development

    Energy Technology Data Exchange (ETDEWEB)

    Mangold, D.

    1993-05-01

    Software quality assurance is an area of concem for DOE, EPA, and other agencies due to the poor quality of software and its documentation they have received in the past. This report briefly summarizes the software development concepts and terminology increasingly employed by these agencies and provides a workable approach to scientific programming under the new requirements. Following this is a practical description of how to qualify a simulation code, based on a software QA plan that has been reviewed and officially accepted by DOE/OCRWM. Two codes have recently been baselined and qualified, so that they can be officially used for QA Level 1 work under the DOE/OCRWM QA requirements. One of them was baselined and qualified within one week. The first of the codes was the multi-phase multi-component flow code TOUGH version 1, an already existing code, and the other was a geochemistry transport code STATEQ that was under development The way to accomplish qualification for both types of codes is summarized in an easy-to-follow step-by step fashion to illustrate how to baseline and qualify such codes through a relatively painless procedure.

  20. Qualifying codes under software quality assurance: Two examples as guidelines for codes that are existing or under development

    International Nuclear Information System (INIS)

    Mangold, D.

    1993-05-01

    Software quality assurance is an area of concern for DOE, EPA, and other agencies due to the poor quality of software and its documentation they have received in the past. This report briefly summarizes the software development concepts and terminology increasingly employed by these agencies and provides a workable approach to scientific programming under the new requirements. Following this is a practical description of how to qualify a simulation code, based on a software QA plan that has been reviewed and officially accepted by DOE/OCRWM. Two codes have recently been baselined and qualified, so that they can be officially used for QA Level 1 work under the DOE/OCRWM QA requirements. One of them was baselined and qualified within one week. The first of the codes was the multi-phase multi-component flow code TOUGH version 1, an already existing code, and the other was a geochemistry transport code STATEQ that was under development The way to accomplish qualification for both types of codes is summarized in an easy-to-follow step-by step fashion to illustrate how to baseline and qualify such codes through a relatively painless procedure

  1. Effects of hot boning and moisture enhancement on the eating quality of cull cow beef.

    Science.gov (United States)

    Pivotto, L M; Campbell, C P; Swanson, K; Mandell, I B

    2014-01-01

    The effects of chilling method and moisture enhancement were examined for improving eating quality of semimembranosus (SM) and longissimus lumborum (LL) from 62 cull beef cows. Chilling method included hot boning muscles after 45 to 60 min postmortem or conventional chilling for 24 h. Moisture enhancement included 1) a non-injected control (CONT) or injection processing (10% of product weight) using 2) Sodium Tripolyphosphate/salt (Na/STP), 3) Sodium Citrate (NaCIT), 4) Calcium Ascorbate (CaASC), or 5) Citrus Juices (CITRUS). Chilling method by moisture enhancement treatment interactions (Pboned vs. conventionally chilled product (SM and LL) for CaASC vs. other moisture enhancement treatments. Chilling method by moisture enhancement treatment interactions (Pboned LL using CaASC vs. Na/STP. Moisture enhancement can improve tenderness of cull cow beef depending on combinations of chilling method and moisture enhancement treatments used. © 2013.

  2. Visualization on supercomputing platform level II ASC milestone (3537-1B) results from Sandia.

    Energy Technology Data Exchange (ETDEWEB)

    Geveci, Berk (Kitware, Inc., Clifton Park, NY); Fabian, Nathan; Marion, Patrick (Kitware, Inc., Clifton Park, NY); Moreland, Kenneth D.

    2010-09-01

    This report provides documentation for the completion of the Sandia portion of the ASC Level II Visualization on the platform milestone. This ASC Level II milestone is a joint milestone between Sandia National Laboratories and Los Alamos National Laboratories. This milestone contains functionality required for performing visualization directly on a supercomputing platform, which is necessary for peta-scale visualization. Sandia's contribution concerns in-situ visualization, running a visualization in tandem with a solver. Visualization and analysis of petascale data is limited by several factors which must be addressed as ACES delivers the Cielo platform. Two primary difficulties are: (1) Performance of interactive rendering, which is most computationally intensive portion of the visualization process. For terascale platforms, commodity clusters with graphics processors(GPUs) have been used for interactive rendering. For petascale platforms, visualization and rendering may be able to run efficiently on the supercomputer platform itself. (2) I/O bandwidth, which limits how much information can be written to disk. If we simply analyze the sparse information that is saved to disk we miss the opportunity to analyze the rich information produced every timestep by the simulation. For the first issue, we are pursuing in-situ analysis, in which simulations are coupled directly with analysis libraries at runtime. This milestone will evaluate the visualization and rendering performance of current and next generation supercomputers in contrast to GPU-based visualization clusters, and evaluate the performance of common analysis libraries coupled with the simulation that analyze and write data to disk during a running simulation. This milestone will explore, evaluate and advance the maturity level of these technologies and their applicability to problems of interest to the ASC program. Scientific simulation on parallel supercomputers is traditionally performed in four

  3. Quality assurance applied to mammographic equipments using phantoms and software for its evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Mayo, Patricia, E-mail: p.mayo@titaniast.co [Titania Servicios Tecnologicos S.L., Grupo Dominguis, Apartado 46015, Valencia (Spain); Rodenas, Francisco [Departamento de Matematica Aplicada, Universidad Politecnica de Valencia, Apartado 46022, Valencia (Spain); Manuel Campayo, Juan [Hospital Clinico Universitario de Valencia, Avda. Blasco Ibanez, Apartado 46017, Valencia (Spain); Verdu, Gumersido [Departamento de Ingenieria Quimica y Nuclear, Universidad Politecnica de Valencia, Apartado 46022, Valencia (Spain)

    2010-07-21

    The image quality assessment in radiographic equipments is a very important item for a complete quality control of the radiographic image chain. The periodic evaluation of the radiographic image quality must guarantee the constancy of this quality to carry out a suitable diagnosis. Mammographic phantom images are usually used to study the quality of images obtained by determined mammographic equipment. The digital image treatment techniques allow to carry out an automatic analysis of the phantom image. In this work we apply some techniques of digital image processing to analyze in an automatic way the image quality of mammographic phantoms, namely CIRS SP01 and RACON for different varying conditions of the mammographic equipment. The CIRS SP01 phantom is usually used in analogic mammographic equipments and the RACON phantom has been specifically developed by authors to be applied to acceptance and constancy tests of the image quality in digital radiographic equipments following recommendations of international associations. The purpose of this work consists in analyzing the image quality for both phantoms by means of an automatic software utility. This analysis allows us to study the functioning of the image chain of the mammographic system in an objective way, so an abnormal functioning of the radiographic equipment might be detected.

  4. Quality assurance applied to mammographic equipments using phantoms and software for its evaluation

    International Nuclear Information System (INIS)

    Mayo, Patricia; Rodenas, Francisco; Manuel Campayo, Juan; Verdu, Gumersido

    2010-01-01

    The image quality assessment in radiographic equipments is a very important item for a complete quality control of the radiographic image chain. The periodic evaluation of the radiographic image quality must guarantee the constancy of this quality to carry out a suitable diagnosis. Mammographic phantom images are usually used to study the quality of images obtained by determined mammographic equipment. The digital image treatment techniques allow to carry out an automatic analysis of the phantom image. In this work we apply some techniques of digital image processing to analyze in an automatic way the image quality of mammographic phantoms, namely CIRS SP01 and RACON for different varying conditions of the mammographic equipment. The CIRS SP01 phantom is usually used in analogic mammographic equipments and the RACON phantom has been specifically developed by authors to be applied to acceptance and constancy tests of the image quality in digital radiographic equipments following recommendations of international associations. The purpose of this work consists in analyzing the image quality for both phantoms by means of an automatic software utility. This analysis allows us to study the functioning of the image chain of the mammographic system in an objective way, so an abnormal functioning of the radiographic equipment might be detected.

  5. Comparison of lysimeter based and calculated ASCE reference evapotranspiration in a subhumid climate

    Science.gov (United States)

    Nolz, Reinhard; Cepuder, Peter; Eitzinger, Josef

    2016-04-01

    The standardized form of the well-known FAO Penman-Monteith equation, published by the Environmental and Water Resources Institute of the American Society of Civil Engineers (ASCE-EWRI), is recommended as a standard procedure for calculating reference evapotranspiration (ET ref) and subsequently plant water requirements. Applied and validated under different climatic conditions it generally achieved good results compared to other methods. However, several studies documented deviations between measured and calculated reference evapotranspiration depending on environmental and weather conditions. Therefore, it seems generally advisable to evaluate the model under local environmental conditions. In this study, reference evapotranspiration was determined at a subhumid site in northeastern Austria from 2005 to 2010 using a large weighing lysimeter (ET lys). The measured data were compared with ET ref calculations. Daily values differed slightly during a year, at which ET ref was generally overestimated at small values, whereas it was rather underestimated when ET was large, which is supported also by other studies. In our case, advection of sensible heat proved to have an impact, but it could not explain the differences exclusively. Obviously, there were also other influences, such as seasonal varying surface resistance or albedo. Generally, the ASCE-EWRI equation for daily time steps performed best at average weather conditions. The outcomes should help to correctly interpret ET ref data in the region and in similar environments and improve knowledge on the dynamics of influencing factors causing deviations.

  6. Software to model AXAF-I image quality

    Science.gov (United States)

    Ahmad, Anees; Feng, Chen

    1995-01-01

    A modular user-friendly computer program for the modeling of grazing-incidence type x-ray optical systems has been developed. This comprehensive computer software GRAZTRACE covers the manipulation of input data, ray tracing with reflectivity and surface deformation effects, convolution with x-ray source shape, and x-ray scattering. The program also includes the capabilities for image analysis, detector scan modeling, and graphical presentation of the results. A number of utilities have been developed to interface the predicted Advanced X-ray Astrophysics Facility-Imaging (AXAF-I) mirror structural and thermal distortions with the ray-trace. This software is written in FORTRAN 77 and runs on a SUN/SPARC station. An interactive command mode version and a batch mode version of the software have been developed.

  7. Software Testing Requires Variability

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak

    2003-01-01

    Software variability is the ability of a software system or artefact to be changed, customized or configured for use in a particular context. Variability in software systems is important from a number of perspectives. Some perspectives rightly receive much attention due to their direct economic...... impact in software production. As is also apparent from the call for papers these perspectives focus on qualities such as reuse, adaptability, and maintainability....

  8. Software quality assurance procedures for radioactive waste risk assessment codes

    International Nuclear Information System (INIS)

    Hill, I.; Mayer, J.

    1990-01-01

    This support study for the evaluation of the safety of geological disposal systems is aimed at identifying the requirements for software quality assurance procedures for radioactive waste risk assessment codes, and to recommend appropriate procedures. The research covers: (i) the analysis of existing procedures and definition of requirements; (ii) a case study of the use of some existing procedures; (iii) the definition and the implementation of procedures. The report is supported by appendices that give more detail on the procedures recommended. It is intended to provide ideas on the steps that should be taken to ensure the quality of the programs used for assessment of the safety case for radioactive waste repositories, and does not represent the introduction of wholly new ideas or techniques. The emphasis throughout is on procedures that will be easily implemented, rather than on the fully rigorous procedures that are required for some application areas. The study has concentrated on measures that will increase the confidence in repository performance assessments among the wider scientific/engineering community, and the lay public

  9. GCE Data Toolbox for MATLAB - a software framework for automating environmental data processing, quality control and documentation

    Science.gov (United States)

    Sheldon, W.; Chamblee, J.; Cary, R. H.

    2013-12-01

    Environmental scientists are under increasing pressure from funding agencies and journal publishers to release quality-controlled data in a timely manner, as well as to produce comprehensive metadata for submitting data to long-term archives (e.g. DataONE, Dryad and BCO-DMO). At the same time, the volume of digital data that researchers collect and manage is increasing rapidly due to advances in high frequency electronic data collection from flux towers, instrumented moorings and sensor networks. However, few pre-built software tools are available to meet these data management needs, and those tools that do exist typically focus on part of the data management lifecycle or one class of data. The GCE Data Toolbox has proven to be both a generalized and effective software solution for environmental data management in the Long Term Ecological Research Network (LTER). This open source MATLAB software library, developed by the Georgia Coastal Ecosystems LTER program, integrates metadata capture, creation and management with data processing, quality control and analysis to support the entire data lifecycle. Raw data can be imported directly from common data logger formats (e.g. SeaBird, Campbell Scientific, YSI, Hobo), as well as delimited text files, MATLAB files and relational database queries. Basic metadata are derived from the data source itself (e.g. parsed from file headers) and by value inspection, and then augmented using editable metadata templates containing boilerplate documentation, attribute descriptors, code definitions and quality control rules. Data and metadata content, quality control rules and qualifier flags are then managed together in a robust data structure that supports database functionality and ensures data validity throughout processing. A growing suite of metadata-aware editing, quality control, analysis and synthesis tools are provided with the software to support managing data using graphical forms and command-line functions, as well as

  10. Software Assurance Curriculum Project Volume 1: Master of Software Assurance Reference Curriculum

    Science.gov (United States)

    2010-08-01

    developed products. The above definition was derived from these references: [IEEE-CS 2008] ISO /IEC 12207 , IEEE Std 12207 -2008, Systems and Software...Systems [CNSS 2009]. Software quality Capability of a software product to satisfy stated and implied needs when used under specified conditions [ ISO ...Curriculum ISO International Organization for Standardization IT information technology KA knowledge area KU knowledge unit MBA Master of

  11. Formalizing the ISDF Software Development Methodology

    Directory of Open Access Journals (Sweden)

    Mihai Liviu DESPA

    2015-01-01

    Full Text Available The paper is aimed at depicting the ISDF software development methodology by emphasizing quality management and software development lifecycle. The ISDF methodology was built especially for innovative software development projects. The ISDF methodology was developed empirically by trial and error in the process of implementing multiple innovative projects. The research process began by analysing key concepts like innovation and software development and by settling the important dilemma of what makes a web application innovative. Innovation in software development is presented from the end-user, project owner and project manager’s point of view. The main components of a software development methodology are identified. Thus a software development methodology should account for people, roles, skills, teams, tools, techniques, processes, activities, standards, quality measuring tools, and team values. Current software development models are presented and briefly analysed. The need for a dedicated innovation oriented software development methodology is emphasized by highlighting shortcomings of current software development methodologies when tackling innovation. The ISDF methodology is presented in the context of developing an actual application. The ALHPA application is used as a case study for emphasizing the characteristics of the ISDF methodology. The development life cycle of the ISDF methodology includes research, planning, prototyping, design, development, testing, setup and maintenance. Artefacts generated by the ISDF methodology are presented. Quality is managed in the ISDF methodology by assessing compliance, usability, reliability, repeatability, availability and security. In order to properly asses each quality component a dedicated indicator is built. A template for interpreting each indicator is provided. Conclusions are formulated and new related research topics are submitted for debate.

  12. Genetic diversity of internalin genes in the ascB-dapE locus among Listeria monocytogenes lineages III and IV strains.

    Science.gov (United States)

    Chen, Jianshun; Cheng, Changyong; Lv, Yonghui; Fang, Weihuan

    2013-09-01

    Listeria monocytogenes is an important foodborne pathogen encompassing four phylogenetic lineages. Lineages III and IV are rare, but have been reported to show considerable biodiversity, providing important clues for the evolutionary history in Listeria. In this study, analysis of the ascB-dapE locus reveals genetic diversity in lineages III and IV, and is consistent with the classification of sublineages. Four of the six genetic patterns (two of sublineage IIIC and two of lineage IV) are specific to these two lineages. The ascB-dapE locus suggests a hot spot for genome diversification, and serves as an attractive molecular marker for better understanding of the biodiversity and population structure of lineages III and IV strains. © 2013 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  13. Software quality assurance documentation for the release of NUFT 2.0 for HP platforms

    International Nuclear Information System (INIS)

    Fernandez, M.W.; Preckshot, G.G.; Johnson, G.L.

    1998-01-01

    This document is the Individual Software Plan (ISP) for version 2.0 of the Non-isothermal Unsaturated-saturated Flow and Transport (NUFT.) analysis computer program. This document addresses the applicable requirements of LLNL YMP procedure 033-YMP-QP 3.2, Section 4.2.1.1. The purpose of this ISP is to plan and organize the activities required to certify the NUFT code for quality affecting work involving problems that include cross drift analysis of the Yucca Mountain Repository facility. NUFT is software for application to the solution of a class of coupled mass and heat transport problems in porous geologic media including Yucca Mountain Repository Cross Drift Problem (YMRCDP- also known as the Enhanced Characterization of the Repository Block (ECRB)). Solution of this class of problems requires a suite of multiphase, multi-component models for numerical solution of non- isothermal flow and transport in porous media with applications to subsurface contaminant transport problems. NUFT is a suite of multiphase, multi-component models for numerical solution of non- isothermal flow and transport in porous media, with application to subsurface contaminant transport problems, and in particular, to the hydrology in and about the Yucca Mountain Repository Site. NUFI is acquired software, as defined by 033-YMP-QP 3.2, and a preliminary baseline of source code, electronic documentation, and paper documentation has been established as required by 033-YMP-QP 3.2, Section 4.1. NUFT runs on Sun Unix platforms, Solaris operating system version 5.5 and HP-UX with operating system version 10.20. The product to be qualified under this ISP is the version running on HP- UX. The HP version will be labeled Version 2.0h. The h is included to distinguish the HP version from possible future versions qualified for Sun or other platforms. The scope of the plans and procedures outlined in this ISP is limited to the effort required to qualify NUFT for the class of problems identified in

  14. A quantitative calculation for software reliability evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Young-Jun; Lee, Jang-Soo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2016-10-15

    To meet these regulatory requirements, the software used in the nuclear safety field has been ensured through the development, validation, safety analysis, and quality assurance activities throughout the entire process life cycle from the planning phase to the installation phase. A variety of activities, such as the quality assurance activities are also required to improve the quality of a software. However, there are limitations to ensure that the quality is improved enough. Therefore, the effort to calculate the reliability of the software continues for a quantitative evaluation instead of a qualitative evaluation. In this paper, we propose a quantitative calculation method for the software to be used for a specific operation of the digital controller in an NPP. After injecting random faults in the internal space of a developed controller and calculating the ability to detect the injected faults using diagnostic software, we can evaluate the software reliability of a digital controller in an NPP. We tried to calculate the software reliability of the controller in an NPP using a new method that differs from a traditional method. It calculates the fault detection coverage after injecting the faults into the software memory space rather than the activity through the life cycle process. We attempt differentiation by creating a new definition of the fault, imitating the software fault using the hardware, and giving a consideration and weights for injection faults.

  15. Four simple recommendations to encourage best practices in research software

    DEFF Research Database (Denmark)

    Jiménez, Rafael C.; Kuzak, Mateusz; Alhamdoosh, Monther

    2017-01-01

    Scientific research relies on computer software, yet software is not always developed following practices that ensure its quality and sustainability. This manuscript does not aim to propose new software development best practices, but rather to provide simple recommendations that encourage...... the adoption of existing best practices. Software development best practices promote better quality software, and better quality software improves the reproducibility and reusability of research. These recommendations are designed around Open Source values, and provide practical suggestions that contribute...... to making research software and its source code more discoverable, reusable and transparent. This manuscript is aimed at developers, but also at organisations, projects, journals and funders that can increase the quality and sustainability of research software by encouraging the adoption...

  16. Four simple recommendations to encourage best practices in research software.

    Science.gov (United States)

    Jiménez, Rafael C; Kuzak, Mateusz; Alhamdoosh, Monther; Barker, Michelle; Batut, Bérénice; Borg, Mikael; Capella-Gutierrez, Salvador; Chue Hong, Neil; Cook, Martin; Corpas, Manuel; Flannery, Madison; Garcia, Leyla; Gelpí, Josep Ll; Gladman, Simon; Goble, Carole; González Ferreiro, Montserrat; Gonzalez-Beltran, Alejandra; Griffin, Philippa C; Grüning, Björn; Hagberg, Jonas; Holub, Petr; Hooft, Rob; Ison, Jon; Katz, Daniel S; Leskošek, Brane; López Gómez, Federico; Oliveira, Luis J; Mellor, David; Mosbergen, Rowland; Mulder, Nicola; Perez-Riverol, Yasset; Pergl, Robert; Pichler, Horst; Pope, Bernard; Sanz, Ferran; Schneider, Maria V; Stodden, Victoria; Suchecki, Radosław; Svobodová Vařeková, Radka; Talvik, Harry-Anton; Todorov, Ilian; Treloar, Andrew; Tyagi, Sonika; van Gompel, Maarten; Vaughan, Daniel; Via, Allegra; Wang, Xiaochuan; Watson-Haigh, Nathan S; Crouch, Steve

    2017-01-01

    Scientific research relies on computer software, yet software is not always developed following practices that ensure its quality and sustainability. This manuscript does not aim to propose new software development best practices, but rather to provide simple recommendations that encourage the adoption of existing best practices. Software development best practices promote better quality software, and better quality software improves the reproducibility and reusability of research. These recommendations are designed around Open Source values, and provide practical suggestions that contribute to making research software and its source code more discoverable, reusable and transparent. This manuscript is aimed at developers, but also at organisations, projects, journals and funders that can increase the quality and sustainability of research software by encouraging the adoption of these recommendations.

  17. An Alternative Optimization Model and Robust Experimental Design for the Assignment Scheduling Capability for Unmanned Aerial Vehicles (ASC-U) Simulation

    National Research Council Canada - National Science Library

    Oliver, Derek M

    2007-01-01

    .... TRAC selected ASC-U to provide insight into the programmatic decisions addressed in the U.S. Army UAV-Mix Analysis that directly affects future development and fielding of UAVs to include the Future Combat System...

  18. 42 CFR 416.43 - Conditions for coverage-Quality assessment and performance improvement.

    Science.gov (United States)

    2010-10-01

    ..., an ongoing program that demonstrates measurable improvement in patient health outcomes, and improves patient safety by using quality indicators or performance measures associated with improved health outcomes and by the identification and reduction of medical errors. (2) The ASC must measure, analyze, and...

  19. Third-Party Software's Trust Quagmire.

    Science.gov (United States)

    Voas, J; Hurlburt, G

    2015-12-01

    Current software development has trended toward the idea of integrating independent software sub-functions to create more complete software systems. Software sub-functions are often not homegrown - instead they are developed by unknown 3 rd party organizations and reside in software marketplaces owned or controlled by others. Such software sub-functions carry plausible concern in terms of quality, origins, functionality, security, interoperability, to name a few. This article surveys key technical difficulties in confidently building systems from acquired software sub-functions by calling out the principle software supply chain actors.

  20. Quantificação de compostos fenólicos e ácido ascórbico em frutos e polpas congeladas de acerola, caju, goiaba e morango

    Directory of Open Access Journals (Sweden)

    Juliana Mesquita Freire

    2013-12-01

    Full Text Available O crescente interesse da população pela promoção da saúde vem despertando a atenção dos consumidores por alimentos ricos em antioxidantes, combinado com a praticidade do consumo de sucos de frutos processados. O presente trabalho objetivou quantificar e avaliar a capacidade antioxidante de compostos fenólicos e ácido ascórbico em frutas in natura e polpas congeladas. As frutas in natura e polpas congeladas obtidas na região de Lavras - MG foram submetidas à quantificação de compostos fenólicos e ácido ascórbico e os respectivos extratos acetônico-etanólico e acetônico-metanólico. Foram determinadas suas atividades antioxidantes pelos métodos radicalares DPPH e ABTS. O extrato acetônico-metanólico foi mais efetivo para extrair os compostos antioxidantes das amostras. Apenas as polpas congeladas de acerola apresentaram uma redução do teor de compostos fenólicos em relação ao fruto. Para o teor de ácido ascórbico, houve uma redução nas polpas congeladas de acerola e goiaba. As amostras de acerola apresentaram os maiores teores de compostos fenólicos, ácido ascórbico e atividade antioxidante em relação aos demais frutos analisados.

  1. Automating Object-Oriented Software Development Methods

    NARCIS (Netherlands)

    Tekinerdogan, B.; Saeki, Motoshi; Sunyé, Gerson; van den Broek, P.M.; Hruby, Pavel; Tekinerdogan, B.; van den Broek, P.M.; Saeki, M.; Hruby, P.; Sunye, G.

    2001-01-01

    Current software projects have generally to deal with producing and managing large and complex software products. It is generally believed that applying software development methods are useful in coping with this complexity and for supporting quality. As such numerous object-oriented software

  2. Automating Object-Oriented Software Development Methods

    NARCIS (Netherlands)

    Tekinerdogan, B.; Frohner, A´ kos; Saeki, Motoshi; Sunyé, Gerson; van den Broek, P.M.; Hruby, Pavel

    2002-01-01

    Current software projects have generally to deal with producing and managing large and complex software products. It is generally believed that applying software development methods are useful in coping with this complexity and for supporting quality. As such numerous object-oriented software

  3. Assessing software quality at each step of its life-cycle to enhance reliability of control systems

    International Nuclear Information System (INIS)

    Hardion, V.; Buteau, A.; Leclercq, N.; Abeille, G.; Pierre-Joseph, Z.; Le, S.

    2012-01-01

    A distributed software control system aims to enhance the upgrade ability and reliability by sharing responsibility between several components. The disadvantage is that it makes it harder to detect problems on a significant number of modules. With Kaizen in mind we have chosen to continuously invest in automation to obtain a complete overview of software quality despite the growth of legacy code. The development process has already been mastered by staging each life-cycle step thanks to a continuous integration server based on JENKINS and MAVEN. We enhanced this process, focusing on 3 objectives: Automatic Test, Static Code Analysis and Post-Mortem Supervision. Now, the build process automatically includes a test section to detect regressions, incorrect behaviour and integration incompatibility. The in-house TANGOUNIT project satisfies the difficulties of testing distributed components such as Tango Devices. In the next step, the programming code has to pass a complete code quality check-up. The SONAR quality server has been integrated in the process, to collect each static code analysis and display the hot topics on summary web pages. Finally, the integration of Google BREAKPAD in every TANGO Devices gives us essential statistics from crash reports and enables us to replay the crash scenarios at any time. We have already gained greater visibility on current developments. Some concrete results will be presented including reliability enhancement, better management of subcontracted software development, quicker adoption of coding standards by new developers and understanding of impacts when moving to a new technology. (authors)

  4. Imperfect Information in Software Design Processes

    NARCIS (Netherlands)

    Noppen, J.A.R.

    2007-01-01

    The process of designing high-quality software systems is one of the major issues in software engineering research. Over the years, this has resulted in numerous design methods, each with specific qualities and drawbacks. For example, the Rational Unified Process is a comprehensive design process,

  5. International Liability Issues for Software Quality

    National Research Council Canada - National Science Library

    Mead, Nancy

    2003-01-01

    This report focuses on international law related to cybercrime, international information security standards, and software liability issues as they relate to information security for critical infrastructure applications...

  6. Quality of Radiomic Features in Glioblastoma Multiforme: Impact of Semi-Automated Tumor Segmentation Software.

    Science.gov (United States)

    Lee, Myungeun; Woo, Boyeong; Kuo, Michael D; Jamshidi, Neema; Kim, Jong Hyo

    2017-01-01

    The purpose of this study was to evaluate the reliability and quality of radiomic features in glioblastoma multiforme (GBM) derived from tumor volumes obtained with semi-automated tumor segmentation software. MR images of 45 GBM patients (29 males, 16 females) were downloaded from The Cancer Imaging Archive, in which post-contrast T1-weighted imaging and fluid-attenuated inversion recovery MR sequences were used. Two raters independently segmented the tumors using two semi-automated segmentation tools (TumorPrism3D and 3D Slicer). Regions of interest corresponding to contrast-enhancing lesion, necrotic portions, and non-enhancing T2 high signal intensity component were segmented for each tumor. A total of 180 imaging features were extracted, and their quality was evaluated in terms of stability, normalized dynamic range (NDR), and redundancy, using intra-class correlation coefficients, cluster consensus, and Rand Statistic. Our study results showed that most of the radiomic features in GBM were highly stable. Over 90% of 180 features showed good stability (intra-class correlation coefficient [ICC] ≥ 0.8), whereas only 7 features were of poor stability (ICC NDR ≥1), while above 35% of the texture features showed poor NDR (software tools provided sufficiently reliable tumor segmentation and feature stability; thus helping to overcome the inherent inter-rater and intra-rater variability of user intervention. However, certain aspects of feature quality, including NDR and redundancy, need to be assessed for determination of representative signature features before further development of radiomics.

  7. A proposed defect tracking model for classifying the inserted defect reports to enhance software quality control.

    Science.gov (United States)

    Sultan, Torky; Khedr, Ayman E; Sayed, Mostafa

    2013-01-01

    NONE DECLARED Defect tracking systems play an important role in the software development organizations as they can store historical information about defects. There are many research in defect tracking models and systems to enhance their capabilities to be more specifically tracking, and were adopted with new technology. Furthermore, there are different studies in classifying bugs in a step by step method to have clear perception and applicable method in detecting such bugs. This paper shows a new proposed defect tracking model for the purpose of classifying the inserted defects reports in a step by step method for more enhancement of the software quality.

  8. NASA software documentation standard software engineering program

    Science.gov (United States)

    1991-01-01

    The NASA Software Documentation Standard (hereinafter referred to as Standard) can be applied to the documentation of all NASA software. This Standard is limited to documentation format and content requirements. It does not mandate specific management, engineering, or assurance standards or techniques. This Standard defines the format and content of documentation for software acquisition, development, and sustaining engineering. Format requirements address where information shall be recorded and content requirements address what information shall be recorded. This Standard provides a framework to allow consistency of documentation across NASA and visibility into the completeness of project documentation. This basic framework consists of four major sections (or volumes). The Management Plan contains all planning and business aspects of a software project, including engineering and assurance planning. The Product Specification contains all technical engineering information, including software requirements and design. The Assurance and Test Procedures contains all technical assurance information, including Test, Quality Assurance (QA), and Verification and Validation (V&V). The Management, Engineering, and Assurance Reports is the library and/or listing of all project reports.

  9. Automated pre-processing and multivariate vibrational spectra analysis software for rapid results in clinical settings

    Science.gov (United States)

    Bhattacharjee, T.; Kumar, P.; Fillipe, L.

    2018-02-01

    Vibrational spectroscopy, especially FTIR and Raman, has shown enormous potential in disease diagnosis, especially in cancers. Their potential for detecting varied pathological conditions are regularly reported. However, to prove their applicability in clinics, large multi-center multi-national studies need to be undertaken; and these will result in enormous amount of data. A parallel effort to develop analytical methods, including user-friendly software that can quickly pre-process data and subject them to required multivariate analysis is warranted in order to obtain results in real time. This study reports a MATLAB based script that can automatically import data, preprocess spectra— interpolation, derivatives, normalization, and then carry out Principal Component Analysis (PCA) followed by Linear Discriminant Analysis (LDA) of the first 10 PCs; all with a single click. The software has been verified on data obtained from cell lines, animal models, and in vivo patient datasets, and gives results comparable to Minitab 16 software. The software can be used to import variety of file extensions, asc, .txt., .xls, and many others. Options to ignore noisy data, plot all possible graphs with PCA factors 1 to 5, and save loading factors, confusion matrices and other parameters are also present. The software can provide results for a dataset of 300 spectra within 0.01 s. We believe that the software will be vital not only in clinical trials using vibrational spectroscopic data, but also to obtain rapid results when these tools get translated into clinics.

  10. IEEE Computer Society/Software Engineering Institute Software Process Achievement (SPA) Award 2009

    Science.gov (United States)

    2011-03-01

    capabilities to our GDM. We also introduced software as a service ( SaaS ) as part our technology solutions and have further enhanced our ability to...model PROSPER Infosys production support methodology Q&P quality and productivity R&D research and development SaaS software as a service ... Software Development Life Cycle (SDLC) 23 Table 10: Scientific Estimation Coverage by Service Line 27 CMU/SEI-2011-TR-008 | vi CMU/SEI-2011

  11. Essential software architecture

    CERN Document Server

    Gorton, Ian

    2011-01-01

    Job titles like ""Technical Architect"" and ""Chief Architect"" nowadays abound in software industry, yet many people suspect that ""architecture"" is one of the most overused and least understood terms in professional software development. Gorton's book tries to resolve this dilemma. It concisely describes the essential elements of knowledge and key skills required to be a software architect. The explanations encompass the essentials of architecture thinking, practices, and supporting technologies. They range from a general understanding of structure and quality attributes through technical i

  12. ISO and software quality assurance - licensing and certification of software professionals

    Energy Technology Data Exchange (ETDEWEB)

    Hare, J.; Rodin, L.

    1997-11-01

    This report contains viewgraphs on licensing and certifing of software professionals. Discussed in this report are: certification programs; licensing programs; why became certified; certification as a condition of empolyment; certification requirements; and examination structures.

  13. Xyce parallel electronic simulator design.

    Energy Technology Data Exchange (ETDEWEB)

    Thornquist, Heidi K.; Rankin, Eric Lamont; Mei, Ting; Schiek, Richard Louis; Keiter, Eric Richard; Russo, Thomas V.

    2010-09-01

    This document is the Xyce Circuit Simulator developer guide. Xyce has been designed from the 'ground up' to be a SPICE-compatible, distributed memory parallel circuit simulator. While it is in many respects a research code, Xyce is intended to be a production simulator. As such, having software quality engineering (SQE) procedures in place to insure a high level of code quality and robustness are essential. Version control, issue tracking customer support, C++ style guildlines and the Xyce release process are all described. The Xyce Parallel Electronic Simulator has been under development at Sandia since 1999. Historically, Xyce has mostly been funded by ASC, the original focus of Xyce development has primarily been related to circuits for nuclear weapons. However, this has not been the only focus and it is expected that the project will diversify. Like many ASC projects, Xyce is a group development effort, which involves a number of researchers, engineers, scientists, mathmaticians and computer scientists. In addition to diversity of background, it is to be expected on long term projects for there to be a certain amount of staff turnover, as people move on to different projects. As a result, it is very important that the project maintain high software quality standards. The point of this document is to formally document a number of the software quality practices followed by the Xyce team in one place. Also, it is hoped that this document will be a good source of information for new developers.

  14. Software quality assurance documentation for the release of NUFT 2.0s for SUN platforms

    International Nuclear Information System (INIS)

    Fernandez, M.W.; Shaffer, R.J.

    1998-01-01

    The purpose of this Individual Software Plan (ISP) is to plan and organize the activities required to certify the NUFT code for quality affecting work involving problems that include cross drift analysis of the Yucca Mountain Repository facility. NUFT is software for application to the solution of a class of coupled mass and heat transport problems in porous geologic media including Yucca Mountain Repository Cross Drift Problem (YMRCDP- also known as the Enhanced Characterization of the Repository Block (ECRB)). Solution of this class of problems requires a suite of multiphase, multi-component models for numerical solution of non-isothermal flow and transport in porous media with applications to subsurface contaminant transport problems

  15. Genetic and physical analysis of a YAC contig spanning the fungal disease resistance locus Asc of tomato (Lycopersicon esculentum)

    NARCIS (Netherlands)

    Mesbah, L.A.; Kneppers, T.J.A.; Takken, F.L.W.; Laurent, P.; Hille, J.; Nijkamp, H.J.J.

    1998-01-01

    The Alternaria stem canker disease of tomato is caused by the necrotrophic fungal pathogen Alternaria alternata f. sp. lycopersici (AAL). The fungus produces AAL toxins that kill the plant tissue. Resistance to the fungus segregates as a single locus, called Asc, and has been genetically mapped on

  16. Genetic and physical analysis of a YAC contig spannig the fungal disease resistance locus Asc of tomato (Lycopersicon esculentum)

    NARCIS (Netherlands)

    Mesbah, L.A.; Kneppers, T.J.A.; Takken, F.L.W.; Laurent, P.J.F.; Hille, J.; Nijkamp, H.J.J.

    1999-01-01

    The Alternaria in stem canker disease of tomato is caused by the necrotrophic fungal pathogen Alternaria alternata f. sp. lycopersici (AAL). The fungus produces AAL toxins that kill the plant tissue. Resistance to the fungus segregates as a single locus, called Asc, and has been genetically mapped

  17. The primary protection system software

    International Nuclear Information System (INIS)

    Tooley, P.A.

    1992-01-01

    This paper continues the detailed description of the Primary Protection System for Sizewell-B by providing an overview of design and implementation of the software, including the features of the design process which ensure that quality is delivered by the contractor. The Nuclear Electric software assessment activities are also described. The argument for the excellence of the software is made on the basis of a quality product delivered by the equipment supplier's design process, and the confirmation of this provided by the Nuclear Electric assessment process, which is as searching and complete an examination as is reasonably practicable to achieve. (author)

  18. Improving the quality of numerical software through user-centered design

    Energy Technology Data Exchange (ETDEWEB)

    Pancake, C. M., Oregon State University

    1998-06-01

    The software interface - whether graphical, command-oriented, menu-driven, or in the form of subroutine calls - shapes the user`s perception of what software can do. It also establishes upper bounds on software usability. Numerical software interfaces typically are based on the designer`s understanding of how the software should be used. That is a poor foundation for usability, since the features that are ``instinctively right`` from the developer`s perspective are often the very ones that technical programmers find most objectionable or most difficult to learn. This paper discusses how numerical software interfaces can be improved by involving users more actively in design, a process known as user-centered design (UCD). While UCD requires extra organization and effort, it results in much higher levels of usability and can actually reduce software costs. This is true not just for graphical user interfaces, but for all software interfaces. Examples show how UCD improved the usability of a subroutine library, a command language, and an invocation interface.

  19. Clinical evaluation of reducing acquisition time on single-photon emission computed tomography image quality using proprietary resolution recovery software.

    Science.gov (United States)

    Aldridge, Matthew D; Waddington, Wendy W; Dickson, John C; Prakash, Vineet; Ell, Peter J; Bomanji, Jamshed B

    2013-11-01

    A three-dimensional model-based resolution recovery (RR) reconstruction algorithm that compensates for collimator-detector response, resulting in an improvement in reconstructed spatial resolution and signal-to-noise ratio of single-photon emission computed tomography (SPECT) images, was tested. The software is said to retain image quality even with reduced acquisition time. Clinically, any improvement in patient throughput without loss of quality is to be welcomed. Furthermore, future restrictions in radiotracer supplies may add value to this type of data analysis. The aims of this study were to assess improvement in image quality using the software and to evaluate the potential of performing reduced time acquisitions for bone and parathyroid SPECT applications. Data acquisition was performed using the local standard SPECT/CT protocols for 99mTc-hydroxymethylene diphosphonate bone and 99mTc-methoxyisobutylisonitrile parathyroid SPECT imaging. The principal modification applied was the acquisition of an eight-frame gated data set acquired using an ECG simulator with a fixed signal as the trigger. This had the effect of partitioning the data such that the effect of reduced time acquisitions could be assessed without conferring additional scanning time on the patient. The set of summed data sets was then independently reconstructed using the RR software to permit a blinded assessment of the effect of acquired counts upon reconstructed image quality as adjudged by three experienced observers. Data sets reconstructed with the RR software were compared with the local standard processing protocols; filtered back-projection and ordered-subset expectation-maximization. Thirty SPECT studies were assessed (20 bone and 10 parathyroid). The images reconstructed with the RR algorithm showed improved image quality for both full-time and half-time acquisitions over local current processing protocols (Pimproved image quality compared with local processing protocols and has been

  20. QRev—Software for computation and quality assurance of acoustic doppler current profiler moving-boat streamflow measurements—Technical manual for version 2.8

    Science.gov (United States)

    Mueller, David S.

    2016-06-21

    The software program, QRev applies common and consistent computational algorithms combined with automated filtering and quality assessment of the data to improve the quality and efficiency of streamflow measurements and helps ensure that U.S. Geological Survey streamflow measurements are consistent, accurate, and independent of the manufacturer of the instrument used to make the measurement. Software from different manufacturers uses different algorithms for various aspects of the data processing and discharge computation. The algorithms used by QRev to filter data, interpolate data, and compute discharge are documented and compared to the algorithms used in the manufacturers’ software. QRev applies consistent algorithms and creates a data structure that is independent of the data source. QRev saves an extensible markup language (XML) file that can be imported into databases or electronic field notes software. This report is the technical manual for version 2.8 of QRev.

  1. Continuous software engineering – a microservices architecture perspective

    OpenAIRE

    O'Connor, Rory; Elger, Peter; Clarke, Paul

    2017-01-01

    From its earliest days, software development has been beset with challenges in relation to timely delivery, appropriateness of features and quality of deliverables. Many advances in software development processes have helped to address these concerns. For example, agile software development has helped to deliver working software more frequently and capability maturity frameworks have brought about improved consistency in quality levels. However, the age-old challenge of better, cheaper, faste...

  2. A new paradigm for the development of analysis software

    International Nuclear Information System (INIS)

    Kelly, D.; Harauz, J.

    2012-01-01

    For the CANDU industry, analysis software is an important tool for scientists and engineers to examine issues related to safety, operation, and design. However, the software quality assurance approach currently used for these tools assumes the software is the delivered product. In this paper, we present a model that shifts the emphasis from software being the end-product to software being support for the end-product, the science. We describe a novel software development paradigm that supports this shift and provides the groundwork for re-examining the quality assurance practices used for analysis software. (author)

  3. Qualidade de kiwis minimamente processados e submetidos a tratamento com ácido ascórbico, ácido cítrico e cloreto de cálcio Quality of kiwis minimally processed and treated with ascorbic acid, citric acid and calcium chloride

    Directory of Open Access Journals (Sweden)

    Ana Vânia Carvalho

    2002-05-01

    Full Text Available Frutos e hortaliças minimamente processados devem apresentar atributos de conveniência e qualidade do produto fresco. O objetivo deste trabalho foi estudar o efeito do processamento mínimo de frutos tratados com soluções a 1% de ácido ascórbico, ácido cítrico e cloreto de cálcio, durante armazenamento refrigerado, na qualidade do kiwi (Actinidia deliciosa cv. Hayward. A perda de massa foi mínima durante o período de armazenamento. O ácido ascórbico fornecido pelo tratamento foi eficientemente absorvido pelos tecidos, mantendo os níveis de vitamina C cerca de 25% mais elevados nesses frutos do que nos demais tratamentos. A análise microbiológica detectou presença de bolores e leveduras e psicrotróficos, somente no tratamento com ácido cítrico, aos 8 e 10 dias, respectivamente. Não se detectaram coliformes totais e fecais e mesófilos, o que indica que o processamento foi realizado em boas condições higiênicas. Os kiwis minimamente processados e tratados com cloreto de cálcio apresentaram uma vida útil de dez dias. Nos demais tratamentos e no controle, esse tempo foi de seis dias.Minimally processed fruits and vegetables might present the same convenience and quality of fresh products. In this work, the influence of minimal processing of fruits treated with 1% solutions of ascorbic acid, citric acid and calcium chloride on the quality of kiwi (Actinidia deliciosa cv. Hayward, during refrigerated storage were investigated. Mass loss was minimal over the storage period. Ascorbic acid furnished by the treatment was effectively absorbed by tissues, keeping the vitamin C levels 25% higher in those fruits, than in other treatments. Microbiological analysis detected the presence of the group molds and yeasts and psychrotrophic in citric acid treatment, at 8 and 10 days, respectively. Total and faecal coliforms and mesophyles were not found, indicating that processing was performed under good hygienic conditions. Minimally

  4. Software engineering design theory and practice

    CERN Document Server

    Otero, Carlos

    2012-01-01

    … intended for use as a textbook for an advanced course in software design. Each chapter ends with review questions and references. … provides an overview of the software development process, something that would not be out of line in a course on software engineering including such topics as software process, software management, balancing conflicting values of stakeholders, testing, quality, and ethics. The author has principally focused on software design though, extracting the design phase from the surrounding software development lifecycle. … Software design strategies are addressed

  5. Secure software development training course

    Directory of Open Access Journals (Sweden)

    Victor S. Gorbatov

    2017-06-01

    Full Text Available Information security is one of the most important criteria for the quality of developed software. To obtain a sufficient level of application security companies implement security process into software development life cycle. At this stage software companies encounter with deficit employees who able to solve problems of software design, implementation and application security. This article provides a description of the secure software development training course. Training course of application security is designed for co-education students of different IT-specializations.

  6. Network protocol changes can improve DisCom WAN performance : evaluating TCP modifications and SCTP in the ASC tri-lab environment.

    Energy Technology Data Exchange (ETDEWEB)

    Tolendino, Lawrence F.; Hu, Tan Chang

    2005-06-01

    The Advanced Simulation and Computing (ASC) Distance Computing (DisCom) Wide Area Network (WAN) is a high performance, long distance network environment that is based on the ubiquitous TCP/IP protocol set. However, the Transmission Control Protocol (TCP) and the algorithms that govern its operation were defined almost two decades ago for a network environment vastly different from the DisCom WAN. In this paper we explore and evaluate possible modifications to TCP that purport to improve TCP performance in environments like the DisCom WAN. We also examine a much newer protocol, SCTP (Stream Control Transmission Protocol) that claims to provide reliable network transport while also implementing multi-streaming, multi-homing capabilities that are appealing in the DisCom high performance network environment. We provide performance comparisons and recommendations for continued development that will lead to network communications protocol implementations capable of supporting the coming ASC Petaflop computing environments.

  7. Software testing in roughness calculation

    International Nuclear Information System (INIS)

    Chen, Y L; Hsieh, P F; Fu, W E

    2005-01-01

    A test method to determine the function quality provided by the software for roughness measurement is presented in this study. The function quality of the software requirements should be part of and assessed through the entire life cycle of the software package. The specific function, or output accuracy, is crucial for the analysis of the experimental data. For scientific applications, however, commercial software is usually embedded with specific instrument, which is used for measurement or analysis during the manufacture process. In general, the error ratio caused by the software would be more apparent especially when dealing with relatively small quantities, like the measurements in the nanometer-scale range. The model of 'using a data generator' proposed by NPL of UK was applied in this study. An example of the roughness software is tested and analyzed by the above mentioned process. After selecting the 'reference results', the 'reference data' was generated by a programmable 'data generator'. The filter function of 0.8 mm long cutoff value, defined in ISO 11562 was tested with 66 sinusoid data at different wavelengths. Test results from commercial software and CMS written program were compared to the theoretical data calculated from ISO standards. As for the filter function in this software, the result showed a significant disagreement between the reference and test results. The short cutoff feature for filtering at the high frequencies does not function properly, while the long cutoff feature has the maximum difference in the filtering ratio, which is more than 70% between the wavelength of 300 μm and 500 μm. Conclusively, the commercial software needs to be tested more extensively for specific application by appropriate design of reference dataset to ensure its function quality

  8. Software Defined Networking Demands on Software Technologies

    DEFF Research Database (Denmark)

    Galinac Grbac, T.; Caba, Cosmin Marius; Soler, José

    2015-01-01

    Software Defined Networking (SDN) is a networking approach based on a centralized control plane architecture with standardised interfaces between control and data planes. SDN enables fast configuration and reconfiguration of the network to enhance resource utilization and service performances....... This new approach enables a more dynamic and flexible network, which may adapt to user needs and application requirements. To this end, systemized solutions must be implemented in network software, aiming to provide secure network services that meet the required service performance levels. In this paper......, we review this new approach to networking from an architectural point of view, and identify and discuss some critical quality issues that require new developments in software technologies. These issues we discuss along with use case scenarios. Here in this paper we aim to identify challenges...

  9. A software engineering process for safety-critical software application

    International Nuclear Information System (INIS)

    Kang, Byung Heon; Kim, Hang Bae; Chang, Hoon Seon; Jeon, Jong Sun

    1995-01-01

    Application of computer software to safety-critical systems in on the increase. To be successful, the software must be designed and constructed to meet the functional and performance requirements of the system. For safety reason, the software must be demonstrated not only to meet these requirements, but also to operate safely as a component within the system. For longer-term cost consideration, the software must be designed and structured to ease future maintenance and modifications. This paper presents a software engineering process for the production of safety-critical software for a nuclear power plant. The presentation is expository in nature of a viable high quality safety-critical software development. It is based on the ideas of a rational design process and on the experience of the adaptation of such process in the production of the safety-critical software for the shutdown system number two of Wolsung 2, 3 and 4 nuclear power generation plants. This process is significantly different from a conventional process in terms of rigorous software development phases and software design techniques, The process covers documentation, design, verification and testing using mathematically precise notations and highly reviewable tabular format to specify software requirements and software requirements and software requirements and code against software design using static analysis. The software engineering process described in this paper applies the principle of information-hiding decomposition in software design using a modular design technique so that when a change is required or an error is detected, the affected scope can be readily and confidently located. it also facilitates a sense of high degree of confidence in the 'correctness' of the software production, and provides a relatively simple and straightforward code implementation effort. 1 figs., 10 refs. (Author)

  10. A software sensor model based on hybrid fuzzy neural network for rapid estimation water quality in Guangzhou section of Pearl River, China.

    Science.gov (United States)

    Zhou, Chunshan; Zhang, Chao; Tian, Di; Wang, Ke; Huang, Mingzhi; Liu, Yanbiao

    2018-01-02

    In order to manage water resources, a software sensor model was designed to estimate water quality using a hybrid fuzzy neural network (FNN) in Guangzhou section of Pearl River, China. The software sensor system was composed of data storage module, fuzzy decision-making module, neural network module and fuzzy reasoning generator module. Fuzzy subtractive clustering was employed to capture the character of model, and optimize network architecture for enhancing network performance. The results indicate that, on basis of available on-line measured variables, the software sensor model can accurately predict water quality according to the relationship between chemical oxygen demand (COD) and dissolved oxygen (DO), pH and NH 4 + -N. Owing to its ability in recognizing time series patterns and non-linear characteristics, the software sensor-based FNN is obviously superior to the traditional neural network model, and its R (correlation coefficient), MAPE (mean absolute percentage error) and RMSE (root mean square error) are 0.8931, 10.9051 and 0.4634, respectively.

  11. ROLE OF DATA MINING CLASSIFICATION TECHNIQUE IN SOFTWARE DEFECT PREDICTION

    OpenAIRE

    Dr.A.R.Pon Periyasamy; Mrs A.Misbahulhuda

    2017-01-01

    Software defect prediction is the process of locating defective modules in software. Software quality may be a field of study and apply that describes the fascinating attributes of software package product. The performance should be excellent with none defects. Software quality metrics are a set of software package metrics that target the standard aspects of the product, process, and project. The software package defect prediction model helps in early detection of defects and contributes to t...

  12. When to make proprietary software open source

    NARCIS (Netherlands)

    Caulkins, J.P.; Feichtinger, G.; Grass, D.; Hartl, R.F.; Kort, P.M.; Seidl, A.

    Software can be distributed closed source (proprietary) or open source (developed collaboratively). While a firm cannot sell open source software, and so loses potential sales revenue, the open source software development process can have a substantial positive impact on the quality of a software,

  13. Software process improvement in a research environment

    NARCIS (Netherlands)

    Velden, van der M.J.; Hendriks, P.R.H.; Udink ten Cate, A.J.

    1995-01-01

    Research organizations pay much attention to the quality of their work, but not always to the quality of the software they produce within research projects. This is not a healthy situation since research organizations are becoming more and more dependent on software development. This paper describes

  14. How does Software Process Improvement Address Global Software Engineering?

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Diebold, Philipp; Münch, Jürgen

    2016-01-01

    a systematic mapping study on the state-of-the-art in SPI from a general perspective, we observed Global Software Engineering (GSE) becoming a topic of interest in recent years. Therefore, in this paper, we provide a detailed investigation of those papers from the overall systematic mapping study that were......For decades, Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. To set up, guide, and carry out SPI projects, and to measure SPI state, impact, and success, a multitude of different SPI approaches and considerable...... experience are available. SPI addresses many aspects ranging from individual developer skills to entire organizations. It comprises for instance the optimization of specific activities in the software lifecycle as well as the creation of organization awareness and project culture. In the course of conducting...

  15. Comparison of quality control software tools for diffusion tensor imaging.

    Science.gov (United States)

    Liu, Bilan; Zhu, Tong; Zhong, Jianhui

    2015-04-01

    Image quality of diffusion tensor imaging (DTI) is critical for image interpretation, diagnostic accuracy and efficiency. However, DTI is susceptible to numerous detrimental artifacts that may impair the reliability and validity of the obtained data. Although many quality control (QC) software tools are being developed and are widely used and each has its different tradeoffs, there is still no general agreement on an image quality control routine for DTIs, and the practical impact of these tradeoffs is not well studied. An objective comparison that identifies the pros and cons of each of the QC tools will be helpful for the users to make the best choice among tools for specific DTI applications. This study aims to quantitatively compare the effectiveness of three popular QC tools including DTI studio (Johns Hopkins University), DTIprep (University of North Carolina at Chapel Hill, University of Iowa and University of Utah) and TORTOISE (National Institute of Health). Both synthetic and in vivo human brain data were used to quantify adverse effects of major DTI artifacts to tensor calculation as well as the effectiveness of different QC tools in identifying and correcting these artifacts. The technical basis of each tool was discussed, and the ways in which particular techniques affect the output of each of the tools were analyzed. The different functions and I/O formats that three QC tools provide for building a general DTI processing pipeline and integration with other popular image processing tools were also discussed. Copyright © 2015 Elsevier Inc. All rights reserved.

  16. Quality control of software in dissimilar systems using a common clinical data base

    International Nuclear Information System (INIS)

    Erickson, J.J.; Price, R.R.; Touya, J.J.; Kronenberg, M.W.; Pederson, R.; Rollo, F.D.

    1980-01-01

    For a long time there has been widespread interest in the quality control of diagnostic instrumentation. The increasing dependence on computational systems for clinical results makes it imperative that methods for quality control of diagnostic software be developed. This paper proposes a method based on the use of a collection of patient studies for which the results have been corroborated by independent methods. The data set will be distributed in a format requiring no special handling by the system being tested and will appear identical to studies actually collected by the host system. An example of the use of a preliminary version of the data set for comparison of two systems is shown. The comparison shows that analyses performed on the two systems agree very well and can be reliably compared for follow-up studies of a patient

  17. Effect of metal artifact reduction software on image quality of C-arm cone-beam computed tomography during intracranial aneurysm treatment.

    Science.gov (United States)

    Enomoto, Yukiko; Yamauchi, Keita; Asano, Takahiko; Otani, Katharina; Iwama, Toru

    2018-01-01

    Background and purpose C-arm cone-beam computed tomography (CBCT) has the drawback that image quality is degraded by artifacts caused by implanted metal objects. We evaluated whether metal artifact reduction (MAR) prototype software can improve the subjective image quality of CBCT images of patients with intracranial aneurysms treated with coils or clips. Materials and methods Forty-four patients with intracranial aneurysms implanted with coils (40 patients) or clips (four patients) underwent one CBCT scan from which uncorrected and MAR-corrected CBCT image datasets were reconstructed. Three blinded readers evaluated the image quality of the image sets using a four-point scale (1: Excellent, 2: Good, 3: Poor, 4: Bad). The median scores of the three readers of uncorrected and MAR-corrected images were compared with the paired Wilcoxon signed-rank and inter-reader agreement of change scores was assessed by weighted kappa statistics. The readers also recorded new clinical findings, such as intracranial hemorrhage, air, or surrounding anatomical structures on MAR-corrected images. Results The image quality of MAR-corrected CBCT images was significantly improved compared with the uncorrected CBCT image ( p software improved image quality of CBCT images degraded by metal artifacts.

  18. Sandia Software Guidelines, Volume 2. Documentation

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1995-09-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standards for software documentation, this volume provides guidance in the selection of an adequate document set for a software project and example formats for many types of software documentation. A tutorial on life cycle documentation is also provided. Extended document thematic outlines and working examples of software documents are available on electronic media as an extension of this volume.

  19. Quality of radiomic features in glioblastoma multiforme: Impact of semi-automated tumor segmentation software

    International Nuclear Information System (INIS)

    Lee, Myung Eun; Kim, Jong Hyo; Woo, Bo Yeong; Ko, Micheal D.; Jamshidi, Neema

    2017-01-01

    The purpose of this study was to evaluate the reliability and quality of radiomic features in glioblastoma multiforme (GBM) derived from tumor volumes obtained with semi-automated tumor segmentation software. MR images of 45 GBM patients (29 males, 16 females) were downloaded from The Cancer Imaging Archive, in which post-contrast T1-weighted imaging and fluid-attenuated inversion recovery MR sequences were used. Two raters independently segmented the tumors using two semi-automated segmentation tools (TumorPrism3D and 3D Slicer). Regions of interest corresponding to contrast-enhancing lesion, necrotic portions, and non-enhancing T2 high signal intensity component were segmented for each tumor. A total of 180 imaging features were extracted, and their quality was evaluated in terms of stability, normalized dynamic range (NDR), and redundancy, using intra-class correlation coefficients, cluster consensus, and Rand Statistic. Our study results showed that most of the radiomic features in GBM were highly stable. Over 90% of 180 features showed good stability (intra-class correlation coefficient [ICC] ≥ 0.8), whereas only 7 features were of poor stability (ICC NDR ≥1), while above 35% of the texture features showed poor NDR (< 1). Features were shown to cluster into only 5 groups, indicating that they were highly redundant. The use of semi-automated software tools provided sufficiently reliable tumor segmentation and feature stability; thus helping to overcome the inherent inter-rater and intra-rater variability of user intervention. However, certain aspects of feature quality, including NDR and redundancy, need to be assessed for determination of representative signature features before further development of radiomics

  20. Study of evaluation techniques of software configuration management and reliability

    Energy Technology Data Exchange (ETDEWEB)

    Youn, Cheong; Baek, Y. W.; Kim, H. C.; Han, H. C.; Choi, C. R. [Chungnam National Univ., Taejon (Korea, Republic of)

    2001-03-15

    The Study of activities to solve software safety and quality must be executed in base of establishing software development process for digitalized nuclear plant. Especially study of software testing and Verification and Validation must executed. For this purpose methodologies and tools which can improve software qualities are evaluated and software Testing, V and V and Configuration Management which can be applied to software life cycle are investigated. This study establish a guideline that can be used to assure software safety and reliability requirements in digitalized nuclear plant systems.

  1. Adoção de Cidadãos Presos e Formação de Professores para a Prisão: Ações de Fraternidade Política e Direitos, aproximando a extensão universitária da ASCES e da UFPE no agreste pernambucano. Adoption of inmates and teacher training for prison: Actions of political fraternity and human rights approaching university extension to ASCES and UFPE in the rural area of Pernambuco state, Brazil

    Directory of Open Access Journals (Sweden)

    Jordão, Maria Perpétua S. D.

    2014-08-01

    na preocupação com a formação de valores civilizatórios que tenham como base a dignidade das pessoas livres ou encarceradas. O referencial teórico que subsidia o trabalho tem, como autores principais: Paulo Freire, Hannah Arendt e Barros. This study aims to analyze how two extension projects - Legal Adoption of Inmates (ASCES and Teacher Training for Penitentiary System (UFPE - enable coordinated actions of citizenship and human rights in a prison facility located in the ‘agreste’ region of Pernambuco state, Brazil, since 2001 (ASCES and 2009 (UFPE. The work was articulated based on the participation of ASCES teachers in the Research Group of UFPE-CNPq, and on the perception of how the integration of institutional activities could improve results and include a larger number of teachers, students, and beneficiaries - prisoners and educators of the penitentiary system. The actions involving broad participation of civil society were carried out fortnightly at the prison unit; the works are presented in the form of training meetings with teachers, lectures, and discussions with inmates, as well as in activities mediated by playful elements such as images, photographs, films, documentaries, music, and poetry. Students from the two institutions are integrated without competition; teachers receive continuing education; and recreational activities with the inmates break their prison routine and encourage them to remain in the classroom, reducing evasion. The activity is used as internship workload for the ASCES students and as supplementary activities for students of Pedagogy. These activities also allow the two institutions to contribute for improving the quality of education in prison; motivate the interaction of students from different institutions, including students form institutions that do not belong to the projects, attracting the participation of many volunteers and alumni from the ASCES and the Faculty of Philosophy, Sciences and Languages of Caruaru

  2. Simple solution to the medical instrumentation software problem

    Science.gov (United States)

    Leif, Robert C.; Leif, Suzanne B.; Leif, Stephanie H.; Bingue, E.

    1995-04-01

    Medical devices now include a substantial software component, which is both difficult and expensive to produce and maintain. Medical software must be developed according to `Good Manufacturing Practices', GMP. Good Manufacturing Practices as specified by the FDA and ISO requires the definition and compliance to a software processes which ensures quality products by specifying a detailed method of software construction. The software process should be based on accepted standards. US Department of Defense software standards and technology can both facilitate the development and improve the quality of medical systems. We describe the advantages of employing Mil-Std-498, Software Development and Documentation, and the Ada programming language. Ada provides the very broad range of functionalities, from embedded real-time to management information systems required by many medical devices. It also includes advanced facilities for object oriented programming and software engineering.

  3. SU-F-J-72: A Clinical Usable Integrated Contouring Quality Evaluation Software for Radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Jiang, S; Dolly, S; Cai, B; Mutic, S; Li, H [Washington University School of Medicine, Saint Louis, MO (United States)

    2016-06-15

    Purpose: To introduce the Auto Contour Evaluation (ACE) software, which is the clinical usable, user friendly, efficient and all-in-one toolbox for automatically identify common contouring errors in radiotherapy treatment planning using supervised machine learning techniques. Methods: ACE is developed with C# using Microsoft .Net framework and Windows Presentation Foundation (WPF) for elegant GUI design and smooth GUI transition animations through the integration of graphics engines and high dots per inch (DPI) settings on modern high resolution monitors. The industrial standard software design pattern, Model-View-ViewModel (MVVM) pattern, is chosen to be the major architecture of ACE for neat coding structure, deep modularization, easy maintainability and seamless communication with other clinical software. ACE consists of 1) a patient data importing module integrated with clinical patient database server, 2) a 2D DICOM image and RT structure simultaneously displaying module, 3) a 3D RT structure visualization module using Visualization Toolkit or VTK library and 4) a contour evaluation module using supervised pattern recognition algorithms to detect contouring errors and display detection results. ACE relies on supervised learning algorithms to handle all image processing and data processing jobs. Implementations of related algorithms are powered by Accord.Net scientific computing library for better efficiency and effectiveness. Results: ACE can take patient’s CT images and RT structures from commercial treatment planning software via direct user input or from patients’ database. All functionalities including 2D and 3D image visualization and RT contours error detection have been demonstrated with real clinical patient cases. Conclusion: ACE implements supervised learning algorithms and combines image processing and graphical visualization modules for RT contours verification. ACE has great potential for automated radiotherapy contouring quality verification

  4. SU-F-J-72: A Clinical Usable Integrated Contouring Quality Evaluation Software for Radiotherapy

    International Nuclear Information System (INIS)

    Jiang, S; Dolly, S; Cai, B; Mutic, S; Li, H

    2016-01-01

    Purpose: To introduce the Auto Contour Evaluation (ACE) software, which is the clinical usable, user friendly, efficient and all-in-one toolbox for automatically identify common contouring errors in radiotherapy treatment planning using supervised machine learning techniques. Methods: ACE is developed with C# using Microsoft .Net framework and Windows Presentation Foundation (WPF) for elegant GUI design and smooth GUI transition animations through the integration of graphics engines and high dots per inch (DPI) settings on modern high resolution monitors. The industrial standard software design pattern, Model-View-ViewModel (MVVM) pattern, is chosen to be the major architecture of ACE for neat coding structure, deep modularization, easy maintainability and seamless communication with other clinical software. ACE consists of 1) a patient data importing module integrated with clinical patient database server, 2) a 2D DICOM image and RT structure simultaneously displaying module, 3) a 3D RT structure visualization module using Visualization Toolkit or VTK library and 4) a contour evaluation module using supervised pattern recognition algorithms to detect contouring errors and display detection results. ACE relies on supervised learning algorithms to handle all image processing and data processing jobs. Implementations of related algorithms are powered by Accord.Net scientific computing library for better efficiency and effectiveness. Results: ACE can take patient’s CT images and RT structures from commercial treatment planning software via direct user input or from patients’ database. All functionalities including 2D and 3D image visualization and RT contours error detection have been demonstrated with real clinical patient cases. Conclusion: ACE implements supervised learning algorithms and combines image processing and graphical visualization modules for RT contours verification. ACE has great potential for automated radiotherapy contouring quality verification

  5. Software engineering in industry

    Science.gov (United States)

    Story, C. M.

    1989-12-01

    Can software be "engineered"? Can a few people with limited resources and a negligible budget produce high quality software solutions to complex software problems? It is possible to resolve the conflict between research activities and the necessity to view software development as a means to an end rather than as an end in itself? The aim of this paper is to encourage further thought and discussion on various topics which, in the author's experience, are becoming increasingly critical in large current software production and development projects, inside and outside high energy physics (HEP). This is done by briefly exploring some of the software engineering ideas and technologies now used in the information industry, using, as a case-study, a project with many similarities to those currently under way in HEP.

  6. Software inspections at Fermilab -- Use and experience

    International Nuclear Information System (INIS)

    Berman, E.F.

    1998-01-01

    Because of the critical nature of DA/Online software it is important to commission software which is correct, usable, reliable, and maintainable, i.e., has the highest quality possible. In order to help meet these goals Fermi National Accelerator Laboratory (Fermilab) has begun implementing a formal software inspection process. Formal Inspections are used to reduce the number of defects in software at as early a stage as possible. These Inspections, in use at a wide variety of institutions (e.g., NASA, Motorola), implement a well-defined procedure that can be used to improve the quality of many different types of deliverables. The inspection process, initially designed by Michael Fagan, will be described as it was developed and as it is currently implemented at Fermilab where it has been used to improve the quality of a variety of different experiment DA/Online software. Benefits of applying inspections at many points in the software life-cycle and benefits to the people involved will be investigated. Experience with many different types of Inspections and the lessons learned about the inspection process itself will be detailed. Finally, the future of Inspections at Fermilab will be given

  7. Six Sigma software development

    CERN Document Server

    Tayntor, Christine B

    2002-01-01

    Since Six Sigma has had marked success in improving quality in other settings, and since the quality of software remains poor, it seems a natural evolution to apply the concepts and tools of Six Sigma to system development and the IT department. Until now however, there were no books available that applied these concepts to the system development process. Six Sigma Software Development fills this void and illustrates how Six Sigma concepts can be applied to all aspects of the evolving system development process. It includes the traditional waterfall model and in the support of legacy systems,

  8. Pangasius in the EU market; Prospects for the position of (ASC-certified) pangasius in the EU retail and food service sector

    NARCIS (Netherlands)

    Beukers, R.; Pijl, van der W.; Duijn, van A.P.

    2014-01-01

    The objective of this market study is to investigate the potential short- and long-term benefits for investors to invest in the production of pangasius with a trademark based on ASC certification, for the retail and food service market segments in the EU market.

  9. Towards a New Paradigm of Software Development: an Ambassador Driven Process in Distributed Software Companies

    Science.gov (United States)

    Kumlander, Deniss

    The globalization of companies operations and competitor between software vendors demand improving quality of delivered software and decreasing the overall cost. The same in fact introduce a lot of problem into software development process as produce distributed organization breaking the co-location rule of modern software development methodologies. Here we propose a reformulation of the ambassador position increasing its productivity in order to bridge communication and workflow gap by managing the entire communication process rather than concentrating purely on the communication result.

  10. Exploration on Automated Software Requirement Document Readability Approaches

    OpenAIRE

    Chen, Mingda; He, Yao

    2017-01-01

    Context. The requirements analysis phase, as the very beginning of software development process, has been identified as a quite important phase in the software development lifecycle. Software Requirement Specification (SRS) is the output of requirements analysis phase, whose quality factors play an important role in the evaluation work. Readability is a quite important SRS quality factor, but there are few available automated approaches for readability measurement, because of the tight depend...

  11. Software Quality and Security in Teachers' and Students' Codes When Learning a New Programming Language

    Directory of Open Access Journals (Sweden)

    Arnon Hershkovitz

    2015-09-01

    Full Text Available In recent years, schools (as well as universities have added cyber security to their computer science curricula. This topic is still new for most of the current teachers, who would normally have a standard computer science background. Therefore the teachers are trained and then teaching their students what they have just learned. In order to explore differences in both populations’ learning, we compared measures of software quality and security between high-school teachers and students. We collected 109 source files, written in Python by 18 teachers and 31 students, and engineered 32 features, based on common standards for software quality (PEP 8 and security (derived from CERT Secure Coding Standards. We use a multi-view, data-driven approach, by (a using hierarchical clustering to bottom-up partition the population into groups based on their code-related features and (b building a decision tree model that predicts whether a student or a teacher wrote a given code (resulting with a LOOCV kappa of 0.751. Overall, our findings suggest that the teachers’ codes have a better quality than the students’ – with a sub-group of the teachers, mostly males, demonstrate better coding than their peers and the students – and that the students’ codes are slightly better secured than the teachers’ codes (although both populations show very low security levels. The findings imply that teachers might benefit from their prior knowledge and experience, but also emphasize the lack of continuous involvement of some of the teachers with code-writing. Therefore, findings shed light on computer science teachers as lifelong learners. Findings also highlight the difference between quality and security in today’s programming paradigms. Implications for these findings are discussed.

  12. Novel approaches to assess the quality of fertility data stored in dairy herd management software.

    Science.gov (United States)

    Hermans, K; Waegeman, W; Opsomer, G; Van Ranst, B; De Koster, J; Van Eetvelde, M; Hostens, M

    2017-05-01

    Scientific journals and popular press magazines are littered with articles in which the authors use data from dairy herd management software. Almost none of such papers include data cleaning and data quality assessment in their study design despite this being a very critical step during data mining. This paper presents 2 novel data cleaning methods that permit identification of animals with good and bad data quality. The first method is a deterministic or rule-based data cleaning method. Reproduction and mutation or life-changing events such as birth and death were converted to a symbolic (alphabetical letter) representation and split into triplets (3-letter code). The triplets were manually labeled as physiologically correct, suspicious, or impossible. The deterministic data cleaning method was applied to assess the quality of data stored in dairy herd management from 26 farms enrolled in the herd health management program from the Faculty of Veterinary Medicine Ghent University, Belgium. In total, 150,443 triplets were created, 65.4% were labeled as correct, 17.4% as suspicious, and 17.2% as impossible. The second method, a probabilistic method, uses a machine learning algorithm (random forests) to predict the correctness of fertility and mutation events in an early stage of data cleaning. The prediction accuracy of the random forests algorithm was compared with a classical linear statistical method (penalized logistic regression), outperforming the latter substantially, with a superior receiver operating characteristic curve and a higher accuracy (89 vs. 72%). From those results, we conclude that the triplet method can be used to assess the quality of reproduction data stored in dairy herd management software and that a machine learning technique such as random forests is capable of predicting the correctness of fertility data. Copyright © 2017 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.

  13. A Model for Quality Optimization in Software Design Processes

    NARCIS (Netherlands)

    Noppen, J.A.R.; van den Broek, P.M.; Aksit, Mehmet

    The main objective of software engineers is to design and implement systems that implement all functional and non-functional requirements. Unfortunately, it is very difficult or even generally impossible to deliver a software system that satisfies all the requirements. Even more seriously, failures

  14. Software architecture analysis of usability

    NARCIS (Netherlands)

    Folmer, Eelke

    2005-01-01

    One of the qualities that has received increased attention in recent decades is usability. A software product with poor usability is likely to fail in a highly competitive market; therefore software developing organizations are paying more and more attention to ensuring the usability of their

  15. Crowdsourcing cloud-based software development

    CERN Document Server

    Li, Wei; Tsai, Wei-Tek; Wu, Wenjun

    2015-01-01

    This book presents the latest research on the software crowdsourcing approach to develop large and complex software in a cloud-based platform. It develops the fundamental principles, management organization and processes, and a cloud-based infrastructure to support this new software development approach. The book examines a variety of issues in software crowdsourcing processes, including software quality, costs, diversity of solutions, and the competitive nature of crowdsourcing processes. Furthermore, the book outlines a research roadmap of this emerging field, including all the key technology and management issues for the foreseeable future. Crowdsourcing, as demonstrated by Wikipedia and Facebook for online web applications, has shown promising results for a variety of applications, including healthcare, business, gold mining exploration, education, and software development. Software crowdsourcing is emerging as a promising solution to designing, developing and maintaining software. Preliminary software cr...

  16. Formalizing the ISDF Software Development Methodology

    OpenAIRE

    Mihai Liviu DESPA

    2015-01-01

    The paper is aimed at depicting the ISDF software development methodology by emphasizing quality management and software development lifecycle. The ISDF methodology was built especially for innovative software development projects. The ISDF methodology was developed empirically by trial and error in the process of implementing multiple innovative projects. The research process began by analysing key concepts like innovation and software development and by settling the important dilemma of wha...

  17. How does agility ensure quality?

    OpenAIRE

    Huo, Ming; Verner, June; Ali Babar, Muhammad; Zhu, Liming

    2004-01-01

    peer-reviewed Software quality is one of our most important software concerns. Agile methods may produce software faster but we also need to know how they meet our quality requirements. In this paper we compare the waterfall model with agile processes to show how agile methods achieve software quality. We also show how agile methods attain quality under time pressure and in an unstable requirements environment, i.e. we analyze agile software quality assurance. We ...

  18. Service Level Agreements as Vehicles for Managing Acquisition of Software-Intensive Systems

    National Research Council Canada - National Science Library

    Gaines, Leonard T; Michael, James B

    2005-01-01

    ... to support quality and process control throughout the entire lifecycle of a software-intensive system. This article defines SLAs, discusses software quality and describes how SLAS can be utilized to incorporate requirements pertaining to product, process, project, and deployment quality throughout the software lifecycle.

  19. The FRISBEE tool, a software for optimising the trade-off between food quality, energy use, and global warming impact of cold chains

    NARCIS (Netherlands)

    Gwanpua, S.G.; Verboven, P.; Leducq, D.; Brown, T.; Verlinden, B.E.; Bekele, E.; Aregawi, W. Evans, J.; Foster, A.; Duret, S.; Hoang, H.M.; Sluis, S. van der; Wissink, E.; Hendriksen, L.J.A.M.; Taoukis, P.; Gogou, E.; Stahl, V.; El Jabri, M.; Le Page, J.F.; Claussen, I.; Indergård, E.; Nicolai, B.M.; Alvarez, G.; Geeraerd, A.H.

    2015-01-01

    Food quality (including safety) along the cold chain, energy use and global warming impact of refrigeration systems are three key aspects in assessing cold chain sustainability. In this paper, we present the framework of a dedicated software, the FRISBEE tool, for optimising quality of refrigerated

  20. Internal quality control indicators of cervical cytopathology exams performed in laboratories monitored by the External Quality Control Laboratory.

    Science.gov (United States)

    Ázara, Cinara Zago Silveira; Manrique, Edna Joana Cláudio; Tavares, Suelene Brito do Nascimento; de Souza, Nadja Lindany Alves; Amaral, Rita Goreti

    2014-09-01

    To evaluate the impact of continued education provided by an external quality control laboratory on the indicators of internal quality control of cytopathology exams. The internal quality assurance indicators for cytopathology exams from 12 laboratories monitored by the External Quality Control Laboratory were evaluated. Overall, 185,194 exams were included, 98,133 of which referred to the period preceding implementation of a continued education program, while 87,061 referred to the period following this intervention. Data were obtained from the Cervical Cancer Database of the Brazilian National Health Service. Following implementation of the continued education program, the positivity index (PI) remained within recommended limits in four laboratories. In another four laboratories, the PI progressed from below the limits to within the recommended standards. In one laboratory, the PI remained low, in two laboratories, it remained very low, and in one, it increased from very low to low. The percentage of exams compatible with a high-grade squamous intraepithelial lesion (HSIL) remained within the recommended limits in five laboratories, while in three laboratories it progressed from below the recommended levels to >0.4% of the total number of satisfactory exams, and in four laboratories it remained below the standard limit. Both the percentage of atypical squamous cells of undetermined significance (ASC-US) in relation to abnormal exams, and the ratio between ASC-US and intraepithelial lesions remained within recommended levels in all the laboratories investigated. An improvement was found in the indicators represented by the positivity index and the percentage of exams compatible with a high-grade squamous intraepithelial lesion, showing that the role played by the external quality control laboratory in providing continued education contributed towards improving laboratory staff skills in detecting cervical cancer precursor lesions.