WorldWideScience

Sample records for implementing software quality

  1. Implementation of a free software for quality control of IMRT

    International Nuclear Information System (INIS)

    Chinillace, N.; Alonso, S.; Cortina, T.; Reinado, D.; Ricos, B.; Diaz, S.; Campayo, J. M.

    2011-01-01

    In this paper we focus on implementation and launch of software that allows us to compare quantitatively the two-dimensional dose distributions calculated and measured experimentally in IMRT treatment. The tool we are using to make this comparison is the free software DoseLab. This is a program written in MatLab and open source, thereby allowing in some cases adapt the program to the needs of each user. This program will be able to calculate the gamma function of these distributions, a parameter that simultaneously evaluates the difference in dose between two pixels of the image and the distance between them, giving us an objective and quantitative, allowing us to decide if both distributions are compatible or not.

  2. An Approach for the Implementation of Software Quality Models Adpoting CERTICS and CMMI-DEV

    Directory of Open Access Journals (Sweden)

    GARCIA, F.W.

    2015-12-01

    Full Text Available This paper proposes a mapping between two product quality and software processes models used in the industry, the CERTICS national model and the CMMI-DEV international model. The stages of mapping are presented step by step, as well as the mapping review, which had the cooperation of one specialist in CERTICS and CMMI-DEV models. It aims to correlate the structures of the two models in order to facilitate and reduce the implementation time and costs, and to stimulate the execution of multi-model implementations in software developers companies.

  3. Software Quality Assurance Metrics

    Science.gov (United States)

    McRae, Kalindra A.

    2004-01-01

    Software Quality Assurance (SQA) is a planned and systematic set of activities that ensures conformance of software life cycle processes and products conform to requirements, standards and procedures. In software development, software quality means meeting requirements and a degree of excellence and refinement of a project or product. Software Quality is a set of attributes of a software product by which its quality is described and evaluated. The set of attributes includes functionality, reliability, usability, efficiency, maintainability, and portability. Software Metrics help us understand the technical process that is used to develop a product. The process is measured to improve it and the product is measured to increase quality throughout the life cycle of software. Software Metrics are measurements of the quality of software. Software is measured to indicate the quality of the product, to assess the productivity of the people who produce the product, to assess the benefits derived from new software engineering methods and tools, to form a baseline for estimation, and to help justify requests for new tools or additional training. Any part of the software development can be measured. If Software Metrics are implemented in software development, it can save time, money, and allow the organization to identify the caused of defects which have the greatest effect on software development. The summer of 2004, I worked with Cynthia Calhoun and Frank Robinson in the Software Assurance/Risk Management department. My task was to research and collect, compile, and analyze SQA Metrics that have been used in other projects that are not currently being used by the SA team and report them to the Software Assurance team to see if any metrics can be implemented in their software assurance life cycle process.

  4. Software quality assurance

    CERN Document Server

    Laporte, Claude Y

    2018-01-01

    This book introduces Software Quality Assurance (SQA) and provides an overview of standards used to implement SQA. It defines ways to assess the effectiveness of how one approaches software quality across key industry sectors such as telecommunications, transport, defense, and aerospace. * Includes supplementary website with an instructor's guide and solutions * Applies IEEE software standards as well as the Capability Maturity Model Integration for Development (CMMI) * Illustrates the application of software quality assurance practices through the use of practical examples, quotes from experts, and tips from the authors

  5. Software quality assurance handbook

    Energy Technology Data Exchange (ETDEWEB)

    1990-09-01

    There are two important reasons for Software Quality Assurance (SQA) at Allied-Signal Inc., Kansas City Division (KCD): First, the benefits from SQA make good business sense. Second, the Department of Energy has requested SQA. This handbook is one of the first steps in a plant-wide implementation of Software Quality Assurance at KCD. The handbook has two main purposes. The first is to provide information that you will need to perform software quality assurance activities. The second is to provide a common thread to unify the approach to SQA at KCD. 2 figs.

  6. SU-E-T-103: Development and Implementation of Web Based Quality Control Software

    International Nuclear Information System (INIS)

    Studinski, R; Taylor, R; Angers, C; La Russa, D; Clark, B

    2014-01-01

    Purpose: Historically many radiation medicine programs have maintained their Quality Control (QC) test results in paper records or Microsoft Excel worksheets. Both these approaches represent significant logistical challenges, and are not predisposed to data review and approval. It has been our group's aim to develop and implement web based software designed not just to record and store QC data in a centralized database, but to provide scheduling and data review tools to help manage a radiation therapy clinics Equipment Quality control program. Methods: The software was written in the Python programming language using the Django web framework. In order to promote collaboration and validation from other centres the code was made open source and is freely available to the public via an online source code repository. The code was written to provide a common user interface for data entry, formalize the review and approval process, and offer automated data trending and process control analysis of test results. Results: As of February 2014, our installation of QAtrack+ has 180 tests defined in its database and has collected ∼22 000 test results, all of which have been reviewed and approved by a physicist via QATrack+'s review tools. These results include records for quality control of Elekta accelerators, CT simulators, our brachytherapy programme, TomoTherapy and Cyberknife units. Currently at least 5 other centres are known to be running QAtrack+ clinically, forming the start of an international user community. Conclusion: QAtrack+ has proven to be an effective tool for collecting radiation therapy QC data, allowing for rapid review and trending of data for a wide variety of treatment units. As free and open source software, all source code, documentation and a bug tracker are available to the public at https://bitbucket.org/tohccmedphys/qatrackplus/

  7. SU-E-T-103: Development and Implementation of Web Based Quality Control Software

    Energy Technology Data Exchange (ETDEWEB)

    Studinski, R; Taylor, R; Angers, C; La Russa, D; Clark, B [The Ottawa Hospital Regional Cancer Ctr., Ottawa, ON (Canada)

    2014-06-01

    Purpose: Historically many radiation medicine programs have maintained their Quality Control (QC) test results in paper records or Microsoft Excel worksheets. Both these approaches represent significant logistical challenges, and are not predisposed to data review and approval. It has been our group's aim to develop and implement web based software designed not just to record and store QC data in a centralized database, but to provide scheduling and data review tools to help manage a radiation therapy clinics Equipment Quality control program. Methods: The software was written in the Python programming language using the Django web framework. In order to promote collaboration and validation from other centres the code was made open source and is freely available to the public via an online source code repository. The code was written to provide a common user interface for data entry, formalize the review and approval process, and offer automated data trending and process control analysis of test results. Results: As of February 2014, our installation of QAtrack+ has 180 tests defined in its database and has collected ∼22 000 test results, all of which have been reviewed and approved by a physicist via QATrack+'s review tools. These results include records for quality control of Elekta accelerators, CT simulators, our brachytherapy programme, TomoTherapy and Cyberknife units. Currently at least 5 other centres are known to be running QAtrack+ clinically, forming the start of an international user community. Conclusion: QAtrack+ has proven to be an effective tool for collecting radiation therapy QC data, allowing for rapid review and trending of data for a wide variety of treatment units. As free and open source software, all source code, documentation and a bug tracker are available to the public at https://bitbucket.org/tohccmedphys/qatrackplus/.

  8. Software component quality evaluation

    Science.gov (United States)

    Clough, A. J.

    1991-01-01

    The paper describes a software inspection process that can be used to evaluate the quality of software components. Quality criteria, process application, independent testing of the process and proposed associated tool support are covered. Early results indicate that this technique is well suited for assessing software component quality in a standardized fashion. With automated machine assistance to facilitate both the evaluation and selection of software components, such a technique should promote effective reuse of software components.

  9. A prolog implementation of pattern search to optimize software quality assurance

    OpenAIRE

    Buzzard, Raymond Karl

    1990-01-01

    Approved for public release, distribution is unlimited Quality Assurance (QA) is a critical factor in the development of successful software systems. Through the use of various QA tools, project managers can ensure that a desired level of performance and reliability is built into the system. However, these tools are not without cost. Project managers must weight all QA costs and benefits for each development environment before weigh all QA costs and benefits for each development environmen...

  10. Computer software quality assurance

    International Nuclear Information System (INIS)

    Ives, K.A.

    1986-06-01

    The author defines some criteria for the evaluation of software quality assurance elements for applicability to the regulation of the nuclear industry. The author then analyses a number of software quality assurance (SQA) standards. The major extracted SQA elements are then discussed, and finally specific software quality assurance recommendations are made for the nuclear industry

  11. Implementing Software Defined Radio

    CERN Document Server

    Grayver, Eugene

    2013-01-01

    Software Defined Radio makes wireless communications easier, more efficient, and more reliable. This book bridges the gap between academic research and practical implementation. When beginning a project, practicing engineers, technical managers, and graduate students can save countless hours by considering the concepts presented in these pages. The author covers the myriad options and trade-offs available when selecting an appropriate hardware architecture. As demonstrated here, the choice between hardware- and software-centric architecture can mean the difference between meeting an aggressive schedule and bogging down in endless design iterations. Because of the author’s experience overseeing dozens of failed and successful developments, he is able to present many real-life examples. Some of the key concepts covered are: Choosing the right architecture for the market – laboratory, military, or commercial Hardware platforms – FPGAs, GPPs, specialized and hybrid devices Standardization efforts to ens...

  12. Software to implement the IAEA-TECDOC 1517: quality control in mammography

    International Nuclear Information System (INIS)

    Mora Rodriguez, Patricia

    2011-01-01

    Mammography quality control is presented; also, guidelines for establishing a quality control program through tests to be performed, equipment, frequencies and tolerances. The work is intended for radiologists, medical physicists and technologists. Also examples of data collection sheets, minimum equipment list and overview of a quality control program are included. Two projects (RLA/9/057 and RLA/9/067) raised the goal of developing a software that allows a simple and effective way to automate the evidence contained in the TECDOC. The V1 is presented in the IRPA 12 (2008), which was tested in some countries to detect problems in their use and include possible improvements. However, the final revised version for distribution in member countries was get in 2010. The way of work on the project is described and who has worked. Also, a complete program, its importance and the projection of the Universidad de Costa Rica to work with the same. (author) [es

  13. Software product quality measurement

    OpenAIRE

    Godliauskas, Eimantas

    2016-01-01

    This paper analyses Ruby product quality measures, suggesting three new measures for Ruby product quality measurement tool Rubocop to measure Ruby product quality characteristics defined in ISO 2502n standard series. This paper consists of four main chapters. The first chapter gives a brief view of software product quality and software product quality measurement. The second chapter analyses object oriented quality measures. The third chapter gives a brief view of the most popular Ruby qualit...

  14. Software quality concepts and practice

    CERN Document Server

    Galin, Daniel

    2018-01-01

    The book presents a comprehensive discussion on software quality issues and software quality assurance (SQA) principles and practices, and lays special emphasis on implementing and managing SQA. Primarily designed to serve three audiences; universities and college students, vocational training participants, and software engineers and software development managers, the book may be applicable to all personnel engaged in a software projects Features: * A broad view of SQA. The book delves into SQA issues, going beyond the classic boundaries of custom-made software development to also cover in-house software development, subcontractors, and readymade software. * An up-to-date wide-range coverage of SQA and SQA related topics. Providing comprehensive coverage on multifarious SQA subjects, including topics, hardly explored till in SQA texts. * A systematic presentation of the SQA function and its tasks: establishing the SQA processes, planning, coordinating, follow-up, review and evaluation of SQA processes. * Fo...

  15. Software quality management

    International Nuclear Information System (INIS)

    Bishop, D.C.; Pymm, P.

    1991-01-01

    As programmable electronic (software-based) systems are increasingly being proposed as design solutions for high integrity applications in nuclear power stations, the need to adopt suitable quality management arrangements is paramount. The authors describe Scottish Nuclear's strategy for software quality management and, using the main on-line monitoring system at Torness Power Station as an example, explain how this strategy is put into practice. Particular attention is given to the topics of software quality planning and change control. (author)

  16. Software quality in 1997

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Inc., Burlington, MA (United States)

    1997-11-01

    For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Deployment (QFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels. Since software is on the critical path for many engineered products, and for internal business systems as well, the new approaches are starting to affect global competition and attract widespread international interest. It can be hypothesized that success in mastering software quality will be a key strategy for dominating global software markets in the 21st century.

  17. Sandia software guidelines: Software quality planning

    Energy Technology Data Exchange (ETDEWEB)

    1987-08-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standard for Software Quality Assurance Plans, this volume identifies procedures to follow in producing a Software Quality Assurance Plan for an organization or a project, and provides an example project SQA plan. 2 figs., 4 tabs.

  18. [Implementation of a web based software for documentation and control of quality of an acute pain service].

    Science.gov (United States)

    Pawlik, Michael T; Abel, Reinhard; Abt, Gregor; Kieninger, Martin; Graf, Bernhard Martin; Taeger, Kai; Ittner, Karl Peter

    2009-07-01

    Providing an acute pain service means accumulation of a large amount of data. The alleviation of data collection, improvement of data quality and data analysis plays a pivotal role. The electronic medical record (EMR) is gaining more and more importance in this context and is continuously spreading in clinical practice. Up to now only a few commercial softwares are available that specifically fit to the needs of an acute pain service. Here we report the development and implementation of such a program (Schmerzvisite, Medlinq, Hamburg, Germany) in the acute pain service of a University Hospital.

  19. Software product quality control

    CERN Document Server

    Wagner, Stefan

    2013-01-01

    Quality is not a fixed or universal property of software; it depends on the context and goals of its stakeholders. Hence, when you want to develop a high-quality software system, the first step must be a clear and precise specification of quality. Yet even if you get it right and complete, you can be sure that it will become invalid over time. So the only solution is continuous quality control: the steady and explicit evaluation of a product's properties with respect to its updated quality goals.This book guides you in setting up and running continuous quality control in your environment. Star

  20. Design and implementation of software for automated quality control and data analysis for a complex LC/MS/MS assay for urine opiates and metabolites.

    Science.gov (United States)

    Dickerson, Jane A; Schmeling, Michael; Hoofnagle, Andrew N; Hoffman, Noah G

    2013-01-16

    Mass spectrometry provides a powerful platform for performing quantitative, multiplexed assays in the clinical laboratory, but at the cost of increased complexity of analysis and quality assurance calculations compared to other methodologies. Here we describe the design and implementation of a software application that performs quality control calculations for a complex, multiplexed, mass spectrometric analysis of opioids and opioid metabolites. The development and implementation of this application improved our data analysis and quality assurance processes in several ways. First, use of the software significantly improved the procedural consistency for performing quality control calculations. Second, it reduced the amount of time technologists spent preparing and reviewing the data, saving on average over four hours per run, and in some cases improving turnaround time by a day. Third, it provides a mechanism for coupling procedural and software changes with the results of each analysis. We describe several key details of the implementation including the use of version control software and automated unit tests. These generally useful software engineering principles should be considered for any software development project in the clinical lab. Copyright © 2012 Elsevier B.V. All rights reserved.

  1. Software metrics to improve software quality in HEP

    International Nuclear Information System (INIS)

    Lancon, E.

    1996-01-01

    The ALEPH reconstruction program maintainability has been evaluated with a case tool implementing an ISO standard methodology based on software metrics. It has been found that the overall quality of the program is good and has shown improvement over the past five years. Frequently modified routines exhibits lower quality; most buys were located in routines with particularly low quality. Implementing from the beginning a quality criteria could have avoided time losses due to bug corrections. (author)

  2. Establishing software quality assurance

    International Nuclear Information System (INIS)

    Malsbury, J.

    1983-01-01

    This paper is concerned with four questions about establishing software QA: What is software QA. Why have software QA. What is the role of software QA. What is necessary to ensure the success of software QA

  3. Modernization of software quality assurance

    Science.gov (United States)

    Bhaumik, Gokul

    1988-01-01

    The customers satisfaction depends not only on functional performance, it also depends on the quality characteristics of the software products. An examination of this quality aspect of software products will provide a clear, well defined framework for quality assurance functions, which improve the life-cycle activities of software development. Software developers must be aware of the following aspects which have been expressed by many quality experts: quality cannot be added on; the level of quality built into a program is a function of the quality attributes employed during the development process; and finally, quality must be managed. These concepts have guided our development of the following definition for a Software Quality Assurance function: Software Quality Assurance is a formal, planned approach of actions designed to evaluate the degree of an identifiable set of quality attributes present in all software systems and their products. This paper is an explanation of how this definition was developed and how it is used.

  4. Roadmap for Peridynamic Software Implementation

    Energy Technology Data Exchange (ETDEWEB)

    Littlewood, David John [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-10-01

    The application of peridynamics for engineering analysis requires an efficient and robust software implementation. Key elements include processing of the discretization, the proximity search for identification of pairwise interactions, evaluation of the con- stitutive model, application of a bond-damage law, and contact modeling. Additional requirements may arise from the choice of time integration scheme, for example esti- mation of the maximum stable time step for explicit schemes, and construction of the tangent stiffness matrix for many implicit approaches. This report summaries progress to date on the software implementation of the peridynamic theory of solid mechanics. Discussion is focused on parallel implementation of the meshfree discretization scheme of Silling and Askari [33] in three dimensions, although much of the discussion applies to computational peridynamics in general.

  5. Software Quality Assurance in Software Projects: A Study of Pakistan

    OpenAIRE

    Faisal Shafique Butt; Sundus Shaukat; M. Wasif Nisar; Ehsan Ullah Munir; Muhammad Waseem; Kashif Ayyub

    2013-01-01

    Software quality is specific property which tells what kind of standard software should have. In a software project, quality is the key factor of success and decline of software related organization. Many researches have been done regarding software quality. Software related organization follows standards introduced by Capability Maturity Model Integration (CMMI) to achieve good quality software. Quality is divided into three main layers which are Software Quality Assurance (SQA), Software Qu...

  6. Methods of Software Quality Assurance under a Nuclear Quality Assurance Program

    International Nuclear Information System (INIS)

    Kim, Jang Yeol; Lee, Young Jun; Cha, Kyung Ho; Cheon, Se Woo; Lee, Jang Soo; Kwon, Kee Choon

    2005-01-01

    This paper addresses a substantial implementation of a software quality assurance under a nuclear quality assurance program. The relationship of the responsibility between a top-level nuclear quality assurance program such as ASME/NQA-1 and its lower level software quality assurance is described. Software quality assurance activities and software quality assurance procedures during the software development life cycle are also described

  7. Software quality testing process analysis

    OpenAIRE

    Mera Paz, Julián

    2016-01-01

    Introduction: This article is the result of reading, review, analysis of books, magazines and articles well known for their scientific and research quality, which have addressed the software quality testing process. The author, based on his work experience in software development companies, teaching and other areas, has compiled and selected information to argue and substantiate the importance of the software quality testing process. Methodology: the existing literature on the software qualit...

  8. Software as quality product

    International Nuclear Information System (INIS)

    Enders, A.

    1975-01-01

    In many discussions on the reliability of computer systems, software is presented as the weak link in the chain. The contribution attempts to identify the reasons for this situation as seen from the software development. The concepts correctness and reliability of programmes are explained as they are understood in the specialist discussion of today. Measures and methods are discussed which are particularly relevant as far as the obtaining of fault-free and reliable programmes is concerned. Conclusions are drawn for the user of software so that he is in the position to judge himself what can be justly expected frm the product software compared to other products. (orig./LH) [de

  9. Software Quality Assurance Audits Guidebooks

    Science.gov (United States)

    1990-01-01

    The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes that are used in software development. The Software Assurance Guidebook, NASA-GB-A201, issued in September, 1989, provides an overall picture of the NASA concepts and practices in software assurance. Second level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the second level Software Quality Assurance Audits Guidebook that describes software quality assurance audits in a way that is compatible with practices at NASA Centers.

  10. lessons and challenges from software quality assessment

    African Journals Online (AJOL)

    DJFLEX

    www.globaljournalseries.com, Email: info@globaljournalseries.com ... ASSESSMENT: THE CASE OF SPACE SYSTEMS SOFTWARE. ... KEYWORDS: Software, Software Quality ,Quality Standard, Characteristics, ... and communication, etc.

  11. Software Quality Assurance for Nuclear Safety Systems

    International Nuclear Information System (INIS)

    Sparkman, D R; Lagdon, R

    2004-01-01

    The US Department of Energy has undertaken an initiative to improve the quality of software used to design and operate their nuclear facilities across the United States. One aspect of this initiative is to revise or create new directives and guides associated with quality practices for the safety software in its nuclear facilities. Safety software includes the safety structures, systems, and components software and firmware, support software and design and analysis software used to ensure the safety of the facility. DOE nuclear facilities are unique when compared to commercial nuclear or other industrial activities in terms of the types and quantities of hazards that must be controlled to protect workers, public and the environment. Because of these differences, DOE must develop an approach to software quality assurance that ensures appropriate risk mitigation by developing a framework of requirements that accomplishes the following goals: (sm b ullet) Ensures the software processes developed to address nuclear safety in design, operation, construction and maintenance of its facilities are safe (sm b ullet) Considers the larger system that uses the software and its impacts (sm b ullet) Ensures that the software failures do not create unsafe conditions Software designers for nuclear systems and processes must reduce risks in software applications by incorporating processes that recognize, detect, and mitigate software failure in safety related systems. It must also ensure that fail safe modes and component testing are incorporated into software design. For nuclear facilities, the consideration of risk is not necessarily sufficient to ensure safety. Systematic evaluation, independent verification and system safety analysis must be considered for software design, implementation, and operation. The software industry primarily uses risk analysis to determine the appropriate level of rigor applied to software practices. This risk-based approach distinguishes safety

  12. Software quality: Process or people

    Science.gov (United States)

    Palmer, Regina; Labaugh, Modenna

    1993-01-01

    This paper will present data related to software development processes and personnel involvement from the perspective of software quality assurance. We examine eight years of data collected from six projects. Data collected varied by project but usually included defect and fault density with limited use of code metrics, schedule adherence, and budget growth information. The data are a blend of AFSCP 800-14 and suggested productivity measures in Software Metrics: A Practioner's Guide to Improved Product Development. A software quality assurance database tool, SQUID, was used to store and tabulate the data.

  13. Model of software quality

    OpenAIRE

    Valencia Ayala, Luz Estela; Villa Sánchez, Paula Andréa; Ocampo S., Carlos Alberto

    2009-01-01

    En un mercado globalizado donde las empresas deben innovar y mejorar continuamente para crecer y ser más competitivas, es necesario tener acceso a certificaciones de calidad internacionales que les den un respaldo y puedan mantenerse en este mercado. Las certificaciones de calidad en la industria del software ayudan a las empresas a ser más productivas disminuyendo costos y tiempo en sus desarrollos. Las empresas de desarrollo de software de nuestro país en su mayoría son micro y pequeñas...

  14. First experiences with the implementation of the European standard EN 62304 on medical device software for the quality assurance of a radiotherapy unit

    International Nuclear Information System (INIS)

    Höss, Angelika; Lampe, Christian; Panse, Ralf; Ackermann, Benjamin; Naumann, Jakob; Jäkel, Oliver

    2014-01-01

    According to the latest amendment of the Medical Device Directive standalone software qualifies as a medical device when intended by the manufacturer to be used for medical purposes. In this context, the EN 62304 standard is applicable which defines the life-cycle requirements for the development and maintenance of medical device software. A pilot project was launched to acquire skills in implementing this standard in a hospital-based environment (in-house manufacture). The EN 62304 standard outlines minimum requirements for each stage of the software life-cycle, defines the activities and tasks to be performed and scales documentation and testing according to its criticality. The required processes were established for the pre-existent decision-support software FlashDumpComparator (FDC) used during the quality assurance of treatment-relevant beam parameters. As the EN 62304 standard implicates compliance with the EN ISO 14971 standard on the application of risk management to medical devices, a risk analysis was carried out to identify potential hazards and reduce the associated risks to acceptable levels. The EN 62304 standard is difficult to implement without proper tools, thus open-source software was selected and integrated into a dedicated development platform. The control measures yielded by the risk analysis were independently implemented and verified, and a script-based test automation was retrofitted to reduce the associated test effort. After all documents facilitating the traceability of the specified requirements to the corresponding tests and of the control measures to the proof of execution were generated, the FDC was released as an accessory to the HIT facility. The implementation of the EN 62304 standard was time-consuming, and a learning curve had to be overcome during the first iterations of the associated processes, but many process descriptions and all software tools can be re-utilized in follow-up projects. It has been demonstrated that a

  15. First experiences with the implementation of the European standard EN 62304 on medical device software for the quality assurance of a radiotherapy unit.

    Science.gov (United States)

    Höss, Angelika; Lampe, Christian; Panse, Ralf; Ackermann, Benjamin; Naumann, Jakob; Jäkel, Oliver

    2014-03-21

    According to the latest amendment of the Medical Device Directive standalone software qualifies as a medical device when intended by the manufacturer to be used for medical purposes. In this context, the EN 62304 standard is applicable which defines the life-cycle requirements for the development and maintenance of medical device software. A pilot project was launched to acquire skills in implementing this standard in a hospital-based environment (in-house manufacture). The EN 62304 standard outlines minimum requirements for each stage of the software life-cycle, defines the activities and tasks to be performed and scales documentation and testing according to its criticality. The required processes were established for the pre-existent decision-support software FlashDumpComparator (FDC) used during the quality assurance of treatment-relevant beam parameters. As the EN 62304 standard implicates compliance with the EN ISO 14971 standard on the application of risk management to medical devices, a risk analysis was carried out to identify potential hazards and reduce the associated risks to acceptable levels. The EN 62304 standard is difficult to implement without proper tools, thus open-source software was selected and integrated into a dedicated development platform. The control measures yielded by the risk analysis were independently implemented and verified, and a script-based test automation was retrofitted to reduce the associated test effort. After all documents facilitating the traceability of the specified requirements to the corresponding tests and of the control measures to the proof of execution were generated, the FDC was released as an accessory to the HIT facility. The implementation of the EN 62304 standard was time-consuming, and a learning curve had to be overcome during the first iterations of the associated processes, but many process descriptions and all software tools can be re-utilized in follow-up projects. It has been demonstrated that a

  16. R D software quality assurance

    Energy Technology Data Exchange (ETDEWEB)

    Hood, F.C.

    1991-10-01

    Research software quality assurance (QA) requirements must be adequate to strengthen development or modification objectives, but flexible enough not to restrict creativity. Application guidelines are needed for the different kinds of research and development (R D) software activities to assure project objectives are achieved.

  17. Software cost/resource modeling: Software quality tradeoff measurement

    Science.gov (United States)

    Lawler, R. W.

    1980-01-01

    A conceptual framework for treating software quality from a total system perspective is developed. Examples are given to show how system quality objectives may be allocated to hardware and software; to illustrate trades among quality factors, both hardware and software, to achieve system performance objectives; and to illustrate the impact of certain design choices on software functionality.

  18. Software quality assurance | News

    Science.gov (United States)

    Financial Officer Finance Section Office of the Chief Operating Officer Facilities Engineering Services Accelerator Division Accelerator Physics Center Office of the Chief Safety Officer Environment, Safety, Health and Quality Section Office of the Chief Project Officer Office of Project Support Services Office of

  19. Software quality engineering a practitioner's approach

    CERN Document Server

    Suryn, Witold

    2014-01-01

    Software quality stems from two distinctive, but associated, topics in software engineering: software functional quality and software structural quality. Software Quality Engineering studies the tenets of both of these notions, which focus on the efficiency and value of a design, respectively. The text addresses engineering quality on both the application and system levels with attention to Information Systems and Embedded Systems as well as recent developments. Targeted at graduate engineering students and software quality specialists, the book analyzes the relationship between functionality

  20. Outsourcing Software quality

    OpenAIRE

    Kaur, Amanpreet

    2013-01-01

    The key factors which have led to a growing trend of outsourcing are:? Lack of expert-labor in some portions of the business process.? Availability of cheaper labor, whilst not comprising on the quality of output.? Ability and feasibility to concentrate on the other crucial business process.These factors have specifically contributed to most of the outsourced partnersacross different locations in the world. Expertise in communication capabilities,technical expertise and favorable financial pa...

  1. Continuous improvement of software quality

    International Nuclear Information System (INIS)

    Sivertsen, Terje

    1999-04-01

    The present report is the first Halden Work Report delivered from the OECD Halden Reactor Project's research activity on formal methods and software quality. Of particular concern in this activity is to reach a consensus between regulators, licensees and the nuclear industry on questions related to the effective, industrial use of formal methods. The report gives considerable attention to the importance of continuous improvement as a characteristic of a living software quality system, and to the need of providing a basis for software process/product quality integration. In particular, the report discusses these aspects from the perspectives of defect prevention, formal methods, Total Quality Management (TQM), and Bayesian Belief Nets. Another concern is to promote controlled experiments on the use of new methods, techniques, and tools. This is achieved partly by reviewing suggestions on the collection and experimental use of data, and by surveying a number of metrics believed to have some potential for comparison studies (author) (ml)

  2. Examining software complexity and quality for scientific software

    International Nuclear Information System (INIS)

    Kelly, D.; Shepard, T.

    2005-01-01

    Research has not found a simple relationship between software complexity and software quality, and particularly no relationship between commonly used software complexity metrics and the occurrence of software faults. A study with an example of scientific software from the nuclear power industry illustrates the importance of addressing cognitive complexity, the complexity related to understanding the intellectual content of the software. Simple practices such as aptly-named variables contributes more to high quality software than limiting code sizes. This paper examines the research into complexity and quality and reports on a longitudinal study using the example of nuclear software. (author)

  3. Linear accelerator quality assurance using EPIQA software

    International Nuclear Information System (INIS)

    Bozhikov, S.; Sokerov, H.; Tonev, A.; Ivanova, K.

    2012-01-01

    Unlike treatment with static fields, using a dynamic multileaf collimator (dMLC), there are significant dosimetric issues which must be assessed before dynamic therapy can be implemented. The advanced techniques require some additional commissioning and quality assurance tests. The results of standard quality assurance (QA) machine tests and commissioning tests for volume modulated arc therapy (VMAT) using electronic portal image device (EPID) and 'EPIQA' software are presented. (authors)

  4. Set up and programming of an ALICE Time-Of-Flight trigger facility and software implementation for its Quality Assurance (QA) during LHC Run 2

    CERN Document Server

    Toschi, Francesco

    2016-01-01

    The Cosmic and Topology Trigger Module (CTTM) is the main component of a trigger based on the ALICE TOF detector. Taking advantage of the TOF fast response, this VME board implements the trigger logic and delivers several L0 trigger outputs, used since Run 1, to provide cosmic triggers and rare triggers in pp, p+Pb and Pb+Pb data taking. Due to TOF DCS architectural change of the PCs controlling the CTTM (from 32 bits to 64 bits) it is mandatory to upgrade the software related to the CTTM including the code programming the FPGA firmware. A dedicated CTTM board will be installed in a CERN lab (Meyrin site), with the aim of recreating the electronics chain of the TOF trigger, to get a comfortable porting of the code to the 64 bit environment. The project proposed to the summer student is the setting up of the CTTM and the porting of the software. Moreover, in order to monitor the CTTM Trigger board during the real data taking, the implementation of a new Quality Assurance (QA) code is also crucial, together wit...

  5. The Effects of Development Team Skill on Software Product Quality

    Science.gov (United States)

    Beaver, Justin M.; Schiavone, Guy A.

    2006-01-01

    This paper provides an analysis of the effect of the skill/experience of the software development team on the quality of the final software product. A method for the assessment of software development team skill and experience is proposed, and was derived from a workforce management tool currently in use by the National Aeronautics and Space Administration. Using data from 26 smallscale software development projects, the team skill measures are correlated to 5 software product quality metrics from the ISO/IEC 9126 Software Engineering Product Quality standard. in the analysis of the results, development team skill is found to be a significant factor in the adequacy of the design and implementation. In addition, the results imply that inexperienced software developers are tasked with responsibilities ill-suited to their skill level, and thus have a significant adverse effect on the quality of the software product. Keywords: software quality, development skill, software metrics

  6. Software Quality Certification: identifying the real obstacles

    Directory of Open Access Journals (Sweden)

    Megan Baker

    1996-05-01

    Full Text Available A case study of software certification reveals the real difficulty of certifying quality beyond superficial assessment - readers are invited to form their own conclusions. AS 3563 Software Quality Management System is the Australian version of ISO 9001, developed specifically for the software industry. For many Australian software houses, gaining certification with AS 3563 is a priority since certification has become a prerequisite to doing business with government departments and major corporations. However, the process of achieving registration with this standard is a lengthy and resource intensive process, and may have little impact on actual software quality. This case study recounts the experience of the consulting arm of one of Australia's accounting firms in its quest for certification. By using a number of specific management strategies this company was able to successfully implement AS 3563 in less than half the time usually taken to achieve certification - a feat for which its management should be congratulated. However, because the focus of the project was on gaining certification, few internal benefits have been realised despite the successful implementation of the standard.

  7. Implementing Software Safety in the NASA Environment

    Science.gov (United States)

    Wetherholt, Martha S.; Radley, Charles F.

    1994-01-01

    Until recently, NASA did not consider allowing computers total control of flight systems. Human operators, via hardware, have constituted the ultimate safety control. In an attempt to reduce costs, NASA has come to rely more and more heavily on computers and software to control space missions. (For example. software is now planned to control most of the operational functions of the International Space Station.) Thus the need for systematic software safety programs has become crucial for mission success. Concurrent engineering principles dictate that safety should be designed into software up front, not tested into the software after the fact. 'Cost of Quality' studies have statistics and metrics to prove the value of building quality and safety into the development cycle. Unfortunately, most software engineers are not familiar with designing for safety, and most safety engineers are not software experts. Software written to specifications which have not been safety analyzed is a major source of computer related accidents. Safer software is achieved step by step throughout the system and software life cycle. It is a process that includes requirements definition, hazard analyses, formal software inspections, safety analyses, testing, and maintenance. The greatest emphasis is placed on clearly and completely defining system and software requirements, including safety and reliability requirements. Unfortunately, development and review of requirements are the weakest link in the process. While some of the more academic methods, e.g. mathematical models, may help bring about safer software, this paper proposes the use of currently approved software methodologies, and sound software and assurance practices to show how, to a large degree, safety can be designed into software from the start. NASA's approach today is to first conduct a preliminary system hazard analysis (PHA) during the concept and planning phase of a project. This determines the overall hazard potential of

  8. Implementing an open-access CASA software for the assessment of stallion sperm motility: Relationship with other sperm quality parameters.

    Science.gov (United States)

    Giaretta, Elisa; Munerato, Mauro; Yeste, Marc; Galeati, Giovanna; Spinaci, Marcella; Tamanini, Carlo; Mari, Gaetano; Bucci, Diego

    2017-01-01

    Setting an open-access computer assisted sperm analysis (CASA) may benefit the evaluation of motility in mammalian sperm, especially when economic constraints do not allow the use of a commercial system. There have been successful attempts to develop such a device in Zebra fish sperm and the system has been used in very few studies on mammalian spermatozoa. Against this background, the present study aimed at developing an open-access CASA system for mammalian sperm using the horse as a model and based upon the Image J software previously established for Zebra fish sperm. Along with determining the sperm progressive motility and other kinetic parameters (such as amplitude of lateral head displacement), the "results" window was adjusted to simplify subsequent statistical analyses. The path window was enriched with colored sperm trajectories on the basis of the subpopulation they belong to and a number that allowed the sperm track to be associated to the sperm motility data shown in the "results" window. Data obtained from the novel plugin (named as CASA_bgm) were compared with those of the commercial CASA Hamilton-Thorn IVOS Vers.12, through Bland Altman's plots. While the percentage of total and progressive motile sperm, VCL, VAP, VSL, LIN and STR and ALH were in agreement with those obtained with the commercial system, BCF significantly differed between the two systems probably due to their settings. Interestingly, a positive and significant correlation between the percentages of total motile sperm evaluated through CASA_bgm and those showing high mitochondrial membrane potential evaluated by JC-1 staining was found. In conclusion, CASA_bgm ImageJ plugin could be useful and reliable for stallion sperm motility analysis and it is our aim to apply this system to other mammalian species. Copyright © 2016 Elsevier B.V. All rights reserved.

  9. Software quality - how is it achieved?

    International Nuclear Information System (INIS)

    Straker, E.A.

    1986-01-01

    Although software quality can't be quantified, the tools and techniques to achieve high quality are available. As management stresses the need for definable software quality programs from vendors and subcontractors and provides the incentives for these programs, the quality of software will improve. EPRI could provide the leadership in establishing guidelines for a balanced software quality program and through workshops provide training to utility staff and management on the methods for evaluating the characteristics of quality software. With the more complex systems discussed at this workshop and particularly with the trend toward the use of artificial intelligence, the importance of quality software will grow dramatically

  10. Quality assurance for software important to safety

    International Nuclear Information System (INIS)

    2000-01-01

    Software applications play an increasingly relevant role in nuclear power plant systems. This is particularly true of software important to safety used in both: calculations for the design, testing and analysis of nuclear reactor systems (design, engineering and analysis software); and monitoring, control and safety functions as an integral part of the reactor systems (monitoring, control and safety system software). Computer technology is advancing at a fast pace, offering new possibilities in nuclear reactor design, construction, commissioning, operation, maintenance and decommissioning. These advances also present new issues which must be considered both by the utility and by the regulatory organization. Refurbishment of ageing instrumentation and control systems in nuclear power plants and new safety related application areas have emerged, with direct (e.g. interfaces with safety systems) and indirect (e.g. operator intervention) implications for safety. Currently, there exist several international standards and guides on quality assurance for software important to safety. However, none of the existing documents provides comprehensive guidance to the developer, manager and regulator during all phases of the software life-cycle. The present publication was developed taking into account the large amount of available documentation, the rapid development of software systems and the need for updated guidance on h ow to do it . It provides information and guidance for defining and implementing quality assurance programmes covering the entire life-cycle of software important to safety. Expected users are managers, performers and assessors from nuclear utilities, regulatory bodies, suppliers and technical support organizations involved with the development and use of software applied in nuclear power plants

  11. Software quality assurance plan for PORFLOW-3D

    International Nuclear Information System (INIS)

    Maheras, S.J.

    1993-03-01

    This plan describes the steps taken by the Idaho National Engineering Laboratory Subsurface and Environmental Modeling Unit personnel to implement software quality assurance procedures for the PORFLOW-3D computer code. PORFLOW-3D was used to conduct radiological performance assessments at the Savannah River Site. software quality assurance procedures for PORFLOW-3D include software acquisition, installation, testing, operation, maintenance, and retirement. Configuration control and quality assurance procedures are also included or referenced in this plan

  12. Software metrics: Software quality metrics for distributed systems. [reliability engineering

    Science.gov (United States)

    Post, J. V.

    1981-01-01

    Software quality metrics was extended to cover distributed computer systems. Emphasis is placed on studying embedded computer systems and on viewing them within a system life cycle. The hierarchy of quality factors, criteria, and metrics was maintained. New software quality factors were added, including survivability, expandability, and evolvability.

  13. Software process improvement, quality assurance and measurement

    NARCIS (Netherlands)

    Trienekens, J.J.M.; Kusters, R.J.; Balla, K.; Kontogiannis, K.; Zou, Y.; Di Penta, M.

    2006-01-01

    The aim of this workshop was to present and discuss emergent software quality improvement approaches, with an emphasis on practical applications. Different views on the improvement of software processes, software products, and their interrelations, have been addressed during the workshop.

  14. Achieving software quality by GQM measurement

    NARCIS (Netherlands)

    Solingen, van D.M.; Veenendaal, van E.P.W.M.; Veenendaal, van E.P.W.M.; McMullan, J.

    1997-01-01

    Software development is a discipline with specific management difficulties. Collecting relevant data during development is a way to overcome these difficulties. Such data collection for sqftware development is termed "Software Measurement". Software measurement is a powerful aid to quality

  15. Integrated modeling of software cost and quality

    International Nuclear Information System (INIS)

    Rone, K.Y.; Olson, K.M.

    1994-01-01

    In modeling the cost and quality of software systems, the relationship between cost and quality must be considered. This explicit relationship is dictated by the criticality of the software being developed. The balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and the developers with respect to the processes being employed

  16. Software Quality Assurance activities of ITER CODAC

    Energy Technology Data Exchange (ETDEWEB)

    Pande, Sopan, E-mail: sopan.pande@iter.org [ITER Organization, Route de Vinon sur Verdon, 13115 St Paul Lez Durance (France); DiMaio, Franck; Kim, Changseung; Kim, Joohan; Klotz, Wolf-Dieter; Makijarvi, Petri; Stepanov, Denis; Wallander, Anders [ITER Organization, Route de Vinon sur Verdon, 13115 St Paul Lez Durance (France)

    2013-10-15

    Highlights: ► Comprehensive and consistent software engineering and quality assurance of CODAC. ► Applicable to all CODAC software projects executed by ITER DAs and contractors. ► Configurable plans for cost effective application of SQA processes. ► CODAC software plans SQAP, SVVP, SDP, and SCMP. ► CODAC software processes based on IEEE 12207-2008. -- Abstract: Software as an integral part of the plant system I and C is crucial in the manufacturing and integrated operation of ITER plant systems. Software Quality Assurance is necessary to ensure the development and maintenance of consistently high quality I and C software throughout the lifetime of ITER. CODAC decided to follow IEEE 12207-2008 software lifecycle processes for Software Engineering and Software Quality Assurance. Software Development Plan, Software Configuration Management Plan and Software Verification and Validation Plan are the mainstay of Software Quality Assurance which is documented in the Software Quality Assurance Plan. This paper describes the Software Quality Assurance (SQA) activities performed by CODAC. The SQA includes development and maintenance of above plans, processes and resources. With the help of Verification and Validation Teams they gather evidence of process conformance and product conformance, and record process data for quality audits and perform process improvements.

  17. Software Quality Assurance activities of ITER CODAC

    International Nuclear Information System (INIS)

    Pande, Sopan; DiMaio, Franck; Kim, Changseung; Kim, Joohan; Klotz, Wolf-Dieter; Makijarvi, Petri; Stepanov, Denis; Wallander, Anders

    2013-01-01

    Highlights: ► Comprehensive and consistent software engineering and quality assurance of CODAC. ► Applicable to all CODAC software projects executed by ITER DAs and contractors. ► Configurable plans for cost effective application of SQA processes. ► CODAC software plans SQAP, SVVP, SDP, and SCMP. ► CODAC software processes based on IEEE 12207-2008. -- Abstract: Software as an integral part of the plant system I and C is crucial in the manufacturing and integrated operation of ITER plant systems. Software Quality Assurance is necessary to ensure the development and maintenance of consistently high quality I and C software throughout the lifetime of ITER. CODAC decided to follow IEEE 12207-2008 software lifecycle processes for Software Engineering and Software Quality Assurance. Software Development Plan, Software Configuration Management Plan and Software Verification and Validation Plan are the mainstay of Software Quality Assurance which is documented in the Software Quality Assurance Plan. This paper describes the Software Quality Assurance (SQA) activities performed by CODAC. The SQA includes development and maintenance of above plans, processes and resources. With the help of Verification and Validation Teams they gather evidence of process conformance and product conformance, and record process data for quality audits and perform process improvements

  18. Software Quality Improvement in the OMC Team

    CERN Document Server

    Maier, Viktor

    Physicists use self-written software as a tool to fulfill their tasks and often the developed software is used for several years or even decades. If a software product lives for a long time, it has to be changed and adapted to external influences. This implies that the source code has to be read, understood and modified. The same applies to the software of the Optics Measurements and Corrections (OMC) team at CERN. Their task is to track, analyze and correct the beams in the LHC and other accelerators. To solve this task, they revert to a self-written software base with more than 150,000 physical lines of code. The base is subject to continuous changes as well. Their software does its job and is effective, but runs regrettably not efficient because some parts of the source code are in a bad shape and has a low quality. The implementation could be faster and more memory efficient. In addition it is difficult to read and understand the code. Source code files and functions are too big and identifiers do not rev...

  19. SQIMSO: Quality Improvement for Small Software Organizations

    OpenAIRE

    Rabih Zeineddine; Nashat Mansour

    2005-01-01

    Software quality improvement process remains incomplete if it is not initiated and conducted through a wide improvement program that considers process quality improvement, product quality improvement and evolution of human resources. But, small software organizations are not capable of bearing the cost of establishing software process improvement programs. In this work, we propose a new software quality improvement model for small organizations, SQIMSO, based on three ...

  20. Evaluating Predictive Models of Software Quality

    Science.gov (United States)

    Ciaschini, V.; Canaparo, M.; Ronchieri, E.; Salomoni, D.

    2014-06-01

    Applications from High Energy Physics scientific community are constantly growing and implemented by a large number of developers. This implies a strong churn on the code and an associated risk of faults, which is unavoidable as long as the software undergoes active evolution. However, the necessities of production systems run counter to this. Stability and predictability are of paramount importance; in addition, a short turn-around time for the defect discovery-correction-deployment cycle is required. A way to reconcile these opposite foci is to use a software quality model to obtain an approximation of the risk before releasing a program to only deliver software with a risk lower than an agreed threshold. In this article we evaluated two quality predictive models to identify the operational risk and the quality of some software products. We applied these models to the development history of several EMI packages with intent to discover the risk factor of each product and compare it with its real history. We attempted to determine if the models reasonably maps reality for the applications under evaluation, and finally we concluded suggesting directions for further studies.

  1. Evaluating predictive models of software quality

    International Nuclear Information System (INIS)

    Ciaschini, V; Canaparo, M; Ronchieri, E; Salomoni, D

    2014-01-01

    Applications from High Energy Physics scientific community are constantly growing and implemented by a large number of developers. This implies a strong churn on the code and an associated risk of faults, which is unavoidable as long as the software undergoes active evolution. However, the necessities of production systems run counter to this. Stability and predictability are of paramount importance; in addition, a short turn-around time for the defect discovery-correction-deployment cycle is required. A way to reconcile these opposite foci is to use a software quality model to obtain an approximation of the risk before releasing a program to only deliver software with a risk lower than an agreed threshold. In this article we evaluated two quality predictive models to identify the operational risk and the quality of some software products. We applied these models to the development history of several EMI packages with intent to discover the risk factor of each product and compare it with its real history. We attempted to determine if the models reasonably maps reality for the applications under evaluation, and finally we concluded suggesting directions for further studies.

  2. The medical software quality deployment method.

    Science.gov (United States)

    Hallberg, N; Timpka, T; Eriksson, H

    1999-03-01

    The objective of this study was to develop a Quality Function Deployment (QFD) model for design of information systems in health-care environments. Consecutive blocked-subject case studies were conducted, based on action research methods. Starting with a QFD model for software development, a model for information system design, the Medical Software Quality Deployment (MSQD) model, was developed. The MSQD model was divided into the pre-study phase, in which the customer categories and their power to influence the design are determined; the data collection phase, in which the voice of customers (VoC) is identified by observations and interviews and quantified by Critical. Incident questionnaires; the need specification phase, where the VoC is specified into ranked customer needs; and the design phase where the customer needs are transformed stepwise to technical requirements and design attributes. QFD showed to be useful for integrating the values of different customer categories in software development for health-care settings. In the later design phases, other quality methods should be used for software implementation and testing.

  3. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan : ASC software quality engineering practices Version 3.0.

    Energy Technology Data Exchange (ETDEWEB)

    Turgeon, Jennifer L.; Minana, Molly A.; Hackney, Patricia; Pilch, Martin M.

    2009-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in the US Department of Energy/National Nuclear Security Agency (DOE/NNSA) Quality Criteria, Revision 10 (QC-1) as 'conformance to customer requirements and expectations'. This quality plan defines the SNL ASC Program software quality engineering (SQE) practices and provides a mapping of these practices to the SNL Corporate Process Requirement (CPR) 001.3.6; 'Corporate Software Engineering Excellence'. This plan also identifies ASC management's and the software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals. This SNL ASC Software Quality Plan establishes the signatories commitments to improving software products by applying cost-effective SQE practices. This plan enumerates the SQE practices that comprise the development of SNL ASC's software products and explains the project teams opportunities for tailoring and implementing the practices.

  4. SWiFT Software Quality Assurance Plan.

    Energy Technology Data Exchange (ETDEWEB)

    Berg, Jonathan Charles [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2016-01-01

    This document describes the software development practice areas and processes which contribute to the ability of SWiFT software developers to provide quality software. These processes are designed to satisfy the requirements set forth by the Sandia Software Quality Assurance Program (SSQAP). APPROVALS SWiFT Software Quality Assurance Plan (SAND2016-0765) approved by: Department Manager SWiFT Site Lead Dave Minster (6121) Date Jonathan White (6121) Date SWiFT Controls Engineer Jonathan Berg (6121) Date CHANGE HISTORY Issue Date Originator(s) Description A 2016/01/27 Jon Berg (06121) Initial release of the SWiFT Software Quality Assurance Plan

  5. Software quality assurance - seven years experience

    International Nuclear Information System (INIS)

    Malsbury, J.A.

    1987-01-01

    This paper describes seven years experience with software quality assurance at PPPL. It covers the early attempts of 1980 and 1981 to establish software quality assurance; the first attempt of 1982 to develop a complete software quality assurance plan; the significant modifications of this plan in 1985; and the future. In addition, the paper describes the role of the Quality Assurance organization within each plan. The scope of this paper is limited to discussions of the software development procedures used in the seven year period. Other software quality topics, such as configuration control or problem identification and resolution, are not discussed

  6. Swarming Robot Design, Construction and Software Implementation

    Science.gov (United States)

    Stolleis, Karl A.

    2014-01-01

    In this paper is presented an overview of the hardware design, construction overview, software design and software implementation for a small, low-cost robot to be used for swarming robot development. In addition to the work done on the robot, a full simulation of the robotic system was developed using Robot Operating System (ROS) and its associated simulation. The eventual use of the robots will be exploration of evolving behaviors via genetic algorithms and builds on the work done at the University of New Mexico Biological Computation Lab.

  7. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1: ASC software quality engineering practices, Version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  8. Software quality metrics aggregation in industry

    NARCIS (Netherlands)

    Mordal, K.; Anquetil, N.; Laval, J.; Serebrenik, A.; Vasilescu, B.N.; Ducasse, S.

    2013-01-01

    With the growing need for quality assessment of entire software systems in the industry, new issues are emerging. First, because most software quality metrics are defined at the level of individual software components, there is a need for aggregation methods to summarize the results at the system

  9. On the Role of Software Quality Management in Software Process Improvement

    DEFF Research Database (Denmark)

    Wiedemann Jacobsen, Jan; Kuhrmann, Marco; Münch, Jürgen

    2016-01-01

    Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. SPI addresses many aspects ranging from individual developer skills to entire organizations. It comprises, for instance, the optimization of specific activities...... and a strong focus on custom review, testing, and documentation techniques, whereas a set of five selected improvement measures is almost equally addressed....

  10. Software quality assurance plans for safety-critical software

    International Nuclear Information System (INIS)

    Liddle, P.

    2006-01-01

    Application software is defined as safety-critical if a fault in the software could prevent the system components from performing their nuclear-safety functions. Therefore, for nuclear-safety systems, the AREVA TELEPERM R XS (TXS) system is classified 1E, as defined in the Inst. of Electrical and Electronics Engineers (IEEE) Std 603-1998. The application software is classified as Software Integrity Level (SIL)-4, as defined in IEEE Std 7-4.3.2-2003. The AREVA NP Inc. Software Program Manual (SPM) describes the measures taken to ensure that the TELEPERM XS application software attains a level of quality commensurate with its importance to safety. The manual also describes how TELEPERM XS correctly performs the required safety functions and conforms to established technical and documentation requirements, conventions, rules, and standards. The program manual covers the requirements definition, detailed design, integration, and test phases for the TELEPERM XS application software, and supporting software created by AREVA NP Inc. The SPM is required for all safety-related TELEPERM XS system applications. The program comprises several basic plans and practices: 1. A Software Quality-Assurance Plan (SQAP) that describes the processes necessary to ensure that the software attains a level of quality commensurate with its importance to safety function. 2. A Software Safety Plan (SSP) that identifies the process to reasonably ensure that safety-critical software performs as intended during all abnormal conditions and events, and does not introduce any new hazards that could jeopardize the health and safety of the public. 3. A Software Verification and Validation (V and V) Plan that describes the method of ensuring the software is in accordance with the requirements. 4. A Software Configuration Management Plan (SCMP) that describes the method of maintaining the software in an identifiable state at all times. 5. A Software Operations and Maintenance Plan (SO and MP) that

  11. Engineering high quality medical software

    CERN Document Server

    Coronato, Antonio

    2018-01-01

    This book focuses on high-confidence medical software in the growing field of e-health, telecare services and health technology. It covers the development of methodologies and engineering tasks together with standards and regulations for medical software.

  12. A SYSTEMATIC STUDY OF SOFTWARE QUALITY MODELS

    OpenAIRE

    Dr.Vilas. M. Thakare; Ashwin B. Tomar

    2011-01-01

    This paper aims to provide a basis for software quality model research, through a systematic study ofpapers. It identifies nearly seventy software quality research papers from journals and classifies paper asper research topic, estimation approach, study context and data set. The paper results combined withother knowledge provides support for recommendations in future software quality model research, toincrease the area of search for relevant studies, carefully select the papers within a set ...

  13. Implementation of a free software for quality control of IMRT; Puesta en marcha de un soltware de libre distribucion para el control de calidad IMRT

    Energy Technology Data Exchange (ETDEWEB)

    Chinillace, N.; Alonso, S.; Cortina, T.; Reinado, D.; Ricos, B.; Diaz, S.; Campayo, J. M.

    2011-07-01

    In this paper we focus on implementation and launch of software that allows us to compare quantitatively the two-dimensional dose distributions calculated and measured experimentally in IMRT treatment. The tool we are using to make this comparison is the free software DoseLab. This is a program written in MatLab and open source, thereby allowing in some cases adapt the program to the needs of each user. This program will be able to calculate the gamma function of these distributions, a parameter that simultaneously evaluates the difference in dose between two pixels of the image and the distance between them, giving us an objective and quantitative, allowing us to decide if both distributions are compatible or not.

  14. Software quality assurance plan for GCS

    Science.gov (United States)

    Duncan, Stephen E.; Bailey, Elizabeth K.

    1990-01-01

    The software quality assurance (SQA) function for the Guidance and Control Software (GCS) project which is part of a software error studies research program is described. The SQA plan outlines all of the procedures, controls, and audits to be carried out by the SQA organization to ensure adherence to the policies, procedures, and standards for the GCS project.

  15. Software Quality in the Objectory Software Development Process

    NARCIS (Netherlands)

    van den Berg, Klaas; Demeyer, S.; Bosch, J.

    In this paper we discuss how software quality assurance is realized in Rational Objectory. Although much support is given through guidelines and checkpoints, the tool fails to provide clear goals and metrics for quality assessments and it only partially supports the phases in a measurement program.

  16. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Fishler, B

    2011-03-18

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  17. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    International Nuclear Information System (INIS)

    Fishler, B.

    2011-01-01

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  18. CHALLENGES OF SOFTWARE QUALITY ASSURANCE AND TESTING

    Directory of Open Access Journals (Sweden)

    Md.Shahadat Hossain

    2018-02-01

    Full Text Available Uncertainty exists in Software Company over the world. Software quality problem is leading issue for the software industry. The issue exists from 40 years or 50 years long. The industry is suffering and closing for this issue. In this circumstance, it is important to address and remove its root cause. Otherwise, day by day industry economic loss will increase. I figure out some vital challenges of software quality assurance and testing which have been facing by software industries. The research focused on several small and medium software companies of the world. This paper represents different category of challenges along with responsible stakeholders. This research finds out that testing tools are available testing elements are available testing process has improved but still software has some testing challenges. My research figured out the bottleneck of challenges and explained in this paper. Here software engineers have scope to improve & overcome those challenges. This paper suggests systematic approach to solve the problem.

  19. Quality assurance of EDP software in practical application

    International Nuclear Information System (INIS)

    Winkler, H.

    1982-01-01

    Alongside the specific properties of the soft software, it is mainly points outside the traditional testing field which apply for the quality assurance thereof. Measures for quality assurance must in particular, start in the development. This presupposes a partial-result orientated development process of software. Due to the high qualitative demands, implements for testing and inspection are of great importance. The problems in software quality assurance are typical for a young technical field where the necessity of which is indisputed, but which has to effect on an empirical-pragmatical level still, due to insufficient scientific foundation. (orig.) [de

  20. Requirements engineering: foundation for software quality

    NARCIS (Netherlands)

    Daneva, Maia; Pastor, Oscar

    2016-01-01

    Welcome to the proceedings of the 22nd edition of REFSQ: the International Working Conference on Requirements Engineering – Foundation for Software Quality! Requirements engineering (RE) has been recognized as a critical factor that impacts the quality of software, systems, and services. Since the

  1. Assessing Software Quality Through Visualised Cohesion Metrics

    Directory of Open Access Journals (Sweden)

    Timothy Shih

    2001-05-01

    Full Text Available Cohesion is one of the most important factors for software quality as well as maintainability, reliability and reusability. Module cohesion is defined as a quality attribute that seeks for measuring the singleness of the purpose of a module. The module of poor quality can be a serious obstacle to the system quality. In order to design a good software quality, software managers and engineers need to introduce cohesion metrics to measure and produce desirable software. A highly cohesion software is thought to be a desirable constructing. In this paper, we propose a function-oriented cohesion metrics based on the analysis of live variables, live span and the visualization of processing element dependency graph. We give six typical cohesion examples to be measured as our experiments and justification. Therefore, a well-defined, well-normalized, well-visualized and well-experimented cohesion metrics is proposed to indicate and thus enhance software cohesion strength. Furthermore, this cohesion metrics can be easily incorporated with software CASE tool to help software engineers to improve software quality.

  2. Helping organizations to address their effort toward the implementation of improvements in their software process

    OpenAIRE

    Muñoz-Mata, Mirna Ariadna; Mejia-Miranda, Jezreel; Valtierra-Alvarado, Claudia

    2015-01-01

    Due to the importance of Software Development Small and Medium Enterprises (SMEs) in the software industry, it is necessary to guarantee the quality of their products. In this context, the implementation of software process improvements offers an attractive way to achieve it. Unfortunately, the lack of knowledge on how to address the improvement effort makes the implementation of software improvements in SMEs a path full of obstacles, and most of the times impossible to achieve. In order to h...

  3. Proceedings of the Fifth Triennial Software Quality Forum 2000, Software for the Next Millennium, Software Quality Forum

    Energy Technology Data Exchange (ETDEWEB)

    Scientific Software Engineering Group, CIC-12

    2000-04-01

    The Software Quality Forum is a triennial conference held by the Software Quality Assurance Subcommittee for the Department of Energy's Quality Managers. The forum centers on key issues, information, and technology important in software development for the Nuclear Weapons Complex. This year it will be opened up to include local information technology companies and software vendors presenting their solutions, ideas, and lessons learned. The Software Quality Forum 2000 will take on a more hands-on, instructional tone than those previously held. There will be an emphasis on providing information, tools, and resources to assist developers in their goal of producing next generation software.

  4. Development and application of new quality model for software projects.

    Science.gov (United States)

    Karnavel, K; Dillibabu, R

    2014-01-01

    The IT industry tries to employ a number of models to identify the defects in the construction of software projects. In this paper, we present COQUALMO and its limitations and aim to increase the quality without increasing the cost and time. The computation time, cost, and effort to predict the residual defects are very high; this was overcome by developing an appropriate new quality model named the software testing defect corrective model (STDCM). The STDCM was used to estimate the number of remaining residual defects in the software product; a few assumptions and the detailed steps of the STDCM are highlighted. The application of the STDCM is explored in software projects. The implementation of the model is validated using statistical inference, which shows there is a significant improvement in the quality of the software projects.

  5. Software for Optimizing Quality Assurance of Other Software

    Science.gov (United States)

    Feather, Martin; Cornford, Steven; Menzies, Tim

    2004-01-01

    Software assurance is the planned and systematic set of activities that ensures that software processes and products conform to requirements, standards, and procedures. Examples of such activities are the following: code inspections, unit tests, design reviews, performance analyses, construction of traceability matrices, etc. In practice, software development projects have only limited resources (e.g., schedule, budget, and availability of personnel) to cover the entire development effort, of which assurance is but a part. Projects must therefore select judiciously from among the possible assurance activities. At its heart, this can be viewed as an optimization problem; namely, to determine the allocation of limited resources (time, money, and personnel) to minimize risk or, alternatively, to minimize the resources needed to reduce risk to an acceptable level. The end result of the work reported here is a means to optimize quality-assurance processes used in developing software.

  6. Software Configuration Management: The Quality Weakness

    International Nuclear Information System (INIS)

    Arrojo, E.; Garcia, P.

    1998-01-01

    At the moment it is very difficult to din any process in the industry where software is not involved. We trust software does minimize the possibility of process failures. In parallel, the quality and safety requirements of our processes have been improved to satisfactory levels. Let's look around us. Every day, thousands of calculations are carried out by our engineers using computer programs. Hundreds of processes are controlled automatically. Safety marging, limits, operation controls..., are derived from them. The tools begin to control our processes but, Who does control the tool? Once they have been installed and once they are running, are they always reliable? NO If you think that your current system are satisfactory, we propose you a game in this report. It is just a test. Which is your score?. Then we revise the concept of Configuration Management and we describe an ideal machine; the ''Perpetuum Mobile'' of the Configuration. We describe some rules to implement and improvement and we comment on the operative experience in ENUSA. (Author)

  7. Experiences with Software Quality Metrics in the EMI middlewate

    OpenAIRE

    Alandes, M; Kenny, E M; Meneses, D; Pucciani, G

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristi...

  8. Design and implementation of embedded Bluetooth software system

    Science.gov (United States)

    Zhou, Zhijian; Zhou, Shujie; Xu, Huimin

    2001-10-01

    This thesis introduces the background knowledge and characteristics of Bluetooth technology. Then it summarizes the architecture and working principle of Bluetooth software. After carefully studying the characteristics of embedded operating system and Bluetooth software, this thesis declared two sets of module about Bluetooth software. Corresponding to these module's characteristics, this thesis introduces the design and implementation of LAN Access and Bluetooth headset. The Headset part introduces a developing method corresponding to the particularity of Bluetooth control software. Although these control software are application entity, the control signaling exchanged between them are regulations according to former definitions and they functions through the interaction of data and control information. These data and control information construct the protocol data unit (PDU), and the former definition can be seen as protocol in fact. This thesis uses the advanced development flow on communication protocol development as reference, a formal method - SDL (Specification and Description Language) - describing, validating and coding manually to C. This method not only reserved the efficiency of manually coded code, but also it ensures the quality of codes. The introduction also involves finite state machine theory while introduces the practical developing method on protocol development with the aid of SDL.

  9. Implementing Large Projects in Software Engineering Courses

    Science.gov (United States)

    Coppit, David

    2006-01-01

    In software engineering education, large projects are widely recognized as a useful way of exposing students to the real-world difficulties of team software development. But large projects are difficult to put into practice. First, educators rarely have additional time to manage software projects. Second, classrooms have inherent limitations that…

  10. Software Power Metric Model: An Implementation | Akwukwuma ...

    African Journals Online (AJOL)

    ... and the execution time (TIME) in each case was recorded. We then obtain the application functions point count. Our result shows that the proposed metric is computable, consistent in its use of unit, and is programming language independent. Keywords: Software attributes, Software power, measurement, Software metric, ...

  11. 12: Assuring the quality of critical software

    International Nuclear Information System (INIS)

    Jacky, J.; Kalet, I.

    1987-01-01

    The authors recommend quality assurance procedures for radiation therapy software. Software quality assurance deals with preventing, detecting and repairing programming errors. Error detection difficulties are most severe in computer-based control systems, for example therapy machine control systems, because it may be impossible for users to confirm correct operation while treatments are in progress, or to intervene if things go wrong. Software quality assurance techniques observed in other industries in which public safety is at risk are reviewed. In some of these industries software must be approved or certified before it can be used. Approval is subject to technical reviews and audits by experts other than the program authors. The main obstacles to adoption of these techniques in the radiation therapy field are costs, lack of familiarity and doubts regarding efficacy. 18 refs

  12. A software perspective of environmental data quality

    International Nuclear Information System (INIS)

    Banerjee, B.

    1995-01-01

    Because of the large amount of complex data in environmental projects, particularly large decontamination and decommissioning projects, the quality of the data has a profound impact on the success and cost of the mission. In every phase of the life cycle of the project, including regulatory intervention and legal proceedings, maintaining the quality of data and presenting data in a timely and meaningful manner are critical. In this paper, a systemic view of data quality management from a software engineering perspective is presented. A method of evaluation evolves from this view. This method complements the principles of the data quality objective. When graded adequately, the method of evaluation establishes a paradigm for ensuring data quality for new and renewed projects. This paper also demonstrates that incorporating good practices of software engineering into the data management process leads to continuous improvement of data quality

  13. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  14. The software quality control for gamma spectrometry

    International Nuclear Information System (INIS)

    Monte, L.

    1986-01-01

    One of major problems with wich the quality control program of an environmental measurements laboratory is confronted is the evaluation of the performances of software packages for the analysis of gamma-ray spectra. A program of tests for evaluating the performances of the software package (SPECTRAN-F, Canberra Inc.) used by our laboratory is being carried out. In this first paper the results of a preliminary study concerning the evaluation of the performance of the doublet analysis routine are presented

  15. Spectrum analysis on quality requirements consideration in software design documents.

    Science.gov (United States)

    Kaiya, Haruhiko; Umemura, Masahiro; Ogata, Shinpei; Kaijiri, Kenji

    2013-12-01

    Software quality requirements defined in the requirements analysis stage should be implemented in the final products, such as source codes and system deployment. To guarantee this meta-requirement, quality requirements should be considered in the intermediate stages, such as the design stage or the architectural definition stage. We propose a novel method for checking whether quality requirements are considered in the design stage. In this method, a technique called "spectrum analysis for quality requirements" is applied not only to requirements specifications but also to design documents. The technique enables us to derive the spectrum of a document, and quality requirements considerations in the document are numerically represented in the spectrum. We can thus objectively identify whether the considerations of quality requirements in a requirements document are adapted to its design document. To validate the method, we applied it to commercial software systems with the help of a supporting tool, and we confirmed that the method worked well.

  16. A Study on the Software Quality Assurance Plan

    International Nuclear Information System (INIS)

    Kim, Hyun Tae

    2006-01-01

    On 25 August 2006, the CMMI V1.2 (Capability Maturity Model Integration Version 1.2) was released with the new title CMMI-DEV (CMMI for Development) which supersedes the CMMI-SE/SW (CMMI for systems engineering and software engineering) V1.1. This study discusses the application of IEEE Std 730-2002, IEEE Standard for Software Quality Assurance Plans, for the implementation of the Process and Product Quality Assurance (PPQA) process area (PA) of the CMMI-DEV

  17. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan part 2 mappings for the ASC software quality engineering practices, version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR001.3.2 and CPR001.3.6 and to a Department of Energy document, ''ASCI Software Quality Engineering: Goals, Principles, and Guidelines''. This document also identifies ASC management and software project teams' responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  18. System Quality Management in Software Testing Laboratory that Chooses Accreditation

    Directory of Open Access Journals (Sweden)

    Yanet Brito R.

    2013-12-01

    Full Text Available The evaluation of software products will reach full maturity when executed by the scheme and provides third party certification. For the validity of the certification, the independent laboratory must be accredited for that function, using internationally recognized standards. This brings with it a challenge for the Industrial Laboratory Testing Software (LIPS, responsible for testing the products developed in Cuban Software Industry, define strategies that will permit it to offer services with a high level of quality. Therefore it is necessary to establish a system of quality management according to NC-ISO/IEC 17025: 2006 to continuously improve the operational capacity and technical competence of the laboratory, with a view to future accreditation of tests performed. This article discusses the process defined in the LIPS for the implementation of a Management System of Quality, from the current standards and trends, as a necessary step to opt for the accreditation of the tests performed.

  19. A Software Reuse Approach and Its Effect On Software Quality, An Empirical Study for The Software Industry

    OpenAIRE

    Mateen, Ahmed; Kausar, Samina; Sattar, Ahsan Raza

    2017-01-01

    Software reusability has become much interesting because of increased quality and reduce cost. A good process of software reuse leads to enhance the reliability, productivity, quality and the reduction of time and cost. Current reuse techniques focuses on the reuse of software artifact which grounded on anticipated functionality whereas, the non-functional (quality) aspect are also important. So, Software reusability used here to expand quality and productivity of software. It improves overal...

  20. MCNP trademark Software Quality Assurance plan

    International Nuclear Information System (INIS)

    Abhold, H.M.; Hendricks, J.S.

    1996-04-01

    MCNP is a computer code that models the interaction of radiation with matter. MCNP is developed and maintained by the Transport Methods Group (XTM) of the Los Alamos National Laboratory (LANL). This plan describes the Software Quality Assurance (SQA) program applied to the code. The SQA program is consistent with the requirements of IEEE-730.1 and the guiding principles of ISO 900

  1. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1 : ASC software quality engineering practices version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Minana, Molly A.; Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in DOE/AL Quality Criteria (QC-1) as conformance to customer requirements and expectations. This quality plan defines the ASC program software quality practices and provides mappings of these practices to the SNL Corporate Process Requirements (CPR 1.3.2 and CPR 1.3.6) and the Department of Energy (DOE) document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines (GP&G). This quality plan identifies ASC management and software project teams' responsibilities for cost-effective software engineering quality practices. The SNL ASC Software Quality Plan establishes the signatories commitment to improving software products by applying cost-effective software engineering quality practices. This document explains the project teams opportunities for tailoring and implementing the practices; enumerates the practices that compose the development of SNL ASC's software products; and includes a sample assessment checklist that was developed based upon the practices in this document.

  2. Software metrics: The key to quality software on the NCC project

    Science.gov (United States)

    Burns, Patricia J.

    1993-01-01

    Network Control Center (NCC) Project metrics are captured during the implementation and testing phases of the NCCDS software development lifecycle. The metrics data collection and reporting function has interfaces with all elements of the NCC project. Close collaboration with all project elements has resulted in the development of a defined and repeatable set of metrics processes. The resulting data are used to plan and monitor release activities on a weekly basis. The use of graphical outputs facilitates the interpretation of progress and status. The successful application of metrics throughout the NCC project has been instrumental in the delivery of quality software. The use of metrics on the NCC Project supports the needs of the technical and managerial staff. This paper describes the project, the functions supported by metrics, the data that are collected and reported, how the data are used, and the improvements in the quality of deliverable software since the metrics processes and products have been in use.

  3. Software Quality Perceptions of Stakeholders Involved in the Software Development Process

    Science.gov (United States)

    Padmanabhan, Priya

    2013-01-01

    Software quality is one of the primary determinants of project management success. Stakeholders involved in software development widely agree that quality is important (Barney and Wohlin 2009). However, they may differ on what constitutes software quality, and which of its attributes are more important than others. Although, software quality…

  4. Software Quality Control at Belle II

    Science.gov (United States)

    Ritter, M.; Kuhr, T.; Hauth, T.; Gebard, T.; Kristof, M.; Pulvermacher, C.; Belle Software Group, II

    2017-10-01

    Over the last seven years the software stack of the next generation B factory experiment Belle II has grown to over one million lines of C++ and Python code, counting only the part included in offline software releases. There are several thousand commits to the central repository by about 100 individual developers per year. To keep a coherent software stack of high quality that it can be sustained and used efficiently for data acquisition, simulation, reconstruction, and analysis over the lifetime of the Belle II experiment is a challenge. A set of tools is employed to monitor the quality of the software and provide fast feedback to the developers. They are integrated in a machinery that is controlled by a buildbot master and automates the quality checks. The tools include different compilers, cppcheck, the clang static analyzer, valgrind memcheck, doxygen, a geometry overlap checker, a check for missing or extra library links, unit tests, steering file level tests, a sophisticated high-level validation suite, and an issue tracker. The technological development infrastructure is complemented by organizational means to coordinate the development.

  5. Measuring health care process quality with software quality measures.

    Science.gov (United States)

    Yildiz, Ozkan; Demirörs, Onur

    2012-01-01

    Existing quality models focus on some specific diseases, clinics or clinical areas. Although they contain structure, process, or output type measures, there is no model which measures quality of health care processes comprehensively. In addition, due to the not measured overall process quality, hospitals cannot compare quality of processes internally and externally. To bring a solution to above problems, a new model is developed from software quality measures. We have adopted the ISO/IEC 9126 software quality standard for health care processes. Then, JCIAS (Joint Commission International Accreditation Standards for Hospitals) measurable elements were added to model scope for unifying functional requirements. Assessment (diagnosing) process measurement results are provided in this paper. After the application, it was concluded that the model determines weak and strong aspects of the processes, gives a more detailed picture for the process quality, and provides quantifiable information to hospitals to compare their processes with multiple organizations.

  6. IMPROVING (SOFTWARE) PATENT QUALITY THROUGH THE ADMINISTRATIVE PROCESS.

    Science.gov (United States)

    Rai, Arti K

    2013-11-24

    The available evidence indicates that patent quality, particularly in the area of software, needs improvement. This Article argues that even an agency as institutionally constrained as the U.S. Patent and Trademark Office ("PTO") could implement a portfolio of pragmatic, cost-effective quality improvement strategies. The argument in favor of these strategies draws upon not only legal theory and doctrine but also new data from a PTO software examination unit with relatively strict practices. Strategies that resolve around Section 112 of the patent statute could usefully be deployed at the initial examination stage. Other strategies could be deployed within the new post-issuance procedures available to the agency under the America Invents Act. Notably, although the strategies the Article discusses have the virtue of being neutral as to technology, they are likely to have a very significant practical impact in the area of software.

  7. Software quality assurance plan for viscometer

    International Nuclear Information System (INIS)

    Gimera, M.

    1994-01-01

    The in situ viscometer is a portable instrument designed to raise and lower a sphere (rheometer ball) through layers of tank waste material while recording ball position, velocity, and cable tension. In the field, the viscometer attaches to a decontamination spool piece which in turn is designed to attach to any 4-inch, 150-pound flange (typical of many available tank risers). The motion of the ball and collection of data is controlled by instrumentation and control equipment housed in a separate remote control console. This document covers the product, Viscometer Data Acquisition Software. This document provides the software quality assurance plan, verification and validation plan, and configuration management plan for developing the software for the instrumentation that will be used to obtain rheology data from Tank SY-101

  8. Prototype implementation of segment assembling software

    Directory of Open Access Journals (Sweden)

    Pešić Đorđe

    2018-01-01

    Full Text Available IT education is very important and a lot of effort is put into the development of tools for helping students to acquire programming knowledge and for helping teachers in automating the examination process. This paper describes a prototype of the program segment assembling software used in the context of making tests in the field of algorithmic complexity. The proposed new program segment assembling model uses rules and templates. A template is a simple program segment. A rule defines combining method and data dependencies if they exist. One example of program segment assembling by the proposed system is given. Graphical user interface is also described.

  9. Design and implement of BESIII online histogramming software

    International Nuclear Information System (INIS)

    Li Fei; Wang Liang; Liu Yingjie; Chinese Academy of Sciences, Beijing; Zhu Kejun; Zhao Jingwei

    2007-01-01

    The online histogramming software is an important part of the BESIII DAQ (Data Acquisition) system. This article introduces the main requirements and design of the online histogramming software and presents how to produce, transmit and gather histograms in the distributed environment in the current software implement. The article also illustrate one smart, simple and easy to expand way of setup with xml configure database. (authors)

  10. Quality assurance of nuclear medicine computer software

    International Nuclear Information System (INIS)

    Cradduck, T.D.

    1986-01-01

    Although quality assurance activities have become well established for the hardware found in nuclear medicine little attention has been paid to computer software. This paper outlines some of the problems that exist and indicates some of the solutions presently under development. The major thrust has been towards establishment of programming standards and comprehensive documentation. Some manufacturers have developed installation verification procedures which programmers are urged to use as models for their own programs. Items that tend to cause erroneous results are discussed with the emphasis for error detection and correction being placed on proper education and training of the computer operator. The concept of interchangeable data files or 'software phantoms' for purposes of quality assurance is discussed. (Author)

  11. Software Defined Radio Datalink Implementation Using PC-Type Computers

    National Research Council Canada - National Science Library

    Zafeiropoulos, Georgios

    2003-01-01

    The objective of this thesis was to examine the feasibility of implementation and the performance of a Software Defined Radio datalink, using a common PC type host computer and a high level programming language...

  12. The dynamic of modern software development project management and the software crisis of quality. An integrated system dynamics approach towards software quality improvement

    OpenAIRE

    Nasirikaljahi, Armindokht

    2012-01-01

    The software industry is plagued by cost-overruns, delays, poor customer satisfaction and quality issues that are costing clients and customers world-wide billions of dollars each year. The phenomenon is coined The Software Crisis", and poses a huge challenge for software project management. This thesis addresses one of the core issues of the software crisis, namely software quality. The challenges of software quality are central for understanding the other symptoms of the software crisis. Th...

  13. Software quality assurance plan for the National Ignition Facility integrated computer control system

    International Nuclear Information System (INIS)

    Woodruff, J.

    1996-11-01

    Quality achievement is the responsibility of the line organizations of the National Ignition Facility (NIF) Project. This Software Quality Assurance Plan (SQAP) applies to the activities of the Integrated Computer Control System (ICCS) organization and its subcontractors. The Plan describes the activities implemented by the ICCS section to achieve quality in the NIF Project's controls software and implements the NIF Quality Assurance Program Plan (QAPP, NIF-95-499, L-15958-2) and the Department of Energy's (DOE's) Order 5700.6C. This SQAP governs the quality affecting activities associated with developing and deploying all control system software during the life cycle of the NIF Project

  14. Implementation of Digital Watermarking Using MATLAB Software

    OpenAIRE

    Karnpriya Vyas; Kirti Sethiya; Sonu Jain

    2012-01-01

    Digital watermarking holds significant promise as one of the keys to protecting proprietary digital content in the coming years. It focuses on embedding information inside a digital object such that the embedded information is in separable bound to the object. The proposed scheme has been implemented on MATLAB, as it is a high level technical computing language and interactive environment for algorithm development, data visualization, data analysis, and numerical computation. We w...

  15. Master Pump Shutdown MPS Software Quality Assurance Plan (SQAP)

    International Nuclear Information System (INIS)

    BEVINS, R.R.

    2000-01-01

    The MPSS Software Quality Assurance (SQAP) describes the tools and strategy used in the development of the MPSS software. The document also describes the methodology for controlling and managing changes to the software

  16. Lessons and challenges from software quality assessment: The ...

    African Journals Online (AJOL)

    Lessons and challenges from software quality assessment: The case of space systems software. ... esoteric software technologies and paradigms such as object oriented development, etc. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  17. Design and implementation of Skype USB user gateway software

    Science.gov (United States)

    Qi, Yang

    2017-08-01

    With the widespread application of VoIP, the client with private protocol becomes more and more popular. Skype is one of the representatives. How to connect Skype with PSTN just by Skype client has gradually become hot. This paper design and implement the software based on a kind of USB User Gateway. With the software Skype user can freely communicate with PSTN phone. FSM is designed as the core of the software, and Skype control is separated by the USB Gateway control. In this way, the communication becomes more flexible and efficient. In the actual user testing, the software obtains good results.

  18. Quality assurance of the modernized Dukovany I and C safety system software

    International Nuclear Information System (INIS)

    Karpeta, C.

    2005-01-01

    The approach to quality assurance of the software that implements the instrumentation and control functions for safety category A as per IEC 61226, which has been adopted within the 'NPP Dukovany I and C Refurbishment' project, is described. A survey of the requirements for software quality assurance of the systems that initiate protection interventions in the event of anticipated operational occurrences or accident conditions is given. The software development process applied by the system designers and manufacturers, from the software requirements specification phase to the software testing phase, is outlined. Basic information on technical audits of the software development process is also provided. (orig.)

  19. N286.7-99, A Canadian standard specifying software quality management system requirements for analytical, scientific, and design computer programs and its implementation at AECL

    International Nuclear Information System (INIS)

    Abel, R.

    2000-01-01

    Analytical, scientific, and design computer programs (referred to in this paper as 'scientific computer programs') are developed for use in a large number of ways by the user-engineer to support and prove engineering calculations and assumptions. These computer programs are subject to frequent modifications inherent in their application and are often used for critical calculations and analysis relative to safety and functionality of equipment and systems. N286.7-99(4) was developed to establish appropriate quality management system requirements to deal with the development, modification, and application of scientific computer programs. N286.7-99 provides particular guidance regarding the treatment of legacy codes

  20. Applying Quality Costs in a Software Development Environment

    Directory of Open Access Journals (Sweden)

    I.P. Hollingsworth

    1999-05-01

    Full Text Available This paper shows how Quality Costs can be a measure of software quality. The relationship between Quality Costs and other software quality metrics is briefly explained, and software development oriented versions of the two principal Quality Cost models are described. Finally the paper discusses the major issues involved in setting up a software Quality Cost programme. The concepts are based on previous research on Quality Costs in manufacturing, coupled with work on software metrics and the work currently being undertaken by the authors in a number of industries.

  1. Adopting software quality measures for healthcare processes.

    Science.gov (United States)

    Yildiz, Ozkan; Demirörs, Onur

    2009-01-01

    In this study, we investigated the adoptability of software quality measures for healthcare process measurement. Quality measures of ISO/IEC 9126 are redefined from a process perspective to build a generic healthcare process quality measurement model. Case study research method is used, and the model is applied to a public hospital's Entry to Care process. After the application, weak and strong aspects of the process can be easily observed. Access audibility, fault removal, completeness of documentation, and machine utilization are weak aspects and these aspects are the candidates for process improvement. On the other hand, functional completeness, fault ratio, input validity checking, response time, and throughput time are the strong aspects of the process.

  2. Quality assessment with the AGIR software results and experience

    International Nuclear Information System (INIS)

    Rauch, D.; Kotter, E.; Kurtz, C.; Schaefer, O.; Ehritt-Braun, C.; Burger, D.; Schaper, J.; Uhrmeister, P.

    2001-01-01

    Purpose: To evaluate whether a new software from the working group for interventional radiology (AGIR) is an appropriate tool for quality assurance in interventional radiology, and presentation of results acquired within the quality improvement process in 1999. Patients and methods: AGIR-defined parameters such as patient data, risk profile, given interventions as well as complications were registered by a recently developed software. Based on monthly data analyses, possible complications were identified and discussed in morbidity and mortality conferences. Results: 1014 interventions were performed in our institution in 1999. According to criteria established by AGIR, the complication rate was 2.7%. In addition and according to SCVIR criteria, complications were distinguished quantitatively in five classes and semiquantitatively in minor and major groups. The result was a minor complication rate of 1.8%, and a major rate of 0.9%. There were no cases of death associated with the intervention. Further strategies were developed in order to reduce the complication rate. Conclusion: Extensive quality assurance methods can be integrated in daily routine work. These methods lead to an intensive transparency of treatment results, and allow the implementation of continuous quality improvements. The development of the software is a first step in establishing a nation-wide quality assurance system. Nevertheless, modification and additional definition of the AGIR predefined parameters are required, for example, to avoid unnecessary procedures. (orig.) [de

  3. Operational excellence (six sigma) philosophy: Application to software quality assurance

    Energy Technology Data Exchange (ETDEWEB)

    Lackner, M.

    1997-11-01

    This report contains viewgraphs on operational excellence philosophy of six sigma applied to software quality assurance. This report outlines the following: goal of six sigma; six sigma tools; manufacturing vs administrative processes; Software quality assurance document inspections; map software quality assurance requirements document; failure mode effects analysis for requirements document; measuring the right response variables; and questions.

  4. Software quality assurance for safety analysis and risk management at the Savannah River Site

    International Nuclear Information System (INIS)

    Ades, M.J.; Toffer, H.; Crowe, R.D.

    1991-01-01

    As part of its Reactor Operations Improvement Program at the Savannah River Site (SRS), Westinghouse Savannah River Company (WSRC), in cooperation with the Westinghouse Hanford Company, has developed and implemented quality assurance for safety-related software for technical programs essential to the safety and reliability of reactor operations. More specifically, the quality assurance process involved the development and implementation of quality standards and attendant procedures based on industry software quality standards. These procedures were then applied to computer codes in reactor safety and probabilistic risk assessment analyses. This paper provides a review of the major aspects of the WSRC safety-related software quality assurance. In particular, quality assurance procedures are described for the different life cycle phases of the software that include the Requirements, Software Design and Implementation, Testing and Installation, Operation and Maintenance, and Retirement Phases. For each phase, specific provisions are made to categorize the range of activities, the level of responsibilities, and the documentation needed to assure the control of the software. The software quality assurance procedures developed and implemented are evolutionary in nature, and thus, prone to further refinements. These procedures, nevertheless, represent an effective controlling tool for the development, production, and operation of safety-related software applicable to reactor safety and probabilistic risk assessment analyses

  5. Rules of thumb to increase the software quality through testing

    Science.gov (United States)

    Buttu, M.; Bartolini, M.; Migoni, C.; Orlati, A.; Poppi, S.; Righini, S.

    2016-07-01

    The software maintenance typically requires 40-80% of the overall project costs, and this considerable variability mostly depends on the software internal quality: the more the software is designed and implemented to constantly welcome new changes, the lower will be the maintenance costs. The internal quality is typically enforced through testing, which in turn also affects the development and maintenance costs. This is the reason why testing methodologies have become a major concern for any company that builds - or is involved in building - software. Although there is no testing approach that suits all contexts, we infer some general guidelines learned during the Development of the Italian Single-dish COntrol System (DISCOS), which is a project aimed at producing the control software for the three INAF radio telescopes (the Medicina and Noto dishes, and the newly-built SRT). These guidelines concern both the development and the maintenance phases, and their ultimate goal is to maximize the DISCOS software quality through a Behavior-Driven Development (BDD) workflow beside a continuous delivery pipeline. We consider different topics and patterns; they involve the proper apportion of the tests (from end-to-end to low-level tests), the choice between hardware simulators and mockers, why and how to apply TDD and the dependency injection to increase the test coverage, the emerging technologies available for test isolation, bug fixing, how to protect the system from the external resources changes (firmware updating, hardware substitution, etc.) and, eventually, how to accomplish BDD starting from functional tests and going through integration and unit tests. We discuss pros and cons of each solution and point out the motivations of our choices either as a general rule or narrowed in the context of the DISCOS project.

  6. Design and Implementation of Modular Software for Programming Mobile Robots

    Directory of Open Access Journals (Sweden)

    Alessandro Farinelli

    2006-03-01

    Full Text Available This article describes a software development toolkit for programming mobile robots, that has been used on different platforms and for different robotic applications. We address design choices, implementation issues and results in the realization of our robot programming environment, that has been devised and built from many people since 1998. We believe that the proposed framework is extremely useful not only for experienced robotic software developers, but also for students approaching robotic research projects.

  7. Design and Implementation of Software for Resistance Welding Process Simulations

    DEFF Research Database (Denmark)

    Zhang, Wenqi

    2003-01-01

    Based on long time engineering research and dedicated collaborations with industry, a new welding software, SORPAS, has been developed for simulation of resistance projection and spot welding processes applying the powerful finite element method (FEM). In order to make the software directly usabl...... equipment manufacturers...... by engineers and technicians in industry, all of the important parameters in resistance welding are considered and automatically implemented into the software. With the specially designed graphic user interface for Windows, engineers (even without prior knowledge of FEM) can quickly learn and easily operate...... and utilize the software. All industrial users, including welding engineers from DaimlerChrysler, Volkswangen, PSA Peugeot Citroen, VOLVO, Siemens, ABB and so on, have started using the software just after taking a one-day training course. With the user-friendly facilities for flexible geometric design...

  8. Effective organizational solutions for implementation of DBMS software packages

    Science.gov (United States)

    Jones, D.

    1984-01-01

    The space telescope management information system development effort is a guideline for discussing effective organizational solutions used in implementing DBMS software. Focus is on the importance of strategic planning. The value of constructing an information system architecture to conform to the organization's managerial needs, the need for a senior decision maker, dealing with shifting user requirements, and the establishment of a reliable working relationship with the DBMS vendor are examined. Requirements for a schedule to demonstrate progress against a defined timeline and the importance of continued monitoring for production software control, production data control, and software enhancements are also discussed.

  9. Improvement of Computer Software Quality through Software Automated Tools.

    Science.gov (United States)

    1986-08-30

    information that are returned from the tools to the human user, and the forms in which these outputs are presented. Page 2 of 4 STAGE OF DEVELOPMENT: What... AUTOMIATED SOFTWARE TOOL MONITORING SYSTEM APPENDIX 2 2-1 INTRODUCTION This document and Automated Software Tool Monitoring Program (Appendix 1) are...t Output Output features provide links from the tool to both the human user and the target machine (where applicable). They describe the types

  10. Unisys' experience in software quality and productivity management of an existing system

    Science.gov (United States)

    Munson, John B.

    1988-01-01

    A summary of Quality Improvement techniques, implementation, and results in the maintenance, management, and modification of large software systems for the Space Shuttle Program's ground-based systems is provided.

  11. Reducing the risk of failure: Software Quality assurance standards and methods

    International Nuclear Information System (INIS)

    Elphick, J.; Cope, H.

    1992-01-01

    An effective Software Quality Assurance (SQA) program provides an overall approach to software engineering and the establishment of proven methods for the production of reliable software. And, in the authors experience, the overall costs for the software life are diminished with the application of quality methods. In their experience, the issues for implementing quality standards and practices are many. This paper addresses those issues as well as the lessons learned from developing and implementing a number of software quality assurance programs. Their experience includes the development and implementation of their own NRC accepted SQA program and an SQA program for an engineering software developer, as well as developing SQA procedures, standards, and methods for utilities, medical and commercial clients. Some of the issues addressed in this paper are: setting goals and defining quality; applying the software life cycle; addressing organizational issues; providing flexibility and increasing productivity; producing effective documentation; maintaining quality records; Imposing software configuration management; conducting reviews, audits, and controls; verification and validation; and controlling software procurement

  12. Implementation Of Carlson Survey Software2009 In Survey Works And Comparison With CDS Software

    Directory of Open Access Journals (Sweden)

    Mohamed Faraj EL Megrahi

    2017-02-01

    Full Text Available The automation surveying is one of the most influential changes to surveying concept and profession has had to go through, this has taken effect in two major courses, hardware (instrumentation used in data collection and presentation, and the software (the applications used in data processing and manipulation. Automation is majorly computer based and just like all such systems is subject to improvement often; this is manifested in the new kinds of instrumentation models every few years such as total station and newer versions of software’s. The software that has the potential to completely affect survey automation is Carlson Surveying Software. This when coupled with total station as data processing and collection methods respectively; is capable of greatly improving productivity while reducing time and cost required in the long run. However, it is only natural for users to desire a competent software and be able to choose from what is available on the market based on guided research and credible information from previous researches. Such studies not only help in choice of software but are also handy when it comes to testing approaches and recommending improvements based on advantages and disadvantages to the manufacturers to help in advancement in the software industry for better and more comfortable use. The expected outcome of the research is a successful implementation of Carlson survey 2009 software in survey works and a comparison with other existing software like Civil Design Software (CDS was highlighted its advantages and disadvantages.

  13. Lessons learned from development and quality assurance of software systems at the Halden Project

    International Nuclear Information System (INIS)

    Bjorlo, T.J.; Berg, O.; Pehrsen, M.; Dahll, G.; Sivertsen, T.

    1996-01-01

    The OECD Halden Reactor Project has developed a number of software systems within the research programmes. These programmes have comprised a wide range of topics, like studies of software for safety-critical applications, development of different operator support systems, and software systems for building and implementing graphical user interfaces. The systems have ranged from simple prototypes to installations in process plants. In the development of these software systems, Halden has gained much experience in quality assurance of different types of software. This paper summarises the accumulated experience at the Halden Project in quality assurance of software systems. The different software systems being developed at the Halden Project may be grouped into three categories. These are plant-specific software systems (one-of-a-kind deliveries), generic software products, and safety-critical software systems. This classification has been found convenient as the categories have different requirements to the quality assurance process. In addition, the experience from use of software development tools and proprietary software systems at Halden, is addressed. The paper also focuses on the experience gained from the complete software life cycle, starting with the software planning phase and ending with software operation and maintenance

  14. Process model for building quality software on internet time ...

    African Journals Online (AJOL)

    The competitive nature of the software construction market and the inherently exhilarating nature of software itself have hinged the success of any software development project on four major pillars: time to market, product quality, innovation and documentation. Unfortunately, however, existing software development models ...

  15. Quality measures and assurance for AI (Artificial Intelligence) software

    Science.gov (United States)

    Rushby, John

    1988-01-01

    This report is concerned with the application of software quality and evaluation measures to AI software and, more broadly, with the question of quality assurance for AI software. Considered are not only the metrics that attempt to measure some aspect of software quality, but also the methodologies and techniques (such as systematic testing) that attempt to improve some dimension of quality, without necessarily quantifying the extent of the improvement. The report is divided into three parts Part 1 reviews existing software quality measures, i.e., those that have been developed for, and applied to, conventional software. Part 2 considers the characteristics of AI software, the applicability and potential utility of measures and techniques identified in the first part, and reviews those few methods developed specifically for AI software. Part 3 presents an assessment and recommendations for the further exploration of this important area.

  16. Software application for quality control protocol of mammography systems

    International Nuclear Information System (INIS)

    Kjosevski, Vladimir; Gershan, Vesna; Ginovska, Margarita; Spasevska, Hristina

    2010-01-01

    Considering the fact that the Quality Control of the technological process of the mammographic system involves testing of a large number of parameters, it is clearly evident that there is a need for using the information technology for gathering, processing and storing of all the parameters that are result of this process. The main goal of this software application is facilitation and automation of the gathering, processing, storing and presenting process of the data related to the qualification of the physical and technical parameters during the quality control of the technological process of the mammographic system. The software application along with its user interface and database has been made with the Microsoft Access 2003 application which is part of the Microsoft Office 2003 software packet and has been chosen as a platform for developing because it is the most commonly used office application today among the computer users in the country. This is important because it will provide the end users a familiar environment to work in, without the need for additional training and improving the computer skills that they posses. Most importantly, the software application is easy to use, fast in calculating the parameters needed and it is an excellent way to store and display the results. There is a possibility for up scaling this software solution so it can be used by many different users at the same time over the Internet. It is highly recommended that this system is implemented as soon as possible in the quality control process of the mammographic systems due to its many advantages.(Author)

  17. SAPHIRE 8 Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Curtis Smith

    2010-02-01

    This Quality Assurance (QA) Plan documents the QA activities that will be managed by the INL related to JCN N6423. The NRC developed the SAPHIRE computer code for performing probabilistic risk assessments (PRAs) using a personal computer (PC) at the Idaho National Laboratory (INL) under Job Code Number (JCN) L1429. SAPHIRE started out as a feasibility study for a PRA code to be run on a desktop personal PC and evolved through several phases into a state-of-the-art PRA code. The developmental activity of SAPHIRE was the result of two concurrent important events: The tremendous expansion of PC software and hardware capability of the 90s and the onset of a risk-informed regulation era.

  18. Software Quality Assessment Tool Based on Meta-Models

    OpenAIRE

    Doneva Rositsa; Gaftandzhieva Silvia; Doneva Zhelyana; Staevsky Nevena

    2015-01-01

    In the software industry it is indisputably essential to control the quality of produced software systems in terms of capabilities for easy maintenance, reuse, portability and others in order to ensure reliability in the software development. But it is also clear that it is very difficult to achieve such a control through a ‘manual’ management of quality.There are a number of approaches for software quality assurance based typically on software quality models (e.g. ISO 9126, McCall’s, Boehm’s...

  19. Image quality dependence on image processing software in ...

    African Journals Online (AJOL)

    Image quality dependence on image processing software in computed radiography. ... Agfa CR readers use MUSICA software, and an upgrade with significantly different image ... Full Text: EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  20. Improving Performance of Software Implemented Floating Point Addition

    DEFF Research Database (Denmark)

    Hindborg, Andreas Erik; Karlsson, Sven

    2011-01-01

    We outline and evaluate hardware extensions to an integer processor pipeline which allow IEEE 754 oating point, FP, addition to be eciently implemented in software. With a very moderate increase in hardware resources, our perfor- mance evaluation shows that, for a benchmark that executes 12.5% FP...... addition instructions, our approach exhibits a rel- ative slowdown of 3.38 to 15.15 as compared to dedicated hardware. This is a signicant improvement of pure software emulation which leads to relative slowdowns up to 45.33....

  1. On Quality and Measures in Software Engineering

    Science.gov (United States)

    Bucur, Ion I.

    2006-01-01

    Complexity measures are mainly used to estimate vital information about reliability and maintainability of software systems from regular analysis of the source code. Such measures also provide constant feedback during a software project to assist the control of the development procedure. There exist several models to classify a software product's…

  2. Software Quality and Copyright: Issues in Computer-Assisted Instruction.

    Science.gov (United States)

    Helm, Virginia

    The two interconnected problems of educational quality and piracy are described and analyzed in this book, which begins with an investigation of the accusations regarding the alleged dismal quality of educational software. The reality behind accusations of rampant piracy and the effect of piracy on the quality of educational software is examined…

  3. Experiences with Software Quality Metrics in the EMI middleware

    International Nuclear Information System (INIS)

    Alandes, M; Meneses, D; Pucciani, G; Kenny, E M

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project to extract “code metrics” on the status of the software products and “process metrics” related to the quality of the development and support process such as reaction time to critical bugs, requirements tracking and delays in product releases.

  4. Requirement analysis of the safety-critical software implementation for the nuclear power plant

    International Nuclear Information System (INIS)

    Chang, Hoon Seon; Jung, Jae Cheon; Kim, Jae Hack; Nam, Sang Ku; Kim, Hang Bae

    2005-01-01

    The safety critical software shall be implemented under the strict regulation and standards along with hardware qualification. In general, the safety critical software has been implemented using functional block language (FBL) and structured language like C in the real project. Software design shall comply with such characteristics as; modularity, simplicity, minimizing the use of sub-routine, and excluding the interrupt logic. To meet these prerequisites, we used the computer-aided software engineering (CASE) tool to substantiate the requirements traceability matrix that were manually developed using Word processors or Spreadsheets. And the coding standard and manual have been developed to confirm the quality of software development process, such as; readability, consistency, and maintainability in compliance with NUREG/CR-6463. System level preliminary hazard analysis (PHA) is performed by analyzing preliminary safety analysis report (PSAR) and FMEA document. The modularity concept is effectively implemented for the overall module configurations and functions using RTP software development tool. The response time imposed on the basis of the deterministic structure of the safety-critical software was measured

  5. Pragmatic quality metrics for evolutionary software development models

    Science.gov (United States)

    Royce, Walker

    1990-01-01

    Due to the large number of product, project, and people parameters which impact large custom software development efforts, measurement of software product quality is a complex undertaking. Furthermore, the absolute perspective from which quality is measured (customer satisfaction) is intangible. While we probably can't say what the absolute quality of a software product is, we can determine the relative quality, the adequacy of this quality with respect to pragmatic considerations, and identify good and bad trends during development. While no two software engineers will ever agree on an optimum definition of software quality, they will agree that the most important perspective of software quality is its ease of change. We can call this flexibility, adaptability, or some other vague term, but the critical characteristic of software is that it is soft. The easier the product is to modify, the easier it is to achieve any other software quality perspective. This paper presents objective quality metrics derived from consistent lifecycle perspectives of rework which, when used in concert with an evolutionary development approach, can provide useful insight to produce better quality per unit cost/schedule or to achieve adequate quality more efficiently. The usefulness of these metrics is evaluated by applying them to a large, real world, Ada project.

  6. Quality Assurance in Software Development: An Exploratory Investigation in Software Project Failures and Business Performance

    Science.gov (United States)

    Ichu, Emmanuel A.

    2010-01-01

    Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…

  7. Building quality into performance and safety assessment software

    International Nuclear Information System (INIS)

    Wojciechowski, L.C.

    2011-01-01

    Quality assurance is integrated throughout the development lifecycle for performance and safety assessment software. The software used in the performance and safety assessment of a Canadian deep geological repository (DGR) follows the CSA quality assurance standard CSA-N286.7 [1], Quality Assurance of Analytical, Scientific and Design Computer Programs for Nuclear Power Plants. Quality assurance activities in this standard include tasks such as verification and inspection; however, much more is involved in producing a quality software computer program. The types of errors found with different verification methods are described. The integrated quality process ensures that defects are found and corrected as early as possible. (author)

  8. A Model for Quality Optimization in Software Design Processes

    NARCIS (Netherlands)

    Noppen, J.A.R.; van den Broek, P.M.; Aksit, Mehmet

    The main objective of software engineers is to design and implement systems that implement all functional and non-functional requirements. Unfortunately, it is very difficult or even generally impossible to deliver a software system that satisfies all the requirements. Even more seriously, failures

  9. Service Quality Strategy: Implementation in Algarve Hotels

    OpenAIRE

    Carlos J. F. Cândido

    2010-01-01

    This chapter addresses the problem of service quality strategy implementation and undertakes a tentative validation of three models. The first focuses on service quality, as a function of quality gaps, while the second and third ones examine strategy implementation. The models aim to help to explain how to implement a service quality strategy that simultaneously avoids quality gaps and resistance to change. Sample data has been collected through questionnaires distributed within the p...

  10. Implementing Resource-aware Multicast Forwarding in Software Defined Networks

    DEFF Research Database (Denmark)

    Poderys, Justas; Sunny, Anjusha; Soler, José

    2018-01-01

    Using multicast data transmissions, data can be eciently distributed to a high number of network users. However, in order to ef-ciently stream multimedia using multicast communication, multicast routing protocols must have knowledge of all network links and their available bandwidth. In Software......-Karp algorithm, by taking into account network topology and links load information. This paper presents the algorithm, implementation details, and an analysis of the testing results....

  11. Software quality assurance and software safety in the Biomed Control System

    International Nuclear Information System (INIS)

    Singh, R.P.; Chu, W.T.; Ludewigt, B.A.; Marks, K.M.; Nyman, M.A.; Renner, T.R.; Stradtner, R.

    1989-01-01

    The Biomed Control System is a hardware/software system used for the delivery, measurement and monitoring of heavy-ion beams in the patient treatment and biology experiment rooms in the Bevalac at the Lawrence Berkeley Laboratory (LBL). This paper describes some aspects of this system including historical background philosophy, configuration management, hardware features that facilitate software testing, software testing procedures, the release of new software quality assurance, safety and operator monitoring. 3 refs

  12. Software Quality Assurance and Controls Standard

    Science.gov (United States)

    2010-04-27

    dassurance a wor pro uc s an processes comply with predefined provisions and plans. • According to International Standard (IS) 12207 – of the 44...from document (plan) focus to process focus – Alignment with framework standard IS 12207 software life cycle (SLC) processes with exact...Books and P blications IEEE Software and Systems Engineering curriculum ABET u Certified Software Development Professional Standards ISO /IEC

  13. International Liability Issues for Software Quality

    National Research Council Canada - National Science Library

    Mead, Nancy

    2003-01-01

    This report focuses on international law related to cybercrime, international information security standards, and software liability issues as they relate to information security for critical infrastructure applications...

  14. The 7 C's for Creating Living Software: A Research Perspective for Quality-Oriented Software Engineering

    NARCIS (Netherlands)

    Aksit, Mehmet

    2004-01-01

    This article proposes the 7 C's for realizing quality-oriented software engineering practices. All the desired qualities of this approach are expressed in short by the term living software. The 7 C's are: Concern-oriented processes, Canonical models, Composable models, Certifiable models,

  15. Measuring the Software Product Quality during the Software Development Life-Cycle: An ISO Standards Perspective

    OpenAIRE

    Rafa E. Al-Qutaish

    2009-01-01

    Problem statement: The International Organization for Standardization (ISO) published a set of international standards related to the software engineering, such as ISO 12207 and ISO 9126. However, there is a set of cross-references between the two standards. Approach: The ISO 9126 on software product quality and ISO 12207 on software life cycle processes had been analyzed to invistigate the relationships between them and to make a mapping from the ISO 9126 quality characteristics to the ISO 1...

  16. Systems and software quality the next step for industrialisation

    CERN Document Server

    Wieczorek, Martin; Bons, Heinz

    2014-01-01

    Software and systems quality is playing an increasingly important role in the growth of almost all - profit and non-profit - organisations. Quality is vital to the success of enterprises in their markets. Most small trade and repair businesses use software systems in their administration and marketing processes. Every doctor's surgery is managing its patients using software. Banking is no longer conceivable without software. Aircraft, trucks and cars use more and more software to handle their increasingly complex technical systems. Innovation, competition and cost pressure are always present i

  17. Continuous software quality analysis for the ATLAS experiment

    CERN Document Server

    Washbrook, Andrew; The ATLAS collaboration

    2017-01-01

    The software for the ATLAS experiment on the Large Hadron Collider at CERN has evolved over many years to meet the demands of Monte Carlo simulation, particle detector reconstruction and data analysis. At present over 3.8 million lines of C++ code (and close to 6 million total lines of code) are maintained by an active worldwide developer community. In order to run the experiment software efficiently at hundreds of computing centres it is essential to maintain a high level of software quality standards. The methods proposed to improve software quality practices by incorporating checks into the new ATLAS software build infrastructure.

  18. Experiences with Software Quality Metrics in the EMI Middleware

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project t...

  19. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    De Guio, Federico

    2013-01-01

    The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release Validation, to constantly validate the functionality and the performance of the reconstruction software; in Monte Carlo productions. The central tool to deliver Data Quality information is a web site for browsing data quality histograms (DQM GUI). In this contribution the usage of the DQM Software in the different environments and its integration in the CMS Reconstruction Software Framework and in all production workflows are presented.

  20. A multiobjective module-order model for software quality enhancement

    NARCIS (Netherlands)

    Khoshgoftaar, TM; Liu, Y; Seliya, N

    2004-01-01

    The knowledge, prior to system operations, of which program modules are problematic is valuable to a software quality assurance team, especially when there is a constraint on software quality enhancement resources. A cost-effective approach for allocating such resources is to obtain a prediction in

  1. A framework for business oriented software quality approaches

    NARCIS (Netherlands)

    Trienekens, J.J.M.; Veenendaal, van E.P.W.M.; McMullan, J.

    1997-01-01

    The importance of software for business systems continues to grow. Software products play an increasingly important role in industry and society. The need f~r delivering "quality products" and "quality services" has become as relevant for companies in the field qf sqfiware development as for any

  2. A pattern framework for software quality assessment and tradeoff analysis

    NARCIS (Netherlands)

    Folmer, Eelke; Boscht, Jan

    The earliest design decisions often have a significant impact on software quality and are the most costly to revoke. One of the challenges in architecture design is to reduce the frequency of retrofit problems in software designs; not being able to improve the quality of a system cost effectively, a

  3. Alignment of Software Product Quality Goals in Two Outsourcing Relationships

    OpenAIRE

    Barney, Sebastian; Wohlin, Claes

    2010-01-01

    Background: Issues with software product quality are commonly reported when organisations engage in outsourcing relationships. To address this issue, value-based software engineering literature emphasises the need for all success-critical stakeholder groups to work towards a mutually agreed goal. Aim: This paper presents a case study that aims to compare and contrast the priority two groups place on software product quality — stakeholders internal to the development organisation, and stakehol...

  4. A software quality model and metrics for risk assessment

    Science.gov (United States)

    Hyatt, L.; Rosenberg, L.

    1996-01-01

    A software quality model and its associated attributes are defined and used as the model for the basis for a discussion on risk. Specific quality goals and attributes are selected based on their importance to a software development project and their ability to be quantified. Risks that can be determined by the model's metrics are identified. A core set of metrics relating to the software development process and its products is defined. Measurements for each metric and their usability and applicability are discussed.

  5. The Implementation of Computer Data Processing Software for EAST NBI

    International Nuclear Information System (INIS)

    Zhang Xiaodan; Hu Chundong; Sheng Peng; Zhao Yuanzhe; Wu Deyun; Cui Qinglong

    2014-01-01

    One of the most important project missions of neutral beam injectors is the implementation of 100 s neutral beam injection (NBI) with high power energy to the plasma of the EAST superconducting tokamak. Correspondingly, it's necessary to construct a high-speed and reliable computer data processing system for processing experimental data, such as data acquisition, data compression and storage, data decompression and query, as well as data analysis. The implementation of computer data processing application software (CDPS) for EAST NBI is presented in this paper in terms of its functional structure and system realization. The set of software is programmed in C language and runs on Linux operating system based on TCP network protocol and multi-threading technology. The hardware mainly includes industrial control computer (IPC), data server, PXI DAQ cards and so on. Now this software has been applied to EAST NBI system, and experimental results show that the CDPS can serve EAST NBI very well. (fusion engineering)

  6. Software to model AXAF image quality

    Science.gov (United States)

    Ahmad, Anees

    1993-01-01

    This draft final report describes the work performed under this delivery order from May 1992 through June 1993. The purpose of this contract was to enhance and develop an integrated optical performance modeling software for complex x-ray optical systems such as AXAF. The GRAZTRACE program developed by the MSFC Optical Systems Branch for modeling VETA-I was used as the starting baseline program. The original program was a large single file program and, therefore, could not be modified very efficiently. The original source code has been reorganized, and a 'Make Utility' has been written to update the original program. The new version of the source code consists of 36 small source files to make it easier for the code developer to manage and modify the program. A user library has also been built and a 'Makelib' utility has been furnished to update the library. With the user library, the users can easily access the GRAZTRACE source files and build a custom library. A user manual for the new version of GRAZTRACE has been compiled. The plotting capability for the 3-D point spread functions and contour plots has been provided in the GRAZTRACE using the graphics package DISPLAY. The Graphics emulator over the network has been set up for programming the graphics routine. The point spread function and the contour plot routines have also been modified to display the plot centroid, and to allow the user to specify the plot range, and the viewing angle options. A Command Mode version of GRAZTRACE has also been developed. More than 60 commands have been implemented in a Code-V like format. The functions covered in this version include data manipulation, performance evaluation, and inquiry and setting of internal parameters. The user manual for these commands has been formatted as in Code-V, showing the command syntax, synopsis, and options. An interactive on-line help system for the command mode has also been accomplished to allow the user to find valid commands, command syntax

  7. Daily quality assurance software for a satellite radiometer system

    Science.gov (United States)

    Keegstra, P. B.; Smoot, G. F.; Bennett, C. L.; Aymon, J.; Backus, C.; Deamici, G.; Hinshaw, G.; Jackson, P. D.; Kogut, A.; Lineweaver, C.

    1992-01-01

    Six Differential Microwave Radiometers (DMR) on COBE (Cosmic Background Explorer) measure the large-angular-scale isotropy of the cosmic microwave background (CMB) at 31.5, 53, and 90 GHz. Quality assurance software analyzes the daily telemetry from the spacecraft to ensure that the instrument is operating correctly and that the data are not corrupted. Quality assurance for DMR poses challenging requirements. The data are differential, so a single bad point can affect a large region of the sky, yet the CMB isotropy requires lengthy integration times (greater than 1 year) to limit potential CMB anisotropies. Celestial sources (with the exception of the moon) are not, in general, visible in the raw differential data. A 'quicklook' software system was developed that, in addition to basic plotting and limit-checking, implements a collection of data tests as well as long-term trending. Some of the key capabilities include the following: (1) stability analysis showing how well the data RMS averages down with increased data; (2) a Fourier analysis and autocorrelation routine to plot the power spectrum and confirm the presence of the 3 mK 'cosmic' dipole signal; (3) binning of the data against basic spacecraft quantities such as orbit angle; (4) long-term trending; and (5) dipole fits to confirm the spacecraft attitude azimuth angle.

  8. Software quality assurance plan for the National Ignition Facility integrated computer control system

    Energy Technology Data Exchange (ETDEWEB)

    Woodruff, J.

    1996-11-01

    Quality achievement is the responsibility of the line organizations of the National Ignition Facility (NIF) Project. This Software Quality Assurance Plan (SQAP) applies to the activities of the Integrated Computer Control System (ICCS) organization and its subcontractors. The Plan describes the activities implemented by the ICCS section to achieve quality in the NIF Project`s controls software and implements the NIF Quality Assurance Program Plan (QAPP, NIF-95-499, L-15958-2) and the Department of Energy`s (DOE`s) Order 5700.6C. This SQAP governs the quality affecting activities associated with developing and deploying all control system software during the life cycle of the NIF Project.

  9. Guidance and Control Software Project Data - Volume 4: Configuration Management and Quality Assurance Documents

    Science.gov (United States)

    Hayhurst, Kelly J. (Editor)

    2008-01-01

    The Guidance and Control Software (GCS) project was the last in a series of software reliability studies conducted at Langley Research Center between 1977 and 1994. The technical results of the GCS project were recorded after the experiment was completed. Some of the support documentation produced as part of the experiment, however, is serving an unexpected role far beyond its original project context. Some of the software used as part of the GCS project was developed to conform to the RTCA/DO-178B software standard, "Software Considerations in Airborne Systems and Equipment Certification," used in the civil aviation industry. That standard requires extensive documentation throughout the software development life cycle, including plans, software requirements, design and source code, verification cases and results, and configuration management and quality control data. The project documentation that includes this information is open for public scrutiny without the legal or safety implications associated with comparable data from an avionics manufacturer. This public availability has afforded an opportunity to use the GCS project documents for DO-178B training. This report provides a brief overview of the GCS project, describes the 4-volume set of documents and the role they are playing in training, and includes configuration management and quality assurance documents from the GCS project. Volume 4 contains six appendices: A. Software Accomplishment Summary for the Guidance and Control Software Project; B. Software Configuration Index for the Guidance and Control Software Project; C. Configuration Management Records for the Guidance and Control Software Project; D. Software Quality Assurance Records for the Guidance and Control Software Project; E. Problem Report for the Pluto Implementation of the Guidance and Control Software Project; and F. Support Documentation Change Reports for the Guidance and Control Software Project.

  10. Criteria and tools for scientific software quality measurements

    Energy Technology Data Exchange (ETDEWEB)

    Tseng, M Y [Previse Inc., Willowdale ON (Canada)

    1995-12-01

    Not all software used in the nuclear industry needs the rigorous formal verification, reliability testing and quality assessment that are being applied to safety critical software. Recently, however, there is increasing recognition that systematic and objective quality assessment of the scientific software used in design and safety analyses of nuclear facilities is necessary to support safety and licensing decisions. Because of the complexity and large size of these programs and the resource constraints faced by the AECB reviewer, it is desirable that appropriate automated tools are used wherever practical. To objectively assess the quality of software, a set of attributes of a software product by which its quality is described and evaluated must be established. These attributes must be relevant to the application domain of software under evaluation. To effectively assess the quality of software, metrics defining quantitative scale and method appropriate to determine the value of attributes need to be applied. To cost-effectively perform the evaluation, use of suitable automated tools is desirable. In this project, criteria for evaluating the quality of scientific software are presented; metrics for which those criteria can be evaluated are identified; a survey of automated tools to measure those metrics was conducted and the most appropriate tool (QA Fortran) was acquired; and the tool usage was demonstrated on three sample programs. (author) 5 refs.

  11. Criteria and tools for scientific software quality measurements

    International Nuclear Information System (INIS)

    Tseng, M.Y.

    1995-12-01

    Not all software used in the nuclear industry needs the rigorous formal verification, reliability testing and quality assessment that are being applied to safety critical software. Recently, however, there is increasing recognition that systematic and objective quality assessment of the scientific software used in design and safety analyses of nuclear facilities is necessary to support safety and licensing decisions. Because of the complexity and large size of these programs and the resource constraints faced by the AECB reviewer, it is desirable that appropriate automated tools are used wherever practical. To objectively assess the quality of software, a set of attributes of a software product by which its quality is described and evaluated must be established. These attributes must be relevant to the application domain of software under evaluation. To effectively assess the quality of software, metrics defining quantitative scale and method appropriate to determine the value of attributes need to be applied. To cost-effectively perform the evaluation, use of suitable automated tools is desirable. In this project, criteria for evaluating the quality of scientific software are presented; metrics for which those criteria can be evaluated are identified; a survey of automated tools to measure those metrics was conducted and the most appropriate tool (QA Fortran) was acquired; and the tool usage was demonstrated on three sample programs. (author) 5 refs

  12. Tool Use Within NASA Software Quality Assurance

    Science.gov (United States)

    Shigeta, Denise; Port, Dan; Nikora, Allen P.; Wilf, Joel

    2013-01-01

    As space mission software systems become larger and more complex, it is increasingly important for the software assurance effort to have the ability to effectively assess both the artifacts produced during software system development and the development process itself. Conceptually, assurance is a straightforward idea - it is the result of activities carried out by an organization independent of the software developers to better inform project management of potential technical and programmatic risks, and thus increase management's confidence in the decisions they ultimately make. In practice, effective assurance for large, complex systems often entails assessing large, complex software artifacts (e.g., requirements specifications, architectural descriptions) as well as substantial amounts of unstructured information (e.g., anomaly reports resulting from testing activities during development). In such an environment, assurance engineers can benefit greatly from appropriate tool support. In order to do so, an assurance organization will need accurate and timely information on the tool support available for various types of assurance activities. In this paper, we investigate the current use of tool support for assurance organizations within NASA, and describe on-going work at JPL for providing assurance organizations with the information about tools they need to use them effectively.

  13. Multistage switching hardware and software implementations for student experiment purpose

    Science.gov (United States)

    Sani, A.; Suherman

    2018-02-01

    Current communication and internet networks are underpinned by the switching technologies that interconnect one network to the others. Students’ understanding on networks rely on how they conver the theories. However, understanding theories without touching the reality may exert spots in the overall knowledge. This paper reports the progress of the multistage switching design and implementation for student laboratory activities. The hardware and software designs are based on three stages clos switching architecture with modular 2x2 switches, controlled by an arduino microcontroller. The designed modules can also be extended for batcher and bayan switch, and working on circuit and packet switching systems. The circuit analysis and simulation show that the blocking probability for each switch combinations can be obtained by generating random or patterned traffics. The mathematic model and simulation analysis shows 16.4% blocking probability differences as the traffic generation is uniform. The circuits design components and interfacing solution have been identified to allow next step implementation.

  14. Cognon Neural Model Software Verification and Hardware Implementation Design

    Science.gov (United States)

    Haro Negre, Pau

    Little is known yet about how the brain can recognize arbitrary sensory patterns within milliseconds using neural spikes to communicate information between neurons. In a typical brain there are several layers of neurons, with each neuron axon connecting to ˜104 synapses of neurons in an adjacent layer. The information necessary for cognition is contained in theses synapses, which strengthen during the learning phase in response to newly presented spike patterns. Continuing on the model proposed in "Models for Neural Spike Computation and Cognition" by David H. Staelin and Carl H. Staelin, this study seeks to understand cognition from an information theoretic perspective and develop potential models for artificial implementation of cognition based on neuronal models. To do so we focus on the mathematical properties and limitations of spike-based cognition consistent with existing neurological observations. We validate the cognon model through software simulation and develop concepts for an optical hardware implementation of a network of artificial neural cognons.

  15. Software Design Improvements. Part 2; Software Quality and the Design and Inspection Process

    Science.gov (United States)

    Lalli, Vincent R.; Packard, Michael H.; Ziemianski, Tom

    1997-01-01

    The application of assurance engineering techniques improves the duration of failure-free performance of software. The totality of features and characteristics of a software product are what determine its ability to satisfy customer needs. Software in safety-critical systems is very important to NASA. We follow the System Safety Working Groups definition for system safety software as: 'The optimization of system safety in the design, development, use and maintenance of software and its integration with safety-critical systems in an operational environment. 'If it is not safe, say so' has become our motto. This paper goes over methods that have been used by NASA to make software design improvements by focusing on software quality and the design and inspection process.

  16. Software quality assurance procedures for radioactive waste risk assessment codes

    International Nuclear Information System (INIS)

    Hill, I.; Mayer, J.

    1990-01-01

    This support study for the evaluation of the safety of geological disposal systems is aimed at identifying the requirements for software quality assurance procedures for radioactive waste risk assessment codes, and to recommend appropriate procedures. The research covers: (i) the analysis of existing procedures and definition of requirements; (ii) a case study of the use of some existing procedures; (iii) the definition and the implementation of procedures. The report is supported by appendices that give more detail on the procedures recommended. It is intended to provide ideas on the steps that should be taken to ensure the quality of the programs used for assessment of the safety case for radioactive waste repositories, and does not represent the introduction of wholly new ideas or techniques. The emphasis throughout is on procedures that will be easily implemented, rather than on the fully rigorous procedures that are required for some application areas. The study has concentrated on measures that will increase the confidence in repository performance assessments among the wider scientific/engineering community, and the lay public

  17. Using software metrics and software reliability models to attain acceptable quality software for flight and ground support software for avionic systems

    Science.gov (United States)

    Lawrence, Stella

    1992-01-01

    This paper is concerned with methods of measuring and developing quality software. Reliable flight and ground support software is a highly important factor in the successful operation of the space shuttle program. Reliability is probably the most important of the characteristics inherent in the concept of 'software quality'. It is the probability of failure free operation of a computer program for a specified time and environment.

  18. Total quality management implementation guidelines

    Energy Technology Data Exchange (ETDEWEB)

    1993-12-01

    These Guidelines were designed by the Energy Quality Council to help managers and supervisors in the Department of Energy Complex bring Total Quality Management to their organizations. Because the Department is composed of a rich mixture of diverse organizations, each with its own distinctive culture and quality history, these Guidelines are intended to be adapted by users to meet the particular needs of their organizations. For example, for organizations that are well along on their quality journeys and may already have achieved quality results, these Guidelines will provide a consistent methodology and terminology reference to foster their alignment with the overall Energy quality initiative. For organizations that are just beginning their quality journeys, these Guidelines will serve as a startup manual on quality principles applied in the Energy context.

  19. Training, Quality Assurance Factors, and Tools Investigation: a Work Report and Suggestions on Software Quality Assurance

    Science.gov (United States)

    Lee, Pen-Nan

    1991-01-01

    Previously, several research tasks have been conducted, some observations were obtained, and several possible suggestions have been contemplated involving software quality assurance engineering at NASA Johnson. These research tasks are briefly described. Also, a brief discussion is given on the role of software quality assurance in software engineering along with some observations and suggestions. A brief discussion on a training program for software quality assurance engineers is provided. A list of assurance factors as well as quality factors are also included. Finally, a process model which can be used for searching and collecting software quality assurance tools is presented.

  20. Continuous Software Quality analysis for the ATLAS experiment

    CERN Document Server

    Washbrook, Andrew; The ATLAS collaboration

    2017-01-01

    The regular application of software quality tools in large collaborative projects is required to reduce code defects to an acceptable level. If left unchecked the accumulation of defects invariably results in performance degradation at scale and problems with the long-term maintainability of the code. Although software quality tools are effective for identification there remains a non-trivial sociological challenge to resolve defects in a timely manner. This is a ongoing concern for the ATLAS software which has evolved over many years to meet the demands of Monte Carlo simulation, detector reconstruction and data analysis. At present over 3.8 million lines of C++ code (and close to 6 million total lines of code) are maintained by a community of hundreds of developers worldwide. It is therefore preferable to address code defects before they are introduced into a widely used software release. Recent wholesale changes to the ATLAS software infrastructure have provided an ideal opportunity to apply software quali...

  1. Quality in Education: An Implementation Handbook.

    Science.gov (United States)

    Arcaro, Jerome S.

    This book describes how the principles of quality can be applied to education. Based on the work of W. Edwards Deming and Joseph M. Juran, the book outlines a systematic and practical approach to implementing quality in educational settings. It also describes how to encourage staff participation in quality initiatives. Total-quality schools are…

  2. Toward Deriving Software Architectures from Quality Attributes

    Science.gov (United States)

    1994-08-01

    administration of Its orograms on the basis of religion creec ancestry. belief, age veteran status sexuai orientation or rn violation of federal state or Ioca...environments rely on the notion of a "tool bus" or an explicit shared repository [ Wasser - man 89] to allow easy integration of tools. 4.7 Unit...attributed parse tree and symbol table that the compiler cre- ates and annotates during its various phases. This results in a very different software

  3. Cluster implementation for parallel computation within MATLAB software environment

    International Nuclear Information System (INIS)

    Santana, Antonio O. de; Dantas, Carlos C.; Charamba, Luiz G. da R.; Souza Neto, Wilson F. de; Melo, Silvio B. Melo; Lima, Emerson A. de O.

    2013-01-01

    A cluster for parallel computation with MATLAB software the COCGT - Cluster for Optimizing Computing in Gamma ray Transmission methods, is implemented. The implementation correspond to creation of a local net of computers, facilities and configurations of software, as well as the accomplishment of cluster tests for determine and optimizing of performance in the data processing. The COCGT implementation was required by data computation from gamma transmission measurements applied to fluid dynamic and tomography reconstruction in a FCC-Fluid Catalytic Cracking cold pilot unity, and simulation data as well. As an initial test the determination of SVD - Singular Values Decomposition - of random matrix with dimension (n , n), n=1000, using the Girco's law modified, revealed that COCGT was faster in comparison to the literature [1] cluster, which is similar and operates at the same conditions. Solution of a system of linear equations provided a new test for the COCGT performance by processing a square matrix with n=10000, computing time was 27 s and for square matrix with n=12000, computation time was 45 s. For determination of the cluster behavior in relation to 'parfor' (parallel for-loop) and 'spmd' (single program multiple data), two codes were used containing those two commands and the same problem: determination of SVD of a square matrix with n= 1000. The execution of codes by means of COCGT proved: 1) for the code with 'parfor', the performance improved with the labs number from 1 to 8 labs; 2) for the code 'spmd', just 1 lab (core) was enough to process and give results in less than 1 s. In similar situation, with the difference that now the SVD will be determined from square matrix with n1500, for code with 'parfor', and n=7000, for code with 'spmd'. That results take to conclusions: 1) for the code with 'parfor', the behavior was the same already described above; 2) for code with 'spmd', the same besides having produced a larger performance, it supports a

  4. Quality Market: Design and Field Study of Prediction Market for Software Quality Control

    Science.gov (United States)

    Krishnamurthy, Janaki

    2010-01-01

    Given the increasing competition in the software industry and the critical consequences of software errors, it has become important for companies to achieve high levels of software quality. While cost reduction and timeliness of projects continue to be important measures, software companies are placing increasing attention on identifying the user…

  5. Evolvability as a Quality Attribute of Software Architectures

    NARCIS (Netherlands)

    Ciraci, S.; van den Broek, P.M.; Duchien, Laurence; D'Hondt, Maja; Mens, Tom

    We review the definition of evolvability as it appears on the literature. In particular, the concept of software evolvability is compared with other system quality attributes, such as adaptability, maintainability and modifiability.

  6. Implementation of an OAIS Repository Using Free, Open Source Software

    Science.gov (United States)

    Flathers, E.; Gessler, P. E.; Seamon, E.

    2015-12-01

    The Northwest Knowledge Network (NKN) is a regional data repository located at the University of Idaho that focuses on the collection, curation, and distribution of research data. To support our home institution and others in the region, we offer services to researchers at all stages of the data lifecycle—from grant application and data management planning to data distribution and archive. In this role, we recognize the need to work closely with other data management efforts at partner institutions and agencies, as well as with larger aggregation efforts such as our state geospatial data clearinghouses, data.gov, DataONE, and others. In the past, one of our challenges with monolithic, prepackaged data management solutions is that customization can be difficult to implement and maintain, especially as new versions of the software are released that are incompatible with our local codebase. Our solution is to break the monolith up into its constituent parts, which offers us several advantages. First, any customizations that we make are likely to fall into areas that can be accessed through Application Program Interfaces (API) that are likely to remain stable over time, so our code stays compatible. Second, as components become obsolete or insufficient to meet new demands that arise, we can replace the individual components with minimal effect on the rest of the infrastructure, causing less disruption to operations. Other advantages include increased system reliability, staggered rollout of new features, enhanced compatibility with legacy systems, reduced dependence on a single software company as a point of failure, and the separation of development into manageable tasks. In this presentation, we describe our application of the Service Oriented Architecture (SOA) design paradigm to assemble a data repository that conforms to the Open Archival Information System (OAIS) Reference Model primarily using a collection of free and open-source software. We detail the design

  7. A survey of Canadian medical physicists: software quality assurance of in-house software.

    Science.gov (United States)

    Salomons, Greg J; Kelly, Diane

    2015-01-05

    This paper reports on a survey of medical physicists who write and use in-house written software as part of their professional work. The goal of the survey was to assess the extent of in-house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple-choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software-related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines.

  8. User-perceptions of embedded software quality

    NARCIS (Netherlands)

    Kusters, R.J.; Solingen, van D.M.; Trienekens, J.J.M.; Budgen, D.; Hoffnagle, G.; Trienekens, J.J.M.

    1997-01-01

    Many researchers and practitioners have recognised that the perception of `quality' is largely influenced by personal view and application context. Depending on personal goals, interests and background, the interpretation of the quality concept is different per individual. In this paper, an approach

  9. Human Factors in Software Development Processes: Measuring System Quality

    DEFF Research Database (Denmark)

    Abrahão, Silvia; Baldassarre, Maria Teresa; Caivano, Danilo

    2016-01-01

    Software Engineering and Human-Computer Interaction look at the development process from different perspectives. They apparently use very different approaches, are inspired by different principles and address different needs. But, they definitively have the same goal: develop high quality software...... in the most effective way. The second edition of the workshop puts particular attention on efforts of the two communities in enhancing system quality. The research question discussed is: who, what, where, when, why, and how should we evaluate?...

  10. Bottlenecks in Software Defect Prediction Implementation in Industrial Projects

    OpenAIRE

    Hryszko Jarosław; Madeyski Lech

    2015-01-01

    Case studies focused on software defect prediction in real, industrial software development projects are extremely rare. We report on dedicated R&D project established in cooperation between Wroclaw University of Technology and one of the leading automotive software development companies to research possibilities of introduction of software defect prediction using an open source, extensible software measurement and defect prediction framework called DePress (Defect Prediction in Software Syst...

  11. A research review of quality assessment for software

    Science.gov (United States)

    1991-01-01

    Measures were recommended to assess the quality of software submitted to the AdaNet program. The quality factors that are important to software reuse are explored and methods of evaluating those factors are discussed. Quality factors important to software reuse are: correctness, reliability, verifiability, understandability, modifiability, and certifiability. Certifiability is included because the documentation of many factors about a software component such as its efficiency, portability, and development history, constitute a class for factors important to some users, not important at all to other, and impossible for AdaNet to distinguish between a priori. The quality factors may be assessed in different ways. There are a few quantitative measures which have been shown to indicate software quality. However, it is believed that there exists many factors that indicate quality and have not been empirically validated due to their subjective nature. These subjective factors are characterized by the way in which they support the software engineering principles of abstraction, information hiding, modularity, localization, confirmability, uniformity, and completeness.

  12. Quality Attributes for Mission Flight Software: A Reference for Architects

    Science.gov (United States)

    Wilmot, Jonathan; Fesq, Lorraine; Dvorak, Dan

    2016-01-01

    In the international standards for architecture descriptions in systems and software engineering (ISO/IEC/IEEE 42010), "concern" is a primary concept that often manifests itself in relation to the quality attributes or "ilities" that a system is expected to exhibit - qualities such as reliability, security and modifiability. One of the main uses of an architecture description is to serve as a basis for analyzing how well the architecture achieves its quality attributes, and that requires architects to be as precise as possible about what they mean in claiming, for example, that an architecture supports "modifiability." This paper describes a table, generated by NASA's Software Architecture Review Board, which lists fourteen key quality attributes, identifies different important aspects of each quality attribute and considers each aspect in terms of requirements, rationale, evidence, and tactics to achieve the aspect. This quality attribute table is intended to serve as a guide to software architects, software developers, and software architecture reviewers in the domain of mission-critical real-time embedded systems, such as space mission flight software.

  13. Implementation of Software Configuration Management Process by Models: Practical Experiments and Learned Lessons

    Directory of Open Access Journals (Sweden)

    Bartusevics Arturs

    2014-12-01

    Full Text Available Nowadays software configuration management process is not only dilemma which system should be used for version control or how to merge changes from one source code branch to other. There are multiple tasks such as version control, build management, deploy management, status accounting, bug tracking and many others that should be solved to support full configuration management process according to most popular quality standards. The main scope of the mentioned process is to include only valid and tested software items to final version of product and prepare a new version as soon as possible. To implement different tasks of software configuration management process, a set of different tools, scripts and utilities should be used. The current paper provides a new model-based approach to implementation of configuration management. Using different models, a new approach helps to organize existing solutions and develop new ones by a parameterized way, thus increasing reuse of solutions. The study provides a general description of new model-based conception and definitions of all models needed to implement a new approach. The second part of the paper contains an overview of criteria, practical experiments and lessons learned from using new models in software configuration management. Finally, further works are defined based on results of practical experiments and lessons learned.

  14. Geologic software for nuclear waste repository studies: A quality assurance program

    International Nuclear Information System (INIS)

    Figuli, S.; English, S.L.

    1987-04-01

    This paper discusses a Quality Assurance (QA) program that Kent State University (KSU) has implemented for the development of geologic software. The software being developed at KSU will be used in the site characterization of nuclear waste repositories and must meet the requirements of federal regulations. This QA program addresses the development of models that will be used in the evaluation of the long-term climatic stability of three sites in the western US

  15. Embedding Quality Function Deployment In Software Development ...

    African Journals Online (AJOL)

    2013-03-01

    Mar 1, 2013 ... This paper focuses on how customers' voice can be heard in order to reduce development and manufacturing costs ..... engineer and customer meet and define the .... Figure 4 Customer's Satisfaction Requirement. Quality ...

  16. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    Batinkov, Atanas Ivanov

    2013-01-01

    The Data Quality Monitoring Software proved to be a central tool in the Compact Muon Solenoid experiment. Its flexibility allowed its integration in several environments: online, for real-time detector monitoring; offline, for the final, fine-grained data certification. The usage of the Data Quality Monitoring software in the different environments and its integration in the Compact Muon Solenoid reconstruction software framework and in all production workflows are presented. The main technical challenges and the adopted solutions to them will be also discussed with emphasis on functionality, long-term robustness and performance.

  17. Sandia National Laboratories ASCI Applications Software Quality Engineering Practices; TOPICAL

    International Nuclear Information System (INIS)

    ZEPPER, JOHN D.; ARAGON, KATHRYN MARY; ELLIS, MOLLY A.; BYLE, KATHLEEN A.; EATON, DONNA SUE

    2002-01-01

    This document provides a guide to the deployment of the software verification activities, software engineering practices, and project management principles that guide the development of Accelerated Strategic Computing Initiative (ASCI) applications software at Sandia National Laboratories (Sandia). The goal of this document is to identify practices and activities that will foster the development of reliable and trusted products produced by the ASCI Applications program. Document contents include an explanation of the structure and purpose of the ASCI Quality Management Council, an overview of the software development lifecycle, an outline of the practices and activities that should be followed, and an assessment tool. These sections map practices and activities at Sandia to the ASCI Software Quality Engineering: Goals, Principles, and Guidelines, a Department of Energy document

  18. Quality Management Systems Implementation Compared With Organizational Maturity in Hospital.

    Science.gov (United States)

    Moradi, Tayebeh; Jafari, Mehdi; Maleki, Mohammad Reza; Naghdi, Seyran; Ghiasvand, Hesam

    2015-07-27

    A quality management system can provide a framework for continuous improvement in order to increase the probability of customers and other stakeholders' satisfaction. The test maturity model helps organizations to assess the degree of maturity in implementing effective and sustained quality management systems; plan based on the current realities of the organization and prioritize their improvement programs. We aim to investigate and compare the level of organizational maturity in hospitals with the status of quality management systems implementation. This analytical cross sectional study was conducted among hospital administrators and quality experts working in hospitals with over 200 beds located in Tehran. In the first step, 32 hospitals were selected and then 96 employees working in the selected hospitals were studied. The data were gathered using the implementation checklist of quality management systems and the organization maturity questionnaire derived from ISO 10014. The content validity was calculated using Lawshe method and the reliability was estimated using test - retest method and calculation of Cronbach's alpha coefficient. The descriptive and inferential statistics were used to analyze the data using SPSS 18 software. According to the table, the mean score of organizational maturity among hospitals in the first stage of quality management systems implementation was equal to those in the third stage and hypothesis was rejected (p-value = 0.093). In general, there is no significant difference in the organizational maturity between the first and third level hospitals (in terms of implementation of quality management systems). Overall, the findings of the study show that there is no significant difference in the organizational maturity between the hospitals in different levels of the quality management systems implementation and in fact, the maturity of the organizations cannot be attributed to the implementation of such systems. As a result, hospitals

  19. [Software for illustrating a cost-quality balance carried out by clinical laboratory practice].

    Science.gov (United States)

    Nishibori, Masahiro; Asayama, Hitoshi; Kimura, Satoshi; Takagi, Yasushi; Hagihara, Michio; Fujiwara, Mutsunori; Yoneyama, Akiko; Watanabe, Takashi

    2010-09-01

    We have no proper reference indicating the quality of clinical laboratory practice, which should clearly illustrates that better medical tests require more expenses. Japanese Society of Laboratory Medicine was concerned about recent difficult medical economy and issued a committee report proposing a guideline to evaluate the good laboratory practice. According to the guideline, we developed software that illustrate a cost-quality balance carried out by clinical laboratory practice. We encountered a number of controversial problems, for example, how to measure and weight each quality-related factor, how to calculate costs of a laboratory test and how to consider characteristics of a clinical laboratory. Consequently we finished only prototype software within the given period and the budget. In this paper, software implementation of the guideline and the above-mentioned problems are summarized. Aiming to stimulate these discussions, the operative software will be put on the Society's homepage for trial

  20. C++ Software Quality in the ATLAS experiment

    CERN Document Server

    Roe, Shaun; The ATLAS collaboration; Kluth, Stefan; Seuster, Rolf; Snyder, Scott; Obreshkov, Emil; Sherwood, Peter; Stewart, Graeme

    2016-01-01

    In this paper we explain how the C++ code quality is managed in ATLAS using a range of tools from compile-time through to run time testing and reflect on the substantial progress made in the last two years largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other available code analysis tools are also discussed, as is the role of unit testing with an example of how the googlemock framework can be applied to our codebase.

  1. Coal Quality Expert: Status and software specifications

    International Nuclear Information System (INIS)

    Harrison, C.D.

    1992-01-01

    Under the Clean Coal Technology Program (Clean Coal Round 1), the US Department of Energy (DOE) and the Electric Power Research Institute (EPRI) are funding the development and demonstration of a computer program called the Coal Quality Expert (CQE trademark). When finished, the CQE will be a comprehensive PC-based program which can be used to evaluate several potential coal cleaning, blending, and switching options to reduce power plant emissions while minimizing generation costs. The CQE will be flxible in nature and capable of evaluating various qualities of coal, available transportation options, performance issues, and alternative emissions control strategies. This allows the CQE to determine the most cost-effective coal and the least expensive emissions control strategy for a given plant. To accomplish this, the CQE will be composed of technical models to evaluate performance issues; environmental models to evaluate environmental and regulatory issues; and cost estimating models to predict costs for installations of new and retrofit coal cleaning processes, power production equipment, and emissions control systems as well as other production costs such as consumables (fuel, scrubber additive, etc.), waste disposal, operating and maintenance, and replacement energy costs. These technical, environmental, and economic models as well as a graphical user interface will be developed for the CQE. And, in addition, to take advantage of already existing capability, the CQE will rely on seamless integration of already proven and extensively used computer programs such as the EPRI Coal Quality Information Systems, Coal Quality Impact Model (CQIM trademark), and NO x Pert. 2 figs

  2. Management of Tutorial Software Quality Programs

    OpenAIRE

    Ion Ivan

    2007-01-01

    The paper presents the requirements for rising the quality of educational processes. There are defined the context in which the activities of our course are unreeled. It is supposed a series of measures for changing the human factor role toward efficiency and creativity.

  3. Study on Software Quality Improvement based on Rayleigh Model and PDCA Model

    OpenAIRE

    Ning Jingfeng; Hu Ming

    2013-01-01

    As the software industry gradually becomes mature, software quality is regarded as the life of a software enterprise. This article discusses how to improve the quality of software, applies Rayleigh model and PDCA model to the software quality management, combines with the defect removal effectiveness index, exerts PDCA model to solve the problem of quality management objectives when using the Rayleigh model in bidirectional quality improvement strategies of software quality management, a...

  4. A survey of Canadian medical physicists: software quality assurance of in‐house software

    Science.gov (United States)

    Kelly, Diane

    2015-01-01

    This paper reports on a survey of medical physicists who write and use in‐house written software as part of their professional work. The goal of the survey was to assess the extent of in‐house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple‐choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software‐related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines. PACS number: 87.55.Qr PMID:25679168

  5. Software quality assurance (SQA) for Savannah River reactors

    Energy Technology Data Exchange (ETDEWEB)

    Schaumann, C.M.

    1990-01-01

    Over the last 25 years, the Savannah River Site (SRS) has developed a strong Software Quality Assurance (SQA) program. It provides the information and management controls required of a high quality auditable system. The SRS SQA program provides the framework to meet the requirements in increasing regulation.

  6. Faster Defect Resolution with Higher Technical Quality of Software

    NARCIS (Netherlands)

    Luijten, B.; Visser, J.

    2010-01-01

    We performed an empirical study of the relation between technical quality of software products and the defect resolution performance of their maintainers. In particular, we tested the hypothesis that ratings for source code maintainability, as employed by the SIG quality model, are correlated with

  7. Spacelab user implementation assessment study. (Software requirements analysis). Volume 2: Technical report

    Science.gov (United States)

    1976-01-01

    The engineering analyses and evaluation studies conducted for the Software Requirements Analysis are discussed. Included are the development of the study data base, synthesis of implementation approaches for software required by both mandatory onboard computer services and command/control functions, and identification and implementation of software for ground processing activities.

  8. The quality analysis system implemented by FRAGEMA

    International Nuclear Information System (INIS)

    Kopff, G.

    1988-01-01

    Systematic statistical processing of measurements and quality control data obtained through manufacturing and conformity inspection is necessary for global knowledge of the fuel quality, which is useful both to the designer and to the manufacturer. For this aim the quality control data management and processing system implemented by FRAGEMA is described and illustrated with examples of the different types of statistical quality reports which are printed out. (orig.)

  9. Application of software quality assurance methods in validation and maintenance of reactor analysis computer codes

    International Nuclear Information System (INIS)

    Reznik, L.

    1994-01-01

    Various computer codes employed at Israel Electricity Company for preliminary reactor design analysis and fuel cycle scoping calculations have been often subject to program source modifications. Although most changes were due to computer or operating system compatibility problems, a number of significant modifications were due to model improvement and enhancements of algorithm efficiency and accuracy. With growing acceptance of software quality assurance requirements and methods, a program of implementing extensive testing of modified software has been adopted within the regular maintenance activities. In this work survey has been performed of various software quality assurance methods of software testing which belong mainly to the two major categories of implementation ('white box') and specification-based ('black box') testing. The results of this survey exhibits a clear preference of specification-based testing. In particular the equivalence class partitioning method and the boundary value method have been selected as especially suitable functional methods for testing reactor analysis codes.A separate study of software quality assurance methods and techniques has been performed in this work objective to establish appropriate pre-test software specification methods. Two methods of software analysis and specification have been selected as the most suitable for this purpose: The method of data flow diagrams has been shown to be particularly valuable for performing the functional/procedural software specification while the entities - relationship diagrams has been approved to be efficient for specifying software data/information domain. Feasibility of these two methods has been analyzed in particular for software uncertainty analysis and overall code accuracy estimation. (author). 14 refs

  10. Quality assurance requirements for the computer software and safety analyses

    International Nuclear Information System (INIS)

    Husarecek, J.

    1992-01-01

    The requirements are given as placed on the development, procurement, maintenance, and application of software for the creation or processing of data during the design, construction, operation, repair, maintenance and safety-related upgrading of nuclear power plants. The verification and validation processes are highlighted, and the requirements put on the software documentation are outlined. The general quality assurance principles applied to safety analyses are characterized. (J.B.). 1 ref

  11. Framework for implementing product portfolio management in software business

    NARCIS (Netherlands)

    Jagroep, Erik; Van De Weerd, Inge; Brinkkemper, Sjaak; Dobbe, Ton

    2014-01-01

    Whether a software product company takes up a project depends on the strategic decisions that are made with regard to an organization's products. A software project needs to fit strategic goals and enable an organization to realize a vision through its software products. Making decisions on a

  12. Metric-based Evaluation of Implemented Software Architectures

    NARCIS (Netherlands)

    Bouwers, E.M.

    2013-01-01

    Software systems make up an important part of our daily lives. Just like all man- made objects, the possibilities of a software system are constrained by the choices made during its creation. The complete set of these choices can be referred to as the software architecture of a system. Since the

  13. ARCHITECTURE SOFTWARE SOLUTION TO SUPPORT AND DOCUMENT MANAGEMENT QUALITY SYSTEM

    Directory of Open Access Journals (Sweden)

    Milan Eric

    2010-12-01

    Full Text Available One of the basis of a series of standards JUS ISO 9000 is quality system documentation. An architecture of the quality system documentation depends on the complexity of business system. An establishment of an efficient management documentation of system of quality is of a great importance for the business system, as well as in the phase of introducing the quality system and in further stages of its improvement. The study describes the architecture and capability of software solutions to support and manage the quality system documentation in accordance with the requirements of standards ISO 9001:2001, ISO 14001:2005 HACCP etc.

  14. IEEE [Institute of Electrical and Electronics Engineers] standards and nuclear software quality engineering

    International Nuclear Information System (INIS)

    Daughtrey, T.

    1988-01-01

    Significant new nuclear-specific software standards have recently been adopted under the sponsorship of the American Nuclear Society and the American Society of Mechanical Engineers. The interest of the US Nuclear Regulatory Commission has also been expressed through their issuance of NUREG/CR-4640. These efforts all indicate a growing awareness of the need for thorough, referenceable expressions of the way to build in and evaluate quality in nuclear software. A broader professional perspective can be seen in the growing number of software engineering standards sponsored by the Institute of Electrical and Electronics Engineers (IEEE) Computer Society. This family of standards represents a systematic effort to capture professional consensus on quality practices throughout the software development life cycle. The only omission-the implementation phase-is treated by accepted American National Standards Institute or de facto standards for programming languages

  15. EMMC guidance on quality assurance for academic materials modelling software engineering

    OpenAIRE

    European Materials Modelling Council

    2015-01-01

    Proposed recommendations for software development in LEIT projects. This document presents the advice of software owners, commercial and academic, on what academic software could do to generate better quality software, ready to be used by third parties.

  16. Cost Analysis of Poor Quality Using a Software Simulation

    Directory of Open Access Journals (Sweden)

    Jana Fabianová

    2017-02-01

    Full Text Available The issues of quality, cost of poor quality and factors affecting quality are crucial to maintaining a competitiveness regarding to business activities. Use of software applications and computer simulation enables more effective quality management. Simulation tools offer incorporating the variability of more variables in experiments and evaluating their common impact on the final output. The article presents a case study focused on the possibility of using computer simulation Monte Carlo in the field of quality management. Two approaches for determining the cost of poor quality are introduced here. One from retrospective scope of view, where the cost of poor quality and production process are calculated based on historical data. The second approach uses the probabilistic characteristics of the input variables by means of simulation, and reflects as a perspective view of the costs of poor quality. Simulation output in the form of a tornado and sensitivity charts complement the risk analysis.

  17. The software improvement process - tools and rules to encourage quality

    International Nuclear Information System (INIS)

    Sigerud, K.; Baggiolini, V.

    2012-01-01

    The Applications section of the CERN accelerator controls group has decided to apply a systematic approach to quality assurance (QA), the 'Software Improvement Process' - SIP. This process focuses on three areas: the development process itself, suitable QA tools, and how to practically encourage developers to do QA. For each stage of the development process we have agreed on the recommended activities and deliverables, and identified tools to automate and support the task. For example we do more code reviews. As peer reviews are resource intensive, we only do them for complex parts of a product. As a complement, we are using static code checking tools, like FindBugs and Checkstyle. We also encourage unit testing and have agreed on a minimum level of test coverage recommended for all products, measured using Clover. Each of these tools is well integrated with our IDE (Eclipse) and give instant feedback to the developer about the quality of their code. The major challenges of SIP have been to 1) agree on common standards and configurations, for example common code formatting and Javadoc documentation guidelines, and 2) how to encourage the developers to do QA. To address the second point, we have successfully implemented 'SIP days', i.e. one day dedicated to QA work to which the whole group of developers participates, and 'Top/Flop' lists, clearly indicating the best and worst products with regards to SIP guidelines and standards, for example test coverage. This paper presents the SIP initiative in more detail, summarizing our experience since two years and our future plans. (authors)

  18. Requirements for guidelines systems: implementation challenges and lessons from existing software-engineering efforts.

    Science.gov (United States)

    Shah, Hemant; Allard, Raymond D; Enberg, Robert; Krishnan, Ganesh; Williams, Patricia; Nadkarni, Prakash M

    2012-03-09

    A large body of work in the clinical guidelines field has identified requirements for guideline systems, but there are formidable challenges in translating such requirements into production-quality systems that can be used in routine patient care. Detailed analysis of requirements from an implementation perspective can be useful in helping define sub-requirements to the point where they are implementable. Further, additional requirements emerge as a result of such analysis. During such an analysis, study of examples of existing, software-engineering efforts in non-biomedical fields can provide useful signposts to the implementer of a clinical guideline system. In addition to requirements described by guideline-system authors, comparative reviews of such systems, and publications discussing information needs for guideline systems and clinical decision support systems in general, we have incorporated additional requirements related to production-system robustness and functionality from publications in the business workflow domain, in addition to drawing on our own experience in the development of the Proteus guideline system (http://proteme.org). The sub-requirements are discussed by conveniently grouping them into the categories used by the review of Isern and Moreno 2008. We cite previous work under each category and then provide sub-requirements under each category, and provide example of similar work in software-engineering efforts that have addressed a similar problem in a non-biomedical context. When analyzing requirements from the implementation viewpoint, knowledge of successes and failures in related software-engineering efforts can guide implementers in the choice of effective design and development strategies.

  19. Quality of Service Attributes for Software as a Service

    Directory of Open Access Journals (Sweden)

    Lukas Burkon

    2013-07-01

    Full Text Available Software as a Service (SaaS has been developing for over ten years and, is reaching a mature level, where quality and its monitoring and management become significant. Although, SaaS is derived from the ASP model, SaaS background and architecture is different and therefore also SaaS quality management is based on different concepts. This paper is focused on the difference between traditional IT outsourcing and SaaS and proposes a set of quality attributes appropriate for the management of the SaaS quality.

  20. Software quality assurance plan for void fraction instrument

    International Nuclear Information System (INIS)

    Gimera, M.

    1994-01-01

    Waste Tank SY-101 has been the focus of extensive characterization work over the past few years. The waste continually generates gases, most notably hydrogen, which are periodically released from the waste. Gas can be trapped in tank waste in three forms: as void gas (bubbles), dissolved gas, or absorbed gas. Void fraction is the volume percentage of a given sample that is comprised of void gas. The void fraction instrument (VFI) acquires the data necessary to calculate void fraction. This document covers the product, Void Fraction Data Acquisition Software. The void fraction software being developed will have the ability to control the void fraction instrument hardware and acquire data necessary to calculate the void fraction in samples. This document provides the software quality assurance plan, verification and validation plan, and configuration management plan for developing the software for the instrumentation that will be used to obtain void fraction data from Tank SY-101

  1. Early experiences building a software quality prediction model

    Science.gov (United States)

    Agresti, W. W.; Evanco, W. M.; Smith, M. C.

    1990-01-01

    Early experiences building a software quality prediction model are discussed. The overall research objective is to establish a capability to project a software system's quality from an analysis of its design. The technical approach is to build multivariate models for estimating reliability and maintainability. Data from 21 Ada subsystems were analyzed to test hypotheses about various design structures leading to failure-prone or unmaintainable systems. Current design variables highlight the interconnectivity and visibility of compilation units. Other model variables provide for the effects of reusability and software changes. Reported results are preliminary because additional project data is being obtained and new hypotheses are being developed and tested. Current multivariate regression models are encouraging, explaining 60 to 80 percent of the variation in error density of the subsystems.

  2. Evaluating the Effect of Software Quality Characteristics on Health Care Quality Indicators

    Directory of Open Access Journals (Sweden)

    Sakineh Aghazadeh

    2015-07-01

    Full Text Available Introduction: Various types of software are used in health care organizations to manage information and care processes. The quality of software has been an important concern for both health authorities and designers of Health Information Technology. Thus, assessing the effect of software quality on the performance quality of healthcare institutions is essential. Method: The most important health care quality indicators in relation to software quality characteristics are provided via an already performed literature review. ISO 9126 standard model is used for definition and integration of various characteristics of software quality. The effects of software quality characteristics and sub-characteristics on the healthcare indicators are evaluated through expert opinion analyses. A questionnaire comprising of 126 questions of 10-point Likert scale was used to gather opinions of experts in the field of Medical/Health Informatics. The data was analyzed using Structural Equation Modeling. Results: Our findings showed that software Maintainability was rated as the most effective factor on user satisfaction (R2 =0.89 and Functionality as the most important and independent variable affecting patient care quality (R2 =0.98. Efficiency was considered as the most effective factor on workflow (R2 =0.97, and Maintainability as the most important factor that affects healthcare communication (R2 =0.95. Usability and Efficiency were rated as the most effectual factor affecting patient satisfaction (R2 =0.80, 0.81. Reliability, Maintainability, and Efficiency were considered as the main factors affecting care costs (R2 =0.87, 0.74, 0.87. Conclusion: We presented a new model based on ISO standards. The model demonstrates and weighs the relations between software quality characteristics and healthcare quality indicators. The clear relationships between variables and the type of the metrics and measurement methods used in the model make it a reliable method to assess

  3. Quality control in diagnostic radiology: software (Visual Basic 6) and database applications

    International Nuclear Information System (INIS)

    Md Saion Salikin; Muhammad Farid Abdul Khalid

    2002-01-01

    Quality Assurance programme in diagnostic Radiology is being implemented by the Ministry of Health (MoH) in Malaysia. Under this program the performance of an x-ray machine used for diagnostic purpose is tested by using the approved procedure which is commonly known as Quality Control in diagnostic radiology. The quality control or performance tests are carried out b a class H licence holder issued the Atomic Energy Licensing Act 1984. There are a few computer applications (software) that are available in the market which can be used for this purpose. A computer application (software) using Visual Basics 6 and Microsoft Access, is being developed to expedite data handling, analysis and storage as well as report writing of the quality control tests. In this paper important features of the software for quality control tests are explained in brief. A simple database is being established for this purpose which is linked to the software. Problems encountered in the preparation of database are discussed in this paper. A few examples of practical usage of the software and database applications are presented in brief. (Author)

  4. The software product assurance metrics study: JPL's software systems quality and productivity

    Science.gov (United States)

    Bush, Marilyn W.

    1989-01-01

    The findings are reported of the Jet Propulsion Laboratory (JPL)/Software Product Assurance (SPA) Metrics Study, conducted as part of a larger JPL effort to improve software quality and productivity. Until recently, no comprehensive data had been assembled on how JPL manages and develops software-intensive systems. The first objective was to collect data on software development from as many projects and for as many years as possible. Results from five projects are discussed. These results reflect 15 years of JPL software development, representing over 100 data points (systems and subsystems), over a third of a billion dollars, over four million lines of code and 28,000 person months. Analysis of this data provides a benchmark for gauging the effectiveness of past, present and future software development work. In addition, the study is meant to encourage projects to record existing metrics data and to gather future data. The SPA long term goal is to integrate the collection of historical data and ongoing project data with future project estimations.

  5. Evaluation of features to support safety and quality in general practice clinical software

    Science.gov (United States)

    2011-01-01

    Background Electronic prescribing is now the norm in many countries. We wished to find out if clinical software systems used by general practitioners in Australia include features (functional capabilities and other characteristics) that facilitate improved patient safety and care, with a focus on quality use of medicines. Methods Seven clinical software systems used in general practice were evaluated. Fifty software features that were previously rated as likely to have a high impact on safety and/or quality of care in general practice were tested and are reported here. Results The range of results for the implementation of 50 features across the 7 clinical software systems was as follows: 17-31 features (34-62%) were fully implemented, 9-13 (18-26%) partially implemented, and 9-20 (18-40%) not implemented. Key findings included: Access to evidence based drug and therapeutic information was limited. Decision support for prescribing was available but varied markedly between systems. During prescribing there was potential for medicine mis-selection in some systems, and linking a medicine with its indication was optional. The definition of 'current medicines' versus 'past medicines' was not always clear. There were limited resources for patients, and some medicines lists for patients were suboptimal. Results were provided to the software vendors, who were keen to improve their systems. Conclusions The clinical systems tested lack some of the features expected to support patient safety and quality of care. Standards and certification for clinical software would ensure that safety features are present and that there is a minimum level of clinical functionality that clinicians could expect to find in any system.

  6. Arquitectura orientada a servicios para software de apoyo para el proceso personal de software A service oriented architecture for the implementation of the personal software process

    Directory of Open Access Journals (Sweden)

    Erick Salinas

    2011-06-01

    Full Text Available El presente trabajo describe una arquitectura orientada a servicios para un software que tiene como objetivo facilitar la implementación de un Proceso Personal de Software en un equipo de desarrollo u organización. Entre las características que posee este software y que son relevantes de mencionar están las de entregar extensibilidad e independencia, esto se ve reflejado en la facilidad para agregar nuevas herramientas al proceso de desarrollo de software integradas al Proceso Personal de Software con un máximo de independencia de sistemas operativos y lenguajes de programación. El software implementado realiza la recolección de los datos necesarios para el Proceso Personal de Software casi completamente automática, considerando que el administrador solamente clasifica los errores que pueden ocurrir cuando se utiliza algún lenguaje de programación en particular, entre otras pequeñas tareas. Esta facilidad de uso hace que la implementación del Proceso Personal de Software se realice exitosamente con un bajo esfuerzo requerido por los integrantes del equipo de desarrollo.This work describes a service oriented architecture of a software application that facilitates the implementation of the Personal Software Process by a development team or an organization. Some of the characteristics of this software and which are important to mention are extensibility and technical environment independence. These characteristics facilitate the process of adding new tools to the software development process integrating them to the Personal Software Process independently of the operating systems and programming languages being used. The implemented software undertakes the data collection necessary to the Personal Software Process almost automatically, since the administrator must only classify the errors that may occur when a particular programming language is used, among other small tasks. This ease of use approach helps to make the implementation of

  7. Continuous integration and quality control for scientific software

    Science.gov (United States)

    Neidhardt, A.; Ettl, M.; Brisken, W.; Dassing, R.

    2013-08-01

    Modern software has to be stable, portable, fast and reliable. This is going to be also more and more important for scientific software. But this requires a sophisticated way to inspect, check and evaluate the quality of source code with a suitable, automated infrastructure. A centralized server with a software repository and a version control system is one essential part, to manage the code basis and to control the different development versions. While each project can be compiled separately, the whole code basis can also be compiled with one central “Makefile”. This is used to create automated, nightly builds. Additionally all sources are inspected automatically with static code analysis and inspection tools, which check well-none error situations, memory and resource leaks, performance issues, or style issues. In combination with an automatic documentation generator it is possible to create the developer documentation directly from the code and the inline comments. All reports and generated information are presented as HTML page on a Web server. Because this environment increased the stability and quality of the software of the Geodetic Observatory Wettzell tremendously, it is now also available for scientific communities. One regular customer is already the developer group of the DiFX software correlator project.

  8. Internal quality control: planning and implementation strategies.

    Science.gov (United States)

    Westgard, James O

    2003-11-01

    The first essential in setting up internal quality control (IQC) of a test procedure in the clinical laboratory is to select the proper IQC procedure to implement, i.e. choosing the statistical criteria or control rules, and the number of control measurements, according to the quality required for the test and the observed performance of the method. Then the right IQC procedure must be properly implemented. This review focuses on strategies for planning and implementing IQC procedures in order to improve the quality of the IQC. A quantitative planning process is described that can be implemented with graphical tools such as power function or critical-error graphs and charts of operating specifications. Finally, a total QC strategy is formulated to minimize cost and maximize quality. A general strategy for IQC implementation is recommended that employs a three-stage design in which the first stage provides high error detection, the second stage low false rejection and the third stage prescribes the length of the analytical run, making use of an algorithm involving the average of normal patients' data.

  9. Develop Quality Characteristics Based Quality Evaluation Process for Ready to Use Software Products

    OpenAIRE

    Daiju Kato; Hiroshi Ishikawa

    2016-01-01

    The users who use ready to use software product had better get the products’ quality information with classified by some kind of global standard metrics or technique for their evaluation. But many of those software products’ co mpanies don’t provide the quality information because of the products are developed b y their own development and evaluation process. But those users want to get quality i...

  10. Quality of Service Attributes for Software as a Service

    OpenAIRE

    Lukas Burkon

    2013-01-01

    Software as a Service (SaaS) has been developing for over ten years and, is reaching a mature level, where quality and its monitoring and management become significant. Although, SaaS is derived from the ASP model, SaaS background and architecture is different and therefore also SaaS quality management is based on different concepts. This paper is focused on the difference between traditional IT outsourcing and SaaS and proposes a set of quality attributes appropriate for the management of th...

  11. Approved Air Quality Implementation Plans in Region 10

    Science.gov (United States)

    Landing page for information about EPA-approved air quality State Implementation Plans (SIPs), Tribal Implementation Plans (TIPs), and Federal Implementation Plans (FIPs) in Alaska, Idaho, Oregon, Washington.

  12. UrQt: an efficient software for the Unsupervised Quality trimming of NGS data.

    Science.gov (United States)

    Modolo, Laurent; Lerat, Emmanuelle

    2015-04-29

    Quality control is a necessary step of any Next Generation Sequencing analysis. Although customary, this step still requires manual interventions to empirically choose tuning parameters according to various quality statistics. Moreover, current quality control procedures that provide a "good quality" data set, are not optimal and discard many informative nucleotides. To address these drawbacks, we present a new quality control method, implemented in UrQt software, for Unsupervised Quality trimming of Next Generation Sequencing reads. Our trimming procedure relies on a well-defined probabilistic framework to detect the best segmentation between two segments of unreliable nucleotides, framing a segment of informative nucleotides. Our software only requires one user-friendly parameter to define the minimal quality threshold (phred score) to consider a nucleotide to be informative, which is independent of both the experiment and the quality of the data. This procedure is implemented in C++ in an efficient and parallelized software with a low memory footprint. We tested the performances of UrQt compared to the best-known trimming programs, on seven RNA and DNA sequencing experiments and demonstrated its optimality in the resulting tradeoff between the number of trimmed nucleotides and the quality objective. By finding the best segmentation to delimit a segment of good quality nucleotides, UrQt greatly increases the number of reads and of nucleotides that can be retained for a given quality objective. UrQt source files, binary executables for different operating systems and documentation are freely available (under the GPLv3) at the following address: https://lbbe.univ-lyon1.fr/-UrQt-.html .

  13. Real-time Kernel Implementation Practice Program for Embedded Software Engineers' Education and its Evaluation

    Science.gov (United States)

    Yoshida, Toshio; Matsumoto, Masahide; Seo, Katsuhiko; Chino, Shinichiro; Sugino, Eiji; Sawamoto, Jun; Koizumi, Hisao

    A real-time kernel (henceforth RTK) is in the center place of embedded software technology, and the understanding of RTK is indispensable for the embedded system design. To implement RTK, it is necessary to understand languages that describe RTK software program code, system programming manners, software development tools, CPU on that RTK runs and the interface between software and hardware, etc. in addition to understanding of RTK itself. This means RTK implementation process largely covers embedded software implementation process. Therefore, it is thought that RTK implementation practice program is very effective as a means of the acquisition of common embedded software skill in addition to deeper acquisition of RTK itself. In this paper, we propose to apply RTK implementing practice program to embedded software engineers educational program. We newly developed very small and step-up type RTK named μK for educational use, and held a seminar that used μK as a teaching material for the students of information science and engineers of the software house. As a result, we confirmed that RTK implementation practice program is very effective for the acquisition of embedded software common skill.

  14. Recommendations for a software quality assurance policy for the nuclear waste disposal risk assessment programme

    International Nuclear Information System (INIS)

    Hill, I.E.

    1985-05-01

    This study reviewed a number of published standards for software quality assurance, and included a series of interviews with software developers aimed at exploring their attitudes to software quality assurance. Recommendations for software quality assurance policy are made based on the above investigations. This document provides a summary of the recommendations made in the full report on project, reference MR-CDS-4. (author)

  15. Software tools for electrical quality assurance in the LHC

    International Nuclear Information System (INIS)

    Bednarek, M.; Ludwin, J.

    2012-01-01

    There are over 1600 superconducting magnet circuits in the LHC machine. Many of them consist of a large number of components electrically connected in series. This enhances the sensitivity of the whole circuits to electrical faults of individual components. Furthermore, circuits are equipped with a large number of instrumentation wires, which are exposed to accidental damage or swapping. In order to ensure safe operation, an Electrical Quality Assurance (ELQA) campaign is needed after each thermal cycle. Due to the complexity of the circuits, as well as their distant geographical distribution (tunnel of 27 km circumference divided in 8 sectors), suitable software and hardware platforms had to be developed. The software combines an Oracle database, LabView data acquisition applications and PHP-based web follow-up tools. This paper describes the software used for the ELQA of the LHC. (authors)

  16. Software Tools for Electrical Quality Assurance in the LHC

    CERN Document Server

    Bednarek, Mateusz

    2011-01-01

    There are over 1600 superconducting magnet circuits in the LHC machine. Many of them consist of a large number of components electrically connected in series. This enhances the sensitivity of the whole circuits to electrical faults of individual components. Furthermore, circuits are equipped with a large number of instrumentation wires, which are exposed to accidental damage or swapping. In order to ensure safe operation, an Electrical Quality Assurance (ELQA) campaign is needed after each thermal cycle. Due to the complexity of the circuits, as well as their distant geographical distribution (tunnel of 27km circumference divided in 8 sectors), suitable software and hardware platforms had to be developed. The software combines an Oracle database, LabView data acquisition applications and PHP-based web follow-up tools. This paper describes the software used for the ELQA of the LHC.

  17. A Discussion of the Software Quality Assurance Role

    Science.gov (United States)

    Kandt, Ronald Kirk

    2010-01-01

    The basic idea underlying this paper is that the conventional understanding of the role of a Software Quality Assurance (SQA) engineer is unduly limited. This is because few have asked who the customers of a SQA engineer are. Once you do this, you can better define what tasks a SQA engineer should perform, as well as identify the knowledge and skills that such a person should have. The consequence of doing this is that a SQA engineer can provide greater value to his or her customers. It is the position of this paper that a SQA engineer providing significant value to his or her customers must not only assume the role of an auditor, but also that of a software and systems engineer. This is because software engineers and their managers particularly value contributions that directly impact products and their development. These ideas are summarized as lessons learned, based on my experience at Jet Propulsion Laboratory (JPL).

  18. METHOD FOR SECURITY SPECIFICATION SOFTWARE REQUIREMENTS AS A MEANS FOR IMPLEMENTING A SOFTWARE DEVELOPMENT PROCESS SECURE - MERSEC

    Directory of Open Access Journals (Sweden)

    Castro Mecías, L.T.

    2015-06-01

    Full Text Available Often security incidents that have the object or use the software as a means of causing serious damage and legal, economic consequences, etc. Results of a survey by Kaspersky Lab reflectvulnerabilities in software are the main cause of security incidents in enterprises, the report shows that 85% of them have reported security incidents and vulnerabilities in software are the main reason is further estimated that incidents can cause significant losses estimated from 50,000 to $ 649.000. (1 In this regard academic and industry research focuses on proposals based on reducing vulnerabilities and failures of technology, with a positive influence on how the software is developed. A development process for improved safety practices and should include activities from the initial phases of the software; so that security needs are identified, manage risk and appropriate measures are implemented. This article discusses a method of analysis, acquisition and requirements specification of the software safety analysis on the basis of various proposals and deficiencies identified from participant observation in software development teams. Experiments performed using the proposed yields positive results regarding the reduction of security vulnerabilities and compliance with the safety objectives of the software.

  19. Automated Theorem Proving in High-Quality Software Design

    Science.gov (United States)

    Schumann, Johann; Swanson, Keith (Technical Monitor)

    2001-01-01

    The amount and complexity of software developed during the last few years has increased tremendously. In particular, programs are being used more and more in embedded systems (from car-brakes to plant-control). Many of these applications are safety-relevant, i.e. a malfunction of hardware or software can cause severe damage or loss. Tremendous risks are typically present in the area of aviation, (nuclear) power plants or (chemical) plant control. Here, even small problems can lead to thousands of casualties and huge financial losses. Large financial risks also exist when computer systems are used in the area of telecommunication (telephone, electronic commerce) or space exploration. Computer applications in this area are not only subject to safety considerations, but also security issues are important. All these systems must be designed and developed to guarantee high quality with respect to safety and security. Even in an industrial setting which is (or at least should be) aware of the high requirements in Software Engineering, many incidents occur. For example, the Warshaw Airbus crash, was caused by an incomplete requirements specification. Uncontrolled reuse of an Ariane 4 software module was the reason for the Ariane 5 disaster. Some recent incidents in the telecommunication area, like illegal "cloning" of smart-cards of D2GSM handies, or the extraction of (secret) passwords from German T-online users show that also in this area serious flaws can happen. Due to the inherent complexity of computer systems, most authors claim that only a rigorous application of formal methods in all stages of the software life cycle can ensure high quality of the software and lead to real safe and secure systems. In this paper, we will have a look, in how far automated theorem proving can contribute to a more widespread application of formal methods and their tools, and what automated theorem provers (ATPs) must provide in order to be useful.

  20. Quantification frameworks and their application for evaluating the software quality factor using quality characteristic value

    International Nuclear Information System (INIS)

    Kim, C.; Chung, C.H.; Won-Ahn, K.

    2004-01-01

    Many problems, related with safety, frequently occur because Digital Instrument and Control Systems are widely used and expanding their ranges to many applications in Nuclear Power Plants. It, however, does not hold a general position to estimate an appropriate software quality. Thus, the Quality Characteristic Value, a software quality factor through each software life cycle, is suggested in this paper. The Quality Characteristic Value is obtained as following procedure: 1) Scoring Quality Characteristic Factors (especially correctness, traceability, completeness, and understandability) onto Software Verification and Validation results, 2) Deriving the diamond-shaped graphs by setting values of Factors at each axis and lining every points, and lastly 3) Measuring the area of the graph for Quality Characteristic Value. In this paper, this methodology is applied to Plant Control System. In addition, the series of quantification frameworks exhibit some good characteristics in the view of software quality factor. More than any thing else, it is believed that introduced framework may be applicable to regulatory guide, software approval procedures, due to its soundness and simple characteristics. (authors)

  1. Software support for environmental measurement in quality at educational institutions

    Directory of Open Access Journals (Sweden)

    Alena Pauliková

    2016-03-01

    Full Text Available The analysed theme of this article is based on the training of environmental measurements for workplaces. This is very important for sustainable quality in technical educational institutions. Applied kinds of software, which are taught at technical educational institutions, have to offer the professional and methodical knowledge concerning conditions of working ambient for students of selected technical specialisations. This skill is performed in such a way that the graduates, after entering the practical professional life, will be able to participate in solutions for actual problems that are related to environmental protection by means of software support. Nowadays, during the training processit is also obligatory to introduce technical science. Taking into consideration the above-mentioned facts it is possible to say that information technology support for environmental study subjects is a relevant aspect, which should be integrated into the university educational process. There is an effective progress that further highlights the focus on the quality of university education not only for environmental engineers. Actual trends require an increasing number of software/hardware educated engineers who can participate in qualitative university preparation, i.e.IT environmentalists. The Department of Environmental Engineering at the Faculty of Mechanical Engineering, TechnicalUniversity in Košice, Slovakia is an institution specified and intended for quality objectivisation. This institution introduced into the study programmes (“Environmental Management” and “Technology of Environmental Protection” study subjects with the software support, which are oriented towards outdoor and indoor ambient and in this way the Department of Process and Environmental Engineering is integrated effectively and intensively into the area of measurement training with regard to the requirement of quality educational processes.

  2. An approach to software quality assurance for robotic inspection systems

    International Nuclear Information System (INIS)

    Kiebel, G.R.

    1993-10-01

    Software quality assurance (SQA) for robotic systems used in nuclear waste applications is vital to ensure that the systems operate safely and reliably and pose a minimum risk to humans and the environment. This paper describes the SQA approach for the control and data acquisition system for a robotic system being developed for remote surveillance and inspection of underground storage tanks (UST) at the Hanford Site

  3. A Software Implementation of a Satellite Interface Message Processor.

    Science.gov (United States)

    Eastwood, Margaret A.; Eastwood, Lester F., Jr.

    A design for network control software for a computer network is described in which some nodes are linked by a communications satellite channel. It is assumed that the network has an ARPANET-like configuration; that is, that specialized processors at each node are responsible for message switching and network control. The purpose of the control…

  4. Factors to Consider When Implementing Automated Software Testing

    Science.gov (United States)

    2016-11-10

    development and integration is a continuous process throughout the acquisition life cycle . Automated Software Testing can improve testing capabilities...requires a lab, conference room, or both, and whether it should be located in- house or an external facility. 2. Ensure space is adequate to support team

  5. Implementing Resource-aware Multicast Forwarding in Software Defined Networks

    DEFF Research Database (Denmark)

    Poderys, Justas; Sunny, Anjusha; Soler, José

    2018-01-01

    Using multicast data transmissions, data can be eciently distributed to a high number of network users. However, in order to ef-ciently stream multimedia using multicast communication, multicast routing protocols must have knowledge of all network links and their available bandwidth. In Software...

  6. Development and implementation of own software for dosimetry multichannel film

    International Nuclear Information System (INIS)

    Jimenez Feltstrom, D.; Reyes Garcia, R.; Luis Simon, F. J.; Carrasco Herrera, M.; Sanchez Carmona, G.; Herrador Cordoba, M.

    2013-01-01

    The objective of this work is to develop its own software for multichannel film dosimetry Radiochromic EBT2. Compare the results obtained with its use in multichannel and single-channel dosimetry. Check that the multi-channel dosimetry eliminates much of the artifacts caused by dirt, fingerprints, scratches, etc. Radiochromic in film and scanner devices. (Author)

  7. QUALITY SERVICES EVALUATION MODEL BASED ON DEDICATED SOFTWARE TOOL

    Directory of Open Access Journals (Sweden)

    ANDREEA CRISTINA IONICĂ

    2012-10-01

    Full Text Available In this paper we introduced a new model, called Service Quality (SQ, which combines QFD and SERVQUAL methods. This model takes from the SERVQUAL method the five dimensions of requirements and three of characteristics and from the QFD method the application methodology. The originality of the SQ model consists in computing a global index that reflects the customers’ requirements accomplishment level by the quality characteristics. In order to prove the viability of the SQ model, there was developed a software tool that was applied for the evaluation of a health care services provider.

  8. Quality system implementation for nuclear analytical techniques

    International Nuclear Information System (INIS)

    2004-01-01

    The international effort (UNIDO, ILAC, BIPM, etc.) to establish a functional infrastructure for metrology and accreditation in many developing countries needs to be complemented by assistance to implement high quality practices and high quality output by service providers and producers in the respective countries. Knowledge of how to approach QA systems that justify a formal accreditation is available in only a few countries and the dissemination of know how and development of skills is needed bottom up from the working level of laboratories and institutes. Awareness building, convincing of management, introduction of good management practices, technical expertise and good documentation will lead to the creation of a quality culture that assures a sustainability and inherent development of quality practices as a prerequisite of economic success. Quality assurance and quality control can be used as a valuable management tool and is a prerequisite for international trade and information exchange. This publication tries to assist quality managers, Laboratory Managers and staff involved in setting up a QA/QC system in a nuclear analytical laboratory to take appropriate action to start and complete the necessary steps for a successful quality system for ultimate national accreditation. This guidebook contributes to a better understanding of the basic ideas behind ISO/IEC 17025, the international standard for 'General requirements for the competence of testing and calibration laboratories'. It provides basic information and detailed explanation about the establishment of the QC system in analytical and nuclear analytical laboratories. It is a proper training material for training of trainers and makes managers with QC management and implementation familiar. This training material aims to facilitate the implementation of internationally accepted quality principles and to promote attempts by Member States' laboratories to obtain accreditation for nuclear analytical

  9. Hanford Tanks Initiative quality assurance implementation plan

    International Nuclear Information System (INIS)

    Huston, J.J.

    1998-01-01

    Hanford Tanks Initiative (HTI) Quality Assurance Implementation Plan for Nuclear Facilities defines the controls for the products and activities developed by HTI. Project Hanford Management Contract (PHMC) Quality Assurance Program Description (QAPD)(HNF-PRO599) is the document that defines the quality requirements for Nuclear Facilities. The QAPD provides direction for compliance to 10 CFR 830.120 Nuclear Safety Management, Quality Assurance Requirements. Hanford Tanks Initiative (HTI) is a five-year activity resulting from the technical and financial partnership of the US Department of Energy's Office of Waste Management (EM-30), and Office of Science and Technology Development (EM-50). HTI will develop and demonstrate technologies and processes for characterization and retrieval of single shell tank waste. Activities and products associated with HTI consist of engineering, construction, procurement, closure, retrieval, characterization, and safety and licensing

  10. Softcopy quality ruler method: implementation and validation

    Science.gov (United States)

    Jin, Elaine W.; Keelan, Brian W.; Chen, Junqing; Phillips, Jonathan B.; Chen, Ying

    2009-01-01

    A softcopy quality ruler method was implemented for the International Imaging Industry Association (I3A) Camera Phone Image Quality (CPIQ) Initiative. This work extends ISO 20462 Part 3 by virtue of creating reference digital images of known subjective image quality, complimenting the hardcopy Standard Reference Stimuli (SRS). The softcopy ruler method was developed using images from a Canon EOS 1Ds Mark II D-SLR digital still camera (DSC) and a Kodak P880 point-and-shoot DSC. Images were viewed on an Apple 30in Cinema Display at a viewing distance of 34 inches. Ruler images were made for 16 scenes. Thirty ruler images were generated for each scene, representing ISO 20462 Standard Quality Scale (SQS) values of approximately 2 to 31 at an increment of one just noticeable difference (JND) by adjusting the system modulation transfer function (MTF). A Matlab GUI was developed to display the ruler and test images side-by-side with a user-adjustable ruler level controlled by a slider. A validation study was performed at Kodak, Vista Point Technology, and Aptina Imaging in which all three companies set up a similar viewing lab to run the softcopy ruler method. The results show that the three sets of data are in reasonable agreement with each other, with the differences within the range expected from observer variability. Compared to previous implementations of the quality ruler, the slider-based user interface allows approximately 2x faster assessments with 21.6% better precision.

  11. Development and implementation of a software for assessment of calculations in radiotherapy monitor units

    International Nuclear Information System (INIS)

    Mueller, Marcio Rogerio; Silva, Marco Antonio da; Rodrigues, Laura Natal

    2005-01-01

    Objective: To develop and implement a software based on the manual calculation algorithm for verification of the calculations in radiotherapy monitor units and to establish acceptance levels as a mechanism of quality assurance. Materials and methods: Data were obtained from Clinac 600C and 2100C Varian linear accelerators and the computerized treatment planning system used was CadPlan TM . Results: For the 6 MV beams the acceptance levels for deviations among the calculations of monitor units, divided by treatment area, were the following: breast (0.0%±1.7%), head and neck (1.5%±0.5%), hypophysis (-1.7%±0.5%), pelvis (2.1%±2.1%) and thorax (0.2%±1.3%). For the 15 MV beams the suggested level for pelvis in all the treatment techniques was (3.2%±1.3%). Conclusion: The present data are sufficient to justify the use of the software in the clinical practice as a tool for the quality assurance program. (author)

  12. Design and Implementation of Software for the ROV Neptunus

    OpenAIRE

    Munz, Jostein

    2015-01-01

    There has been an increased interest in researching and developing technical solutions for underwater vehicles the last years. In most cases, underwater vehicles are expensive and not affordable to private individuals. Development of low cost ROVs has lately been increasing supported by an active Do-It-Yourself (DIY) community. The OpenROV project has contributed to this community by producing cheap ROVs with open source software, encouraging users to develop their own plugins. Development of...

  13. EARLY IDENTIFICATION OF SOFTWARE DEFECTS USING OCL PREDICATES TO IMPROVE SOFTWARE QUALITY

    Directory of Open Access Journals (Sweden)

    A. JALILA

    2015-03-01

    Full Text Available Formal specification-based testing has been used widely to assess potential faults or prove their absence in a given system at the earliest. This research work has proposed an automated fault-based testing framework to test the specification of the system using Object Constraint Language (OCL. Accordingly, the possible faults in OCL specification of the system has been anticipated by mutating its method based on OCL predicate-based fault classes. Then, test cases are generated using Genetic Algorithm with simulated annealing technique. In this paper, a novel OCL-predicate based fitness function is defined to evaluate the generated test data. Finally, this paper presents the experimental results, which indicate that the proposed methodology provides more test coverage with the reduced test suite and test run. This results in cost-effective software development so as to improve software quality.

  14. Design, Implementation, and Performance of CREAM Data Acquisition Software

    CERN Document Server

    Zinn, S Y; Bagliesi, M G; Beatty, J J; Childers, J T; Coutu, S; Duvernois, M A; Ganel, O; Kim, H J; Lee, M H; Lutz, L; Malinine, A; Maestro, P; Marrocchesi, P S; Park, I H; Seo, E S; Song, C; Swordy, S; Wu, J

    2005-01-01

    Cosmic Ray Energetics and Mass (CREAM) is a balloon-borne experiment scheduled for launching from Antarctica in late 2004. Its aim is to measure the energy spectrum and composition of cosmic rays from proton to iron nuclei at ultra high energies from 1 to 1,000 TeV. Ultra long duration balloons are expected to fly about 100 days. One special feature of the CREAM data acquisition software (CDAQ) is the telemetric operation of the instrument using satellites. During a flight the science event and housekeeping data are sent from the instrument to a ground facility. Likewise, commands for controlling both the hardware and the software are uploaded from the ground facility. This requires a robust, reliable, and fast software system. CDAQ has been developed and tested during three beam tests at CERN in July, September, and November 2003. Recently the interfaces to the transition radiation detector (TRD) and to the timing-based charge detector (TCD) have been added. These new additions to CDAQ will be checked at a t...

  15. Software quality for 1997 - what works and what doesn`t?

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Burlington, MA (United States)

    1997-11-01

    This presentation provides a view of software quality for 1997 - what works and what doesn`t. For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Development (WFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels.

  16. Quality assurance (QA) procedures for software: Evaluation of an ADC quality system

    International Nuclear Information System (INIS)

    Efstathopoulos, E. P.; Benekos, O.; Molfetas, M.; Charou, E.; Kottou, S.; Argentos, S.; Kelekis, N. L.

    2005-01-01

    Image viewing and processing software in computed radiography manipulates image contrast in such a way that all relevant image features are rendered to an appropriate degree of visibility, and improves image quality using enhancement algorithms. The purpose of this study was to investigate procedures for the quality assessment of image processing software for computed radiography with the use of existing test objects and to assess the influence that processing introduces on physical image quality characteristics. Measurements of high-contrast resolution, low-contrast resolution, spatial resolution, grey scale (characteristic curve) and geometric distortion were performed 'subjectively' by three independent observers and 'objectively' by the use of criteria based on pixel intensity values. Results show quality assessment is possible without the need for human evaluators, using digital images. It was discovered that the processing software evaluated in this study was able to improve some aspects of image quality, without introducing geometric distortion. (authors)

  17. The influence of software filtering in digital mammography image quality

    Science.gov (United States)

    Michail, C.; Spyropoulou, V.; Kalyvas, N.; Valais, I.; Dimitropoulos, N.; Fountos, G.; Kandarakis, I.; Panayiotakis, G.

    2009-05-01

    Breast cancer is one of the most frequently diagnosed cancers among women. Several techniques have been developed to help in the early detection of breast cancer such as conventional and digital x-ray mammography, positron and single-photon emission mammography, etc. A key advantage in digital mammography is that images can be manipulated as simple computer image files. Thus non-dedicated commercially available image manipulation software can be employed to process and store the images. The image processing tools of the Photoshop (CS 2) software usually incorporate digital filters which may be used to reduce image noise, enhance contrast and increase spatial resolution. However, improving an image quality parameter may result in degradation of another. The aim of this work was to investigate the influence of three sharpening filters, named hereafter sharpen, sharpen more and sharpen edges on image resolution and noise. Image resolution was assessed by means of the Modulation Transfer Function (MTF).In conclusion it was found that the correct use of commercial non-dedicated software on digital mammograms may improve some aspects of image quality.

  18. The influence of software filtering in digital mammography image quality

    International Nuclear Information System (INIS)

    Michail, C; Spyropoulou, V; Valais, I; Panayiotakis, G; Kalyvas, N; Fountos, G; Kandarakis, I; Dimitropoulos, N

    2009-01-01

    Breast cancer is one of the most frequently diagnosed cancers among women. Several techniques have been developed to help in the early detection of breast cancer such as conventional and digital x-ray mammography, positron and single-photon emission mammography, etc. A key advantage in digital mammography is that images can be manipulated as simple computer image files. Thus non-dedicated commercially available image manipulation software can be employed to process and store the images. The image processing tools of the Photoshop (CS 2) software usually incorporate digital filters which may be used to reduce image noise, enhance contrast and increase spatial resolution. However, improving an image quality parameter may result in degradation of another. The aim of this work was to investigate the influence of three sharpening filters, named hereafter sharpen, sharpen more and sharpen edges on image resolution and noise. Image resolution was assessed by means of the Modulation Transfer Function (MTF).In conclusion it was found that the correct use of commercial non-dedicated software on digital mammograms may improve some aspects of image quality.

  19. Quality factors in the life cycle of software oriented to safety systems in nuclear power plants

    International Nuclear Information System (INIS)

    Nunez McLeod, J.E.; Rivera, S.S.

    1997-01-01

    The inclusion of software in safety related systems for nuclear power plants, makes it necessary to include the software quality assurance concept. The software quality can be defined as the adjustment degree between the software and the specified requirements and user expectations. To guarantee a certain software quality level it is necessary to make a systematic and planned set of tasks, that constitute a software quality guaranty plan. The application of such a plan involves activities that should be performed all along the software life cycle, and that can be evaluated through the so called quality factors, due to the fact that the quality itself cannot be directly measured, but indirectly as some of it manifestations. In this work, a software life cycle model is proposed, for nuclear power plant safety related systems. A set os software quality factors is also proposed , with its corresponding classification according to the proposed model. (author) [es

  20. Global Software Engineering: A Software Process Approach

    Science.gov (United States)

    Richardson, Ita; Casey, Valentine; Burton, John; McCaffery, Fergal

    Our research has shown that many companies are struggling with the successful implementation of global software engineering, due to temporal, cultural and geographical distance, which causes a range of factors to come into play. For example, cultural, project managementproject management and communication difficulties continually cause problems for software engineers and project managers. While the implementation of efficient software processes can be used to improve the quality of the software product, published software process models do not cater explicitly for the recent growth in global software engineering. Our thesis is that global software engineering factors should be included in software process models to ensure their continued usefulness in global organisations. Based on extensive global software engineering research, we have developed a software process, Global Teaming, which includes specific practices and sub-practices. The purpose is to ensure that requirements for successful global software engineering are stipulated so that organisations can ensure successful implementation of global software engineering.

  1. E-learning Materials Development: Applying and Implementing Software Reuse Principles and Granularity Levels in the Small

    OpenAIRE

    Nabil Arman

    2010-01-01

    E-learning materials development is typically acknowledged as an expensive, complicated, and lengthy process, often producing materials that are of low quality and difficult to adaptand maintain. It has always been a challenge to identify proper e-learning materials that can be reused at a reasonable cost and effort. In this paper, software engineering reuse principlesare applied to e-learning materials development process. These principles are then applied and implemented in a prototype that...

  2. An Analysis of Earned Value Management Implementation Within the F-22 System Program Office's Software Development

    National Research Council Canada - National Science Library

    Dibert, John C; Velez, John C

    2006-01-01

    .... Acquisition Category ID programs like the US Air Force F-22 fighter program use EVM to manage their software development efforts, but has the program's implementation of EVM followed the industry...

  3. Freight Advanced Traveler Information System (FRATIS) Dallas-Fort Worth : software architecture design and implementation options.

    Science.gov (United States)

    2013-05-01

    This document describes the Software Architecture Design and Implementation Options for FRATIS : system. The demonstration component of this task will serve to test the technical feasibility of the : FRATIS prototype while also facilitating the colle...

  4. Software development to implement the TxDOT culvert rating guide.

    Science.gov (United States)

    2013-05-01

    This implementation project created CULVLR: Culvert Load Rating, Version 1.0.0, a Windows-based : desktop application software package that automates the process by which Texas Department of Transportation : (TxDOT) engineers and their consultants ...

  5. Opening Up Architectures of Software-Intensive Systems: A First Prototype Implementation

    National Research Council Canada - National Science Library

    Charland, Philippe; Dessureault, Dany; Ouellet, David; Lizotte, Michel

    2007-01-01

    .... One approach to deal with this problem is information hiding. This technical memorandum presents a prototype which implements this technique to reverse engineer dynamic models from Java software systems...

  6. Maximize Your Investment 10 Key Strategies for Effective Packaged Software Implementations

    CERN Document Server

    Beaubouef, Grady Brett

    2009-01-01

    This is a handbook covering ten principles for packaged software implementations that project managers, business owners, and IT developers should pay attention to. The book also has practical real-world coverage including a sample agenda for conducting business solution modeling, customer case studies, and a road map to implement guiding principles. This book is aimed at enterprise architects, development leads, project managers, business systems analysts, business systems owners, and anyone who wants to implement packaged software effectively. If you are a customer looking to implement COTS s

  7. Software for improving the quality of project management, a case study: international manufacture of electrical equipment

    Science.gov (United States)

    Preradović, D. M.; Mićić, Lj S.; Barz, C.

    2017-05-01

    Production conditions in today’s world require software support at every stage of production and development of new products, for quality assurance and compliance with ISO standards. In addition to ISO standards such as usual metrics of quality, companies today are focused on other optional standards, such as CMMI (Capability Maturity Model Integrated) or prescribing they own standards. However, while there is intensive progress being made in the PM (project management), there is still a significant number of projects, at the global level, that are failures. These have failed to achieve their goals, within budget or timeframe. This paper focuses on checking the role of software tools through the rate of success in projects implemented in the case of internationally manufactured electrical equipment. The results of this research show the level of contribution of the project management software used to manage and develop new products to improve PM processes and PM functions, and how selection of the software tools affects the quality of PM processes and successfully completed projects.

  8. Implementing an composition architecture for an online game software

    Directory of Open Access Journals (Sweden)

    Jaime Alberto Guzmán Luna

    2011-01-01

    Full Text Available Este artículo presenta una arquitectura para la composición de servicios Web, en la que un plan de composición se construye sobre la base de un agente de planificación, que puede ser ejecutado en forma concurrente durante su composición a fin de estimar la siguiente acción a ejecutar en lugar de preparar un completo plan que con frecuencia será invalidado. Esta característica es muy valiosa cuando se trata de abordar los problemas en tiempo real. Específicamente se propone como un dominio de prueba un software de juegos en línea, llamado ENVIRO.

  9. Laboratory quality stepwise implementation tool: National reference TB laboratory of Iran

    OpenAIRE

    Ali Naghi Kebriaee; Donya Malekshahian; Mojtaba Ahmadi; Parissa Farnia

    2015-01-01

    Background and objective: During recent years, the World Health Organization (WHO) proposed new software for improving the tuberculosis (TB) laboratory services. The protocol is known as “quality stepwise implementation tool” and is based on enforcement of quality assurance services through accreditation by the International Organization for Standardization (ISO) 15189. As a national reference TB laboratory (NRL) of Iran, the benefit and challenges of implementing this standard were analyzed....

  10. Sci-Thur AM: Planning - 04: Evaluation of the fluence complexity, solution quality, and run efficiency produced by five fluence parameterizations implemented in PARETO multiobjective radiotherapy treatment planning software.

    Science.gov (United States)

    Champion, H; Fiege, J; McCurdy, B; Potrebko, P; Cull, A

    2012-07-01

    PARETO (Pareto-Aware Radiotherapy Evolutionary Treatment Optimization) is a novel multiobjective treatment planning system that performs beam orientation and fluence optimization simultaneously using an advanced evolutionary algorithm. In order to reduce the number of parameters involved in this enormous search space, we present several methods for modeling the beam fluence. The parameterizations are compared using innovative tools that evaluate fluence complexity, solution quality, and run efficiency. A PARETO run is performed using the basic weight (BW), linear gradient (LG), cosine transform (CT), beam group (BG), and isodose-projection (IP) methods for applying fluence modulation over the projection of the Planning Target Volume in the beam's-eye-view plane. The solutions of each run are non-dominated with respect to other trial solutions encountered during the run. However, to compare the solution quality of independent runs, each run competes against every other run in a round robin fashion. Score is assigned based on the fraction of solutions that survive when a tournament selection operator is applied to the solutions of the two competitors. To compare fluence complexity, a modulation index, fractal dimension, and image gradient entropy are calculated for the fluence maps of each optimal plan. We have found that the LG method results in superior solution quality for a spine phantom, lung patient, and cauda equina patient. The BG method produces solutions with the highest degree of fluence complexity. Most methods result in comparable run times. The LG method produces superior solution quality using a moderate degree of fluence modulation. © 2012 American Association of Physicists in Medicine.

  11. Development of Software to Model AXAF-I Image Quality

    Science.gov (United States)

    Ahmad, Anees; Hawkins, Lamar

    1996-01-01

    This draft final report describes the work performed under the delivery order number 145 from May 1995 through August 1996. The scope of work included a number of software development tasks for the performance modeling of AXAF-I. A number of new capabilities and functions have been added to the GT software, which is the command mode version of the GRAZTRACE software, originally developed by MSFC. A structural data interface has been developed for the EAL (old SPAR) finite element analysis FEA program, which is being used by MSFC Structural Analysis group for the analysis of AXAF-I. This interface utility can read the structural deformation file from the EAL and other finite element analysis programs such as NASTRAN and COSMOS/M, and convert the data to a suitable format that can be used for the deformation ray-tracing to predict the image quality for a distorted mirror. There is a provision in this utility to expand the data from finite element models assuming 180 degrees symmetry. This utility has been used to predict image characteristics for the AXAF-I HRMA, when subjected to gravity effects in the horizontal x-ray ground test configuration. The development of the metrology data processing interface software has also been completed. It can read the HDOS FITS format surface map files, manipulate and filter the metrology data, and produce a deformation file, which can be used by GT for ray tracing for the mirror surface figure errors. This utility has been used to determine the optimum alignment (axial spacing and clocking) for the four pairs of AXAF-I mirrors. Based on this optimized alignment, the geometric images and effective focal lengths for the as built mirrors were predicted to cross check the results obtained by Kodak.

  12. Development and case study of a science-based software platform to support policy making on air quality.

    Science.gov (United States)

    Zhu, Yun; Lao, Yanwen; Jang, Carey; Lin, Chen-Jen; Xing, Jia; Wang, Shuxiao; Fu, Joshua S; Deng, Shuang; Xie, Junping; Long, Shicheng

    2015-01-01

    This article describes the development and implementations of a novel software platform that supports real-time, science-based policy making on air quality through a user-friendly interface. The software, RSM-VAT, uses a response surface modeling (RSM) methodology and serves as a visualization and analysis tool (VAT) for three-dimensional air quality data obtained by atmospheric models. The software features a number of powerful and intuitive data visualization functions for illustrating the complex nonlinear relationship between emission reductions and air quality benefits. The case study of contiguous U.S. demonstrates that the enhanced RSM-VAT is capable of reproducing the air quality model results with Normalized Mean Bias quality policy making in near real time. Copyright © 2014. Published by Elsevier B.V.

  13. Software quality assurance in the 1996 performance assessment for the Waste Isolation Pilot Plant

    International Nuclear Information System (INIS)

    Froehlich, Gary K.; Ogden, Harvey C.; Byle, Kathleen A.

    2000-01-01

    The US Department of Energy (DOE) Waste Isolation Pilot Plant (WIPP), located in southeast New Mexico, is a deep geologic repository for the permanent disposal of transuranic waste generated by DOE defense-related activities. Sandia National Laboratories (SNL), in its role as scientific advisor to the DOE, is responsible for evaluating the long-term performance of the WIPP. This risk-based Performance Assessment (PA) is accomplished in part through the use of numerous scientific modeling codes, which rely for some of their inputs on data gathered during characterization of the site. The PA is subject to formal requirements set forth in federal regulations. In particular, the components of the calculation fall under the configuration management and software quality assurance aegis of the American Society of Mechanical Engineers(ASME) Nuclear Quality Assurance (NQA) requirements. This paper describes SNL's implementation of the NQA requirements regarding software quality assurance (SQA). The description of the implementation of SQA for a PA calculation addresses not only the interpretation of the NQA requirements, it also discusses roles, deliverables, and the resources necessary for effective implementation. Finally, examples are given which illustrate the effectiveness of SNL's SQA program, followed by a detailed discussion of lessons learned

  14. Software quality assurance in the 1996 performance assessment for the Waste Isolation Pilot Plant

    International Nuclear Information System (INIS)

    Froehlich, G.K.; Ogden, H.C.; Byle, K.A.

    2000-01-01

    The US Department of Energy (DOE) Waste Isolation Pilot Plant (WIPP), located in southeast New Mexico, is a deep geologic repository for the permanent disposal of transuranic waste generated by DOE defense-related activities. Sandia National Laboratories (SNL), in its role as scientific advisor to the DOE, is responsible for evaluating the long-term performance of the WIPP. This risk-based Performance Assessment (PA) is accomplished in part through the use of numerous scientific modeling codes, which rely for some of their inputs on data gathered during characterization of the site. The PA is subject to formal requirements set forth in federal regulations. In particular, the components of the calculation fall under the configuration management and software quality assurance aegis of the American Society of Mechanical Engineers (ASME) Nuclear Quality Assurance (NQA) requirements. This paper describes SNL's implementation of the NQA requirements regarding software quality assurance (SQA). The description of the implementation of SQA for a PA calculation addresses not only the interpretation of the NQA requirements, it also discusses roles, deliverables, and the resources necessary for effective implementation. Finally, examples are given which illustrate the effectiveness of SNL's SQA program, followed by a detailed discussion of lessons learned

  15. Maintaining Quality and Confidence in Open-Source, Evolving Software: Lessons Learned with PFLOTRAN

    Science.gov (United States)

    Frederick, J. M.; Hammond, G. E.

    2017-12-01

    Software evolution in an open-source framework poses a major challenge to a geoscientific simulator, but when properly managed, the pay-off can be enormous for both the developers and the community at large. Developers must juggle implementing new scientific process models, adopting increasingly efficient numerical methods and programming paradigms, changing funding sources (or total lack of funding), while also ensuring that legacy code remains functional and reported bugs are fixed in a timely manner. With robust software engineering and a plan for long-term maintenance, a simulator can evolve over time incorporating and leveraging many advances in the computational and domain sciences. In this positive light, what practices in software engineering and code maintenance can be employed within open-source development to maximize the positive aspects of software evolution and community contributions while minimizing its negative side effects? This presentation will discusses steps taken in the development of PFLOTRAN (www.pflotran.org), an open source, massively parallel subsurface simulator for multiphase, multicomponent, and multiscale reactive flow and transport processes in porous media. As PFLOTRAN's user base and development team continues to grow, it has become increasingly important to implement strategies which ensure sustainable software development while maintaining software quality and community confidence. In this presentation, we will share our experiences and "lessons learned" within the context of our open-source development framework and community engagement efforts. Topics discussed will include how we've leveraged both standard software engineering principles, such as coding standards, version control, and automated testing, as well unique advantages of object-oriented design in process model coupling, to ensure software quality and confidence. We will also be prepared to discuss the major challenges faced by most open-source software teams, such

  16. Effectiveness of Software Quality Assurance in Offshore Development Enterprises in Sri Lanka

    OpenAIRE

    Malinda G. Sirisena

    2014-01-01

    The aim of this research is to evaluate the effectiveness of software quality assurance approaches of Sri Lankan offshore software development organizations, and to propose a framework which could be used across all offshore software development organizations. An empirical study was conducted using derived framework from popular software quality evaluation models. The research instrument employed was a questionnaire survey among thirty seven Sri Lankan registered offshore software develop...

  17. Automated Improvement of Software Architecture Models for Performance and Other Quality Attributes

    OpenAIRE

    Koziolek, Anne

    2013-01-01

    Quality attributes, such as performance or reliability, are crucial for the success of a software system and largely influenced by the software architecture. Their quantitative prediction supports systematic, goal-oriented software design and forms a base of an engineering approach to software design. This thesis proposes a method and tool to automatically improve component-based software architecture (CBA) models based on such quantitative quality prediction techniques.

  18. Evaluation of Software Quality to Improve Application Performance Using Mc Call Model

    Directory of Open Access Journals (Sweden)

    Inda D Lestantri

    2018-04-01

    Full Text Available The existence of software should have more value to improve the performance of the organization in addition to having the primary function to automate. Before being implemented in an operational environment, software must pass the test gradually to ensure that the software is functioning properly, meeting user needs and providing convenience for users to use it. This test is performed on a web-based application, by taking a test case in an e-SAP application. E-SAP is an application used to monitor teaching and learning activities used by a university in Jakarta. To measure software quality, testing can be done on users randomly. The user samples selected in this test are users with an age range of 18 years old up to 25 years, background information technology. This test was conducted on 30 respondents. This test is done by using Mc Call model. Model of testing Mc Call consists of 11 dimensions are grouped into 3 categories. This paper describes the testing with reference to the category of product operation, which includes 5 dimensions. The dimensions of testing performed include the dimensions of correctness, usability, efficiency, reliability, and integrity. This paper discusses testing on each dimension to measure software quality as an effort to improve performance. The result of research is e-SAP application has good quality with product operation value equal to 85.09%. This indicates that the e-SAP application has a great quality, so this application deserves to be examined in the next stage on the operational environment.

  19. Analysis of quality raw data of second generation sequencers with Quality Assessment Software.

    Science.gov (United States)

    Ramos, Rommel Tj; Carneiro, Adriana R; Baumbach, Jan; Azevedo, Vasco; Schneider, Maria Pc; Silva, Artur

    2011-04-18

    Second generation technologies have advantages over Sanger; however, they have resulted in new challenges for the genome construction process, especially because of the small size of the reads, despite the high degree of coverage. Independent of the program chosen for the construction process, DNA sequences are superimposed, based on identity, to extend the reads, generating contigs; mismatches indicate a lack of homology and are not included. This process improves our confidence in the sequences that are generated. We developed Quality Assessment Software, with which one can review graphs showing the distribution of quality values from the sequencing reads. This software allow us to adopt more stringent quality standards for sequence data, based on quality-graph analysis and estimated coverage after applying the quality filter, providing acceptable sequence coverage for genome construction from short reads. Quality filtering is a fundamental step in the process of constructing genomes, as it reduces the frequency of incorrect alignments that are caused by measuring errors, which can occur during the construction process due to the size of the reads, provoking misassemblies. Application of quality filters to sequence data, using the software Quality Assessment, along with graphing analyses, provided greater precision in the definition of cutoff parameters, which increased the accuracy of genome construction.

  20. Using iKidTools™ Software Support Systems to Develop and Implement Self-Monitoring Interventions

    Science.gov (United States)

    Patti, Angela L.; Miller, Kevin J.

    2011-01-01

    Educational teams often are faced with the task of developing and implementing Behavioral Intervention Plans (BIPs) for students who present challenging and/or disruptive behaviors. This article describes the steps used to develop and implement a self-monitoring BIP that incorporated an innovative software system, iKidTools™. An authentic case…

  1. Development and implementation of a 'Mental Health Finder' software tool within an electronic medical record system.

    Science.gov (United States)

    Swan, D; Hannigan, A; Higgins, S; McDonnell, R; Meagher, D; Cullen, W

    2017-02-01

    In Ireland, as in many other healthcare systems, mental health service provision is being reconfigured with a move toward more care in the community, and particularly primary care. Recording and surveillance systems for mental health information and activities in primary care are needed for service planning and quality improvement. We describe the development and initial implementation of a software tool ('mental health finder') within a widely used primary care electronic medical record system (EMR) in Ireland to enable large-scale data collection on the epidemiology and management of mental health and substance use problems among patients attending general practice. In collaboration with the Irish Primary Care Research Network (IPCRN), we developed the 'Mental Health Finder' as a software plug-in to a commonly used primary care EMR system to facilitate data collection on mental health diagnoses and pharmacological treatments among patients. The finder searches for and identifies patients based on diagnostic coding and/or prescribed medicines. It was initially implemented among a convenience sample of six GP practices. Prevalence of mental health and substance use problems across the six practices, as identified by the finder, was 9.4% (range 6.9-12.7%). 61.9% of identified patients were female; 25.8% were private patients. One-third (33.4%) of identified patients were prescribed more than one class of psychotropic medication. Of the patients identified by the finder, 89.9% were identifiable via prescribing data, 23.7% via diagnostic coding. The finder is a feasible and promising methodology for large-scale data collection on mental health problems in primary care.

  2. Software Quality in the Trenches: Two Case Studies of Quality Assurance Practices in Free/Libre and Open Source Software (FLOSS)

    OpenAIRE

    Vestbø, Tor Arne

    2007-01-01

    When proponents of open source software are asked to explain the success of their movement they typically point to the quality of the software produced, which is in turn attributed to the rather unconventional development model of releasing unfinished versions of the software and having users look over the code and report and fix bugs. This thesis investigates the open source quality assurance model from a knowledge management perspective based on the assumption that debugging involves a h...

  3. A survey of quality assurance practices in biomedical open source software projects.

    Science.gov (United States)

    Koru, Günes; El Emam, Khaled; Neisa, Angelica; Umarji, Medha

    2007-05-07

    Open source (OS) software is continuously gaining recognition and use in the biomedical domain, for example, in health informatics and bioinformatics. Given the mission critical nature of applications in this domain and their potential impact on patient safety, it is important to understand to what degree and how effectively biomedical OS developers perform standard quality assurance (QA) activities such as peer reviews and testing. This would allow the users of biomedical OS software to better understand the quality risks, if any, and the developers to identify process improvement opportunities to produce higher quality software. A survey of developers working on biomedical OS projects was conducted to examine the QA activities that are performed. We took a descriptive approach to summarize the implementation of QA activities and then examined some of the factors that may be related to the implementation of such practices. Our descriptive results show that 63% (95% CI, 54-72) of projects did not include peer reviews in their development process, while 82% (95% CI, 75-89) did include testing. Approximately 74% (95% CI, 67-81) of developers did not have a background in computing, 80% (95% CI, 74-87) were paid for their contributions to the project, and 52% (95% CI, 43-60) had PhDs. A multivariate logistic regression model to predict the implementation of peer reviews was not significant (likelihood ratio test = 16.86, 9 df, P = .051) and neither was a model to predict the implementation of testing (likelihood ratio test = 3.34, 9 df, P = .95). Less attention is paid to peer review than testing. However, the former is a complementary, and necessary, QA practice rather than an alternative. Therefore, one can argue that there are quality risks, at least at this point in time, in transitioning biomedical OS software into any critical settings that may have operational, financial, or safety implications. Developers of biomedical OS applications should invest more effort

  4. Contribuição dos modelos de qualidade e maturidade na melhoria dos processos de software Contribution of quality and maturity models to software process improvement

    Directory of Open Access Journals (Sweden)

    Antonio Carlos Tonini

    2008-01-01

    Full Text Available Grande parte das empresas desenvolvedoras de software criou seu próprio processo de trabalho. Devido à rápida expansão do mercado de software, a concorrência ocorre muito mais em custo do que em diferenciação. Para obter vantagem competitiva, as empresas devem atualizar-se continuamente na tecnologia, buscar a maturidade nos processos e eliminar a ineficiência operacional. Isso requer um envolvimento das pessoas, dos processos e da organização como um todo. O artigo discute a implementação de melhorias nos processos de software segundo os principais modelos de qualidade e de maturidade. Com base em um Estudo de Casos Múltiplos, verifica-se que a melhoria dos processos de software requer que a melhoria ocorra primeiramente entre cada um dos desenvolvedores e, a seguir, envolva os grupos de desenvolvimento e por fim, a organização como um todo. A pesquisa conclui que os modelos de qualidade e maturidade servem como orientadores do processo de melhoria.Many software development companies have developed their own work method. Due to the fast software market growth, the competition focuses more on cost than on differentiation. To achieve competitive advantage, software developer organizations must continually update their technology, reach high level process maturity and eliminate all the operational inefficiency. These procedures involve people, processes and the whole organization. The aim of the paper is to discuss software process improvement implementation according to the most important quality and maturity models. Based on a Multiple Case Study, it is verified that the software process improvement needs firstly individual improvement and, later, it involves the developer teams and the whole organization. The research concludes that the quality and maturity models must be used as improvement process drivers.

  5. Improving the quality of care of patients with rheumatic disease using patient-centric electronic redesign software.

    Science.gov (United States)

    Newman, Eric D; Lerch, Virginia; Billet, Jon; Berger, Andrea; Kirchner, H Lester

    2015-04-01

    Electronic health records (EHRs) are not optimized for chronic disease management. To improve the quality of care for patients with rheumatic disease, we developed electronic data capture, aggregation, display, and documentation software. The software integrated and reassembled information from the patient (via a touchscreen questionnaire), nurse, physician, and EHR into a series of actionable views. Core functions included trends over time, rheumatology-related demographics, and documentation for patient and provider. Quality measures collected included patient-reported outcomes, disease activity, and function. The software was tested and implemented in 3 rheumatology departments, and integrated into routine care delivery. Post-implementation evaluation measured adoption, efficiency, productivity, and patient perception. Over 2 years, 6,725 patients completed 19,786 touchscreen questionnaires. The software was adopted for use by 86% of patients and rheumatologists. Chart review and documentation time trended downward, and productivity increased by 26%. Patient satisfaction, activation, and adherence remained unchanged, although pre-implementation values were high. A strong correlation was seen between use of the software and disease control (weighted Pearson's correlation coefficient 0.5927, P = 0.0095), and a relative increase in patients with low disease activity of 3% per quarter was noted. We describe innovative software that aggregates, stores, and displays information vital to improving the quality of care for patients with chronic rheumatic disease. The software was well-adopted by patients and providers. Post-implementation, significant improvements in quality of care, efficiency of care, and productivity were demonstrated. Copyright © 2015 by the American College of Rheumatology.

  6. A software toolkit for implementing low-cost virtual reality training systems

    International Nuclear Information System (INIS)

    Louka, Michael N.

    1999-04-01

    VR is a powerful technology for implementing training systems but better tools are needed to achieve wider usage and acceptance for desktop computer-based training applications. A need has been identified for a software tool kit to support the efficient implementation of well-structured desktop VR training systems. A powerful toolkit for implementing scalable low-cost VR training applications is described in this report (author) (ml)

  7. Comparison of quality control software tools for diffusion tensor imaging.

    Science.gov (United States)

    Liu, Bilan; Zhu, Tong; Zhong, Jianhui

    2015-04-01

    Image quality of diffusion tensor imaging (DTI) is critical for image interpretation, diagnostic accuracy and efficiency. However, DTI is susceptible to numerous detrimental artifacts that may impair the reliability and validity of the obtained data. Although many quality control (QC) software tools are being developed and are widely used and each has its different tradeoffs, there is still no general agreement on an image quality control routine for DTIs, and the practical impact of these tradeoffs is not well studied. An objective comparison that identifies the pros and cons of each of the QC tools will be helpful for the users to make the best choice among tools for specific DTI applications. This study aims to quantitatively compare the effectiveness of three popular QC tools including DTI studio (Johns Hopkins University), DTIprep (University of North Carolina at Chapel Hill, University of Iowa and University of Utah) and TORTOISE (National Institute of Health). Both synthetic and in vivo human brain data were used to quantify adverse effects of major DTI artifacts to tensor calculation as well as the effectiveness of different QC tools in identifying and correcting these artifacts. The technical basis of each tool was discussed, and the ways in which particular techniques affect the output of each of the tools were analyzed. The different functions and I/O formats that three QC tools provide for building a general DTI processing pipeline and integration with other popular image processing tools were also discussed. Copyright © 2015 Elsevier Inc. All rights reserved.

  8. Software quality assurance on the Yucca Mountain Site Characterization Project

    International Nuclear Information System (INIS)

    Matras, J.R.

    1993-01-01

    The Yucca Mountain Site Characterization Project (YMP) has been involved over the years in the continuing struggle with establishing acceptable Software Quality Assurance (SQA) requirements for the development, modification, and acquisition of computer programs used to support the Mined Geologic Disposal System. These computer programs will be used to produce or manipulate data used directly in site characterization, design, analysis, performance assessment, and operation of repository structures, systems, and components. Scientists and engineers working on the project have claimed that the SQA requirements adopted by the project are too restrictive to allow them to perform their work. This paper will identify the source of the original SQA requirements adopted by the project. It will delineate the approach used by the project to identify concerns voiced by project engineers and scientists regarding the original SQA requirements. It will conclude with a discussion of methods used to address these problems in the rewrite of the original SQA requirements

  9. Implementing a modeling software for animated protein-complex interactions using a physics simulation library.

    Science.gov (United States)

    Ueno, Yutaka; Ito, Shuntaro; Konagaya, Akihiko

    2014-12-01

    To better understand the behaviors and structural dynamics of proteins within a cell, novel software tools are being developed that can create molecular animations based on the findings of structural biology. This study proposes our method developed based on our prototypes to detect collisions and examine the soft-body dynamics of molecular models. The code was implemented with a software development toolkit for rigid-body dynamics simulation and a three-dimensional graphics library. The essential functions of the target software system included the basic molecular modeling environment, collision detection in the molecular models, and physical simulations of the movement of the model. Taking advantage of recent software technologies such as physics simulation modules and interpreted scripting language, the functions required for accurate and meaningful molecular animation were implemented efficiently.

  10. Quality of Design, Analysis and Reporting of Software Engineering Experiments:A Systematic Review

    OpenAIRE

    By Kampenes, Vigdis

    2007-01-01

    Background: Like any research discipline, software engineering research must be of a certain quality to be valuable. High quality research in software engineering ensures that knowledge is accumulated and helpful advice is given to the industry. One way of assessing research quality is to conduct systematic reviews of the published research literature. Objective: The purpose of this work was to assess the quality of published experiments in software engineering with respect to the validit...

  11. SOFTWARE TRAINING AIDS DEVELOPMENT AND IMPLEMENTATION IN PROFESSIONAL PREPARATION PRACTICE OF TECHNOLOGICAL EDUCATION TEACHER

    Directory of Open Access Journals (Sweden)

    Anatoliy G. Gritchenko

    2013-03-01

    Full Text Available The article outlines the theoretical and practical aspects of software training aids development and implementation in professional preparation practice of technological education teacher. The myriad opportunities of new information technologies are described; the characteristic features of modern software training tool (STT are revealed; the main algorithmic structure circuits of training programs construction (linear, cyclic, with hyperlinks, to the labels, which enable the development of STT variety and functionality are given; the methodology of STT creating is described based on the analysis of the technology teacher preparation in HEE content, MITE didactic functions and selection criteria of educational software for this area of specialist’s preparation.

  12. Quality Systems Implementation in the Pharmaceutical Industry

    African Journals Online (AJOL)

    Nafiisah

    quality standards imposed on local pharmaceutical manufacturers. Keywords: Quality/ .... GMP is concerned with both production and quality control. It is aimed ... in the European pharmaceutical industry in designing their quality systems:.

  13. Recommendations for a Software Quality Assurance Plan for the CMR Facility at LANL

    International Nuclear Information System (INIS)

    Adams, K.; Matthews, S. D.; McQueen, M. A.

    1998-01-01

    The Nuclear Materials Technology (NMT) organizations 1 and 3 within the Chemical and Metallurgical Research (CMR) facility at the Los Alamos National Laboratory are working to achieve Waste Isolation Pilot Plant (WIPP) certification to enable them to transport their TRU waste to WIPP. This document is intended to provide not only recommendations to address the necessary software quality assurance activities to enable the NMT-1 and NMT-3 organizations to be WIPP compliant but is also meant to provide a template for the final Software Quality Assurance Plan (SQAP). This document specifically addresses software quality assurance for all software used in support of waste characterization and analysis. Since NMT-1 and NMT-3 currently have several operational software products that are used for waste characterization and analysis, these software quality assurance recommendations apply to the operations, maintenance and retirement of the software and the creation and development of any new software required for waste characterization and analyses

  14. Software Quality Evaluation Models Applicable in Health Information and Communications Technologies. A Review of the Literature.

    Science.gov (United States)

    Villamor Ordozgoiti, Alberto; Delgado Hito, Pilar; Guix Comellas, Eva María; Fernandez Sanchez, Carlos Manuel; Garcia Hernandez, Milagros; Lluch Canut, Teresa

    2016-01-01

    Information and Communications Technologies in healthcare has increased the need to consider quality criteria through standardised processes. The aim of this study was to analyse the software quality evaluation models applicable to healthcare from the perspective of ICT-purchasers. Through a systematic literature review with the keywords software, product, quality, evaluation and health, we selected and analysed 20 original research papers published from 2005-2016 in health science and technology databases. The results showed four main topics: non-ISO models, software quality evaluation models based on ISO/IEC standards, studies analysing software quality evaluation models, and studies analysing ISO standards for software quality evaluation. The models provide cost-efficiency criteria for specific software, and improve use outcomes. The ISO/IEC25000 standard is shown as the most suitable for evaluating the quality of ICTs for healthcare use from the perspective of institutional acquisition.

  15. SU-F-J-72: A Clinical Usable Integrated Contouring Quality Evaluation Software for Radiotherapy

    International Nuclear Information System (INIS)

    Jiang, S; Dolly, S; Cai, B; Mutic, S; Li, H

    2016-01-01

    Purpose: To introduce the Auto Contour Evaluation (ACE) software, which is the clinical usable, user friendly, efficient and all-in-one toolbox for automatically identify common contouring errors in radiotherapy treatment planning using supervised machine learning techniques. Methods: ACE is developed with C# using Microsoft .Net framework and Windows Presentation Foundation (WPF) for elegant GUI design and smooth GUI transition animations through the integration of graphics engines and high dots per inch (DPI) settings on modern high resolution monitors. The industrial standard software design pattern, Model-View-ViewModel (MVVM) pattern, is chosen to be the major architecture of ACE for neat coding structure, deep modularization, easy maintainability and seamless communication with other clinical software. ACE consists of 1) a patient data importing module integrated with clinical patient database server, 2) a 2D DICOM image and RT structure simultaneously displaying module, 3) a 3D RT structure visualization module using Visualization Toolkit or VTK library and 4) a contour evaluation module using supervised pattern recognition algorithms to detect contouring errors and display detection results. ACE relies on supervised learning algorithms to handle all image processing and data processing jobs. Implementations of related algorithms are powered by Accord.Net scientific computing library for better efficiency and effectiveness. Results: ACE can take patient’s CT images and RT structures from commercial treatment planning software via direct user input or from patients’ database. All functionalities including 2D and 3D image visualization and RT contours error detection have been demonstrated with real clinical patient cases. Conclusion: ACE implements supervised learning algorithms and combines image processing and graphical visualization modules for RT contours verification. ACE has great potential for automated radiotherapy contouring quality verification

  16. SU-F-J-72: A Clinical Usable Integrated Contouring Quality Evaluation Software for Radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Jiang, S; Dolly, S; Cai, B; Mutic, S; Li, H [Washington University School of Medicine, Saint Louis, MO (United States)

    2016-06-15

    Purpose: To introduce the Auto Contour Evaluation (ACE) software, which is the clinical usable, user friendly, efficient and all-in-one toolbox for automatically identify common contouring errors in radiotherapy treatment planning using supervised machine learning techniques. Methods: ACE is developed with C# using Microsoft .Net framework and Windows Presentation Foundation (WPF) for elegant GUI design and smooth GUI transition animations through the integration of graphics engines and high dots per inch (DPI) settings on modern high resolution monitors. The industrial standard software design pattern, Model-View-ViewModel (MVVM) pattern, is chosen to be the major architecture of ACE for neat coding structure, deep modularization, easy maintainability and seamless communication with other clinical software. ACE consists of 1) a patient data importing module integrated with clinical patient database server, 2) a 2D DICOM image and RT structure simultaneously displaying module, 3) a 3D RT structure visualization module using Visualization Toolkit or VTK library and 4) a contour evaluation module using supervised pattern recognition algorithms to detect contouring errors and display detection results. ACE relies on supervised learning algorithms to handle all image processing and data processing jobs. Implementations of related algorithms are powered by Accord.Net scientific computing library for better efficiency and effectiveness. Results: ACE can take patient’s CT images and RT structures from commercial treatment planning software via direct user input or from patients’ database. All functionalities including 2D and 3D image visualization and RT contours error detection have been demonstrated with real clinical patient cases. Conclusion: ACE implements supervised learning algorithms and combines image processing and graphical visualization modules for RT contours verification. ACE has great potential for automated radiotherapy contouring quality verification

  17. 78 FR 16474 - Extension of the Period for Comments on the Enhancement of Quality of Software-Related Patents

    Science.gov (United States)

    2013-03-15

    ...] Extension of the Period for Comments on the Enhancement of Quality of Software-Related Patents AGENCY... announcing the formation of a partnership with the software community to enhance the quality of software... quality of software-related patents and the preparation of software-related patent applications including...

  18. Results of the EC research project REQUEST on software quality and reliability

    International Nuclear Information System (INIS)

    Kersken, M.; Saglietti, F.

    1990-01-01

    GRS work in software safety was mainly concerned with the qualitative assessment of software reliability and quality. As a supplement to these activities the work within the REQUEST project emphasized the quantitative determination of the respective parameters. The three-level quality model COQUAMO serves for the computation - and partly for the prediction - of quality factors during the software life cycle. PERFIDE controls the application of software reliability models during the test phase and in early operational life. Specific attention was paid to the assessment of fault-tolerant diverse software systems. (orig.) [de

  19. Caliko: An Inverse Kinematics Software Library Implementation of the FABRIK Algorithm

    OpenAIRE

    Lansley, Alastair; Vamplew, Peter; Smith, Philip; Foale, Cameron

    2016-01-01

    The Caliko library is an implementation of the FABRIK (Forward And Backward Reaching Inverse Kinematics) algorithm written in Java. The inverse kinematics (IK) algorithm is implemented in both 2D and 3D, and incorporates a variety of joint constraints as well as the ability to connect multiple IK chains together in a hierarchy. The library allows for the simple creation and solving of multiple IK chains as well as visualisation of these solutions. It is licensed under the MIT software license...

  20. A Behavior-Preserving Translation From FBD Design to C Implementation for Reactor Protection System Software

    International Nuclear Information System (INIS)

    Yoo, Junbeom; Kim, Euisub; Lee, Jangsoo

    2013-01-01

    Software safety for nuclear reactor protection systems (RPSs) is the most important requirement for the obtainment of permission for operation and export from government authorities, which is why it should be managed with well-experienced software development processes. The RPS software is typically modeled with function block diagrams (FBDs) in the design phase, and then mechanically translated into C programs in the implementation phase, which is finally compiled into executable machine codes and loaded on RPS hardware - PLC (Programmable Logic Controller). Whereas C Compilers are fully-verified COTS (Commercial Off-The-Shelf) software, translators from FBDs to C programs are provided by PLC vendors. Long-term experience, experiments and simulations have validated their correctness and function safety. This paper proposes a behavior-preserving translation from FBD design to C implementation for RPS software. It includes two sets of translation algorithms and rules as well as a prototype translator. We used an example of RPS software in a Korean nuclear power plant to demonstrate the correctness and effectiveness of the proposed translation

  1. A Behavior-Preserving Translation From FBD Design to C Implementation for Reactor Protection System Software

    Energy Technology Data Exchange (ETDEWEB)

    Yoo, Junbeom; Kim, Euisub [Konkuk Univ., Seoul (Korea, Republic of); Lee, Jangsoo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2013-08-15

    Software safety for nuclear reactor protection systems (RPSs) is the most important requirement for the obtainment of permission for operation and export from government authorities, which is why it should be managed with well-experienced software development processes. The RPS software is typically modeled with function block diagrams (FBDs) in the design phase, and then mechanically translated into C programs in the implementation phase, which is finally compiled into executable machine codes and loaded on RPS hardware - PLC (Programmable Logic Controller). Whereas C Compilers are fully-verified COTS (Commercial Off-The-Shelf) software, translators from FBDs to C programs are provided by PLC vendors. Long-term experience, experiments and simulations have validated their correctness and function safety. This paper proposes a behavior-preserving translation from FBD design to C implementation for RPS software. It includes two sets of translation algorithms and rules as well as a prototype translator. We used an example of RPS software in a Korean nuclear power plant to demonstrate the correctness and effectiveness of the proposed translation.

  2. Integrated software environment dedicated for implementation of control systems based on PLC controllers

    Directory of Open Access Journals (Sweden)

    Szymon SURMA

    2007-01-01

    Full Text Available Industrial processes’ control systems based on PLC controllers play today a very important role in all fields of transport, including also sea transport. Construction of control systems is the field of engineering, which has been continuously evolving towards maximum simplification of system design path. Up to now the time needed forthe system construction from the design to commissioning had to be divided into a few stages. A mistake made in an earlier stage caused that in most cases the next stages had to be restarted. Available debugging systems allows defect detection at an early stage of theproject implementation. The paper presents general characteristic of integrated software for implementation of complex control systems. The issues related to the software use for programming of the visualisation environment, control computer, selection oftransmission medium and transmission protocol as well as PLC controllers’ configuration, software and control have been analysed.

  3. Design and Implementation of a Media Access Component at Picsearch Using a Rigorous Software Engineering Approach

    OpenAIRE

    Silva, Diego Núñez

    2011-01-01

    With the arrival of a new generation of sophisticated smartphones, possibilities for mobile video usage are presenting exciting new opportunities. This Master's thesis is based on a collaboration with Picsearch to design and implement a software component that enables the company's video services in Android smartphones.

  4. Assessing and updating the reliability of concrete bridges subjected to spatial deterioration - principles and software implementation

    DEFF Research Database (Denmark)

    Schneider, Ronald; Fischer, Johannes; Bügler, Maximilian

    2015-01-01

    to implement the method presented here. The software prototype is applied to a typical highway bridge and the influence of inspection information on the system deterioration state and the structural reliability is quantified taking into account the spatial correlation of the corrosion process. This work...

  5. Software

    Energy Technology Data Exchange (ETDEWEB)

    Macedo, R.; Budd, G.; Ross, E.; Wells, P.

    2010-07-15

    The software section of this journal presented new software programs that have been developed to help in the exploration and development of hydrocarbon resources. Software provider IHS Inc. has made additions to its geological and engineering analysis software tool, IHS PETRA, a product used by geoscientists and engineers to visualize, analyze and manage well production, well log, drilling, reservoir, seismic and other related information. IHS PETRA also includes a directional well module and a decline curve analysis module to improve analysis capabilities in unconventional reservoirs. Petris Technology Inc. has developed a software to help manage the large volumes of data. PetrisWinds Enterprise (PWE) helps users find and manage wellbore data, including conventional wireline and MWD core data; analysis core photos and images; waveforms and NMR; and external files documentation. Ottawa-based Ambercore Software Inc. has been collaborating with Nexen on the Petroleum iQ software for steam assisted gravity drainage (SAGD) producers. Petroleum iQ integrates geology and geophysics data with engineering data in 3D and 4D. Calgary-based Envirosoft Corporation has developed a software that reduces the costly and time-consuming effort required to comply with Directive 39 of the Alberta Energy Resources Conservation Board. The product includes an emissions modelling software. Houston-based Seismic Micro-Technology (SMT) has developed the Kingdom software that features the latest in seismic interpretation. Holland-based Joa Oil and Gas and Calgary-based Computer Modelling Group have both supplied the petroleum industry with advanced reservoir simulation software that enables reservoir interpretation. The 2010 software survey included a guide to new software applications designed to facilitate petroleum exploration, drilling and production activities. Oil and gas producers can use the products for a range of functions, including reservoir characterization and accounting. In

  6. The Implementation of Satellite Attitude Control System Software Using Object Oriented Design

    Science.gov (United States)

    Reid, W. Mark; Hansell, William; Phillips, Tom; Anderson, Mark O.; Drury, Derek

    1998-01-01

    NASA established the Small Explorer (SNMX) program in 1988 to provide frequent opportunities for highly focused and relatively inexpensive space science missions. The SMEX program has produced five satellites, three of which have been successfully launched. The remaining two spacecraft are scheduled for launch within the coming year. NASA has recently developed a prototype for the next generation Small Explorer spacecraft (SMEX-Lite). This paper describes the object-oriented design (OOD) of the SMEX-Lite Attitude Control System (ACS) software. The SMEX-Lite ACS is three-axis controlled and is capable of performing sub-arc-minute pointing. This paper first describes high level requirements governing the SMEX-Lite ACS software architecture. Next, the context in which the software resides is explained. The paper describes the principles of encapsulation, inheritance, and polymorphism with respect to the implementation of an ACS software system. This paper will also discuss the design of several ACS software components. Specifically, object-oriented designs are presented for sensor data processing, attitude determination, attitude control, and failure detection. Finally, this paper will address the establishment of the ACS Foundation Class (AFC) Library. The AFC is a large software repository, requiring a minimal amount of code modifications to produce ACS software for future projects.

  7. 78 FR 53270 - Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2013-08-29

    ... Quality Implementation Plan; California; Sacramento Metropolitan Air Quality Management District... to the Sacramento Metropolitan Air Quality Management District (SMAQMD or District) portion of the..., Sacramento Metropolitan Air Quality Management District, Rule 214 (Federal New Source Review), Rule 203...

  8. Proteomics Quality Control: Quality Control Software for MaxQuant Results.

    Science.gov (United States)

    Bielow, Chris; Mastrobuoni, Guido; Kempa, Stefan

    2016-03-04

    Mass spectrometry-based proteomics coupled to liquid chromatography has matured into an automatized, high-throughput technology, producing data on the scale of multiple gigabytes per instrument per day. Consequently, an automated quality control (QC) and quality analysis (QA) capable of detecting measurement bias, verifying consistency, and avoiding propagation of error is paramount for instrument operators and scientists in charge of downstream analysis. We have developed an R-based QC pipeline called Proteomics Quality Control (PTXQC) for bottom-up LC-MS data generated by the MaxQuant software pipeline. PTXQC creates a QC report containing a comprehensive and powerful set of QC metrics, augmented with automated scoring functions. The automated scores are collated to create an overview heatmap at the beginning of the report, giving valuable guidance also to nonspecialists. Our software supports a wide range of experimental designs, including stable isotope labeling by amino acids in cell culture (SILAC), tandem mass tags (TMT), and label-free data. Furthermore, we introduce new metrics to score MaxQuant's Match-between-runs (MBR) functionality by which peptide identifications can be transferred across Raw files based on accurate retention time and m/z. Last but not least, PTXQC is easy to install and use and represents the first QC software capable of processing MaxQuant result tables. PTXQC is freely available at https://github.com/cbielow/PTXQC .

  9. Implementation of the Quality Management System in the telecommunications companies

    OpenAIRE

    Menshikova Ekaterina; Khazanov Oleg; Styazhkin Mark

    2016-01-01

    This article describes the advantages of the implementing of the quality management system based on ISO 9001 in the telecommunications companies. The stages of the implementation of the quality management system in the telecommunications companies are discussed. The map of the main process and criteria of processes is developed. The expediency of implementing the quality management system based on the requirements of ISO 9001 in Telecommunications is showed.

  10. Development of Automatic Quality Check Software in Mailbox Declaration For Nuclear Fuel Fabrication Plants

    International Nuclear Information System (INIS)

    Kim, Minsu; Shim, Hye Won; Jo, Seong Yeon; Lee, Kwang Yeol; Ban, Myoung Jin

    2014-01-01

    Short Notice Random Inspection (SNRI) is a new IAEA safeguards inspection regime for bulk handing facility, which utilities random inspection through a mailbox system. Its main objective is to verify 100% of the flow components of the safeguarded nuclear material at such a facility. To achieve the SNRI objective, it is required to provide daily mailbox declaration, by a facility's operator, to the IAEA with regard to information, such as the receipt and shipment of nuclear materials. Mailbox declarations are then later compared with accounting records so as to examine the accuracy and consistency of the facility operator's declaration at the time of the SNRI. The IAEA has emphasized the importance of accurate mailbox declarations and recommended that the ROK initiate its own independent quality control system in order to improve and maintain its mailbox declarations as a part of the SSAC activities. In an effort to improve the transparency of operational activities at fuel fabrication plants and to satisfy IAEA recommendation, an automatic quality check software application has been developed to improve mailbox declarations at fabrication plants in Korea. The ROK and the IAEA have recognized the importance of providing good quality mailbox declaration for an effective and efficient SNRI at fuel fabrication plants in Korea. The SRA developed an automatic quality check software program in order to provide an independent QC system of mailbox declaration, as well as to improve the quality of mailbox declaration. Once the automatic QC system is implemented, it will improve the quality of an operator's mailbox declaration by examining data before sending it to the IAEA. The QC system will be applied to fuel fabrication plants in the first half of 2014

  11. Portal for Families Overcoming Neurodevelopmental Disorders (PFOND): Implementation of a Software Framework for Facilitated Community Website Creation by Nontechnical Volunteers.

    Science.gov (United States)

    Ye, Xin Cynthia; Ng, Isaiah; Seid-Karbasi, Puya; Imam, Tuhina; Lee, Cheryl E; Chen, Shirley Yu; Herman, Adam; Sharma, Balraj; Johal, Gurinder; Gu, Bobby; Wasserman, Wyeth W

    2013-08-06

    The Portal for Families Overcoming Neurodevelopmental Disorders (PFOND) provides a structured Internet interface for the sharing of information with individuals struggling with the consequences of rare developmental disorders. Large disease-impacted communities can support fundraising organizations that disseminate Web-based information through elegant websites run by professional staff. Such quality resources for families challenged by rare disorders are infrequently produced and, when available, are often dependent upon the continued efforts of a single individual. The project endeavors to create an intuitive Web-based software system that allows a volunteer with limited technical computer skills to produce a useful rare disease website in a short time period. Such a system should provide access to emerging news and research findings, facilitate community participation, present summary information about the disorder, and allow for transient management by volunteers who are likely to change periodically. The prototype portal was implemented using the WordPress software system with both existing and customized supplementary plug-in software modules. Gamification scoring features were implemented in a module, allowing editors to measure progress. The system was installed on a Linux-based computer server, accessible across the Internet through standard Web browsers. A prototype PFOND system was implemented and tested. The prototype system features a structured organization with distinct partitions for background information, recent publications, and community discussions. The software design allows volunteer editors to create a themed website, implement a limited set of topic pages, and connect the software to dynamic RSS feeds providing information about recent news or advances. The prototype was assessed by a fraction of the disease sites developed (8 out of 27), including Aarskog-Scott syndrome, Aniridia, Adams-Oliver syndrome, Cat Eye syndrome, Kabuki syndrome

  12. Software engineering and Ada (Trademark) training: An implementation model for NASA

    Science.gov (United States)

    Legrand, Sue; Freedman, Glenn

    1988-01-01

    The choice of Ada for software engineering for projects such as the Space Station has resulted in government and industrial groups considering training programs that help workers become familiar with both a software culture and the intricacies of a new computer language. The questions of how much time it takes to learn software engineering with Ada, how much an organization should invest in such training, and how the training should be structured are considered. Software engineering is an emerging, dynamic discipline. It is defined by the author as the establishment and application of sound engineering environments, tools, methods, models, principles, and concepts combined with appropriate standards, guidelines, and practices to support computing which is correct, modifiable, reliable and safe, efficient, and understandable throughout the life cycle of the application. Neither the training programs needed, nor the content of such programs, have been well established. This study addresses the requirements for training for NASA personnel and recommends an implementation plan. A curriculum and a means of delivery are recommended. It is further suggested that a knowledgeable programmer may be able to learn Ada in 5 days, but that it takes 6 to 9 months to evolve into a software engineer who uses the language correctly and effectively. The curriculum and implementation plan can be adapted for each NASA Center according to the needs dictated by each project.

  13. Nuclear power project construction quality acceptance standards Implementation and discussion

    International Nuclear Information System (INIS)

    Lai Shifu

    2013-01-01

    Project quality construction acceptance is an important element of nuclear power project quality management. Based on characteristics of NPP construction and refer to national construction quality acceptance code and industry regulation, this paper introduces and discusses NPP construction quality acceptance regulation improvement and implementation under current situation. (authors)

  14. Concerning the problem of polygraphic wire calculation: theoretical aspects, software, practical implementation

    Directory of Open Access Journals (Sweden)

    Selkina A. V.

    2016-05-01

    Full Text Available the article analyzes the problems arising while organizing the workflow in printing companies. We suggest to address these problems by means of implementing computer-based accounting systems. Online and offline calculators used by printing enterprises for accounting are discussed. The author outlined the functional and specified requirements to such software. They were considered in the calculation module of accounting polygraphic wire used for block bonding. The software allows to increase the calculation process speed, to reduce the amount of errors in calculation and to decrease the labour intensity of the accounting process.

  15. Research of quality control during development of NPP DCS 1E classified software

    International Nuclear Information System (INIS)

    Shi Weihua; Lu Zhenguo; Xie Qi

    2012-01-01

    The Nuclear safety depends on right behavior of 1E software, which is a important part of 1E DCS system. Nowadays, user focus on good function of 1E system, but pay little attention to quality control of 1E software. In fact, it's declared in IEC61513 and IEC60880 that 1E software should under strict quality control during all stages of development. This article is related to the practice of 1E DCS system quality control and explores the QC surveillance for 1E software from the user's point of view. (authors)

  16. ISO and software quality assurance - licensing and certification of software professionals

    Energy Technology Data Exchange (ETDEWEB)

    Hare, J.; Rodin, L.

    1997-11-01

    This report contains viewgraphs on licensing and certifing of software professionals. Discussed in this report are: certification programs; licensing programs; why became certified; certification as a condition of empolyment; certification requirements; and examination structures.

  17. Application of software quality assurance to a specific scientific code development task

    International Nuclear Information System (INIS)

    Dronkers, J.J.

    1986-03-01

    This paper describes an application of software quality assurance to a specific scientific code development program. The software quality assurance program consists of three major components: administrative control, configuration management, and user documentation. The program attempts to be consistent with existing local traditions of scientific code development while at the same time providing a controlled process of development

  18. SQuAVisiT : A Software Quality Assessment and Visualisation Toolset

    NARCIS (Netherlands)

    Roubtsov, Serguei; Telea, Alexandru; Holten, Danny

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the Software Quality Assessment and Visualisation Toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  19. SQuAVisiT: a software quality assessment and visualisation toolset

    NARCIS (Netherlands)

    Roubtsov, S.; Telea, A.C.; Holten, D.H.R.

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the software quality assessment and visualisation toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully

  20. An empirical study of software architectures' effect on product quality

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Jonasson, Kristjan; Neukirchen, Helmut

    2011-01-01

    Software architectures shift the focus of developers from lines-of-code to coarser-grained components and their interconnection structure. Unlike 2ne-grained objects, these components typically encompass business functionality and need to be aware of the underlying business processes. Hence......, the interface of a component should re4ect relevant parts of the business process and the software architecture should emphasize the coordination among components. To shed light on these issues, we provide a framework for component-based software architectures focusing on the process perspective. The interface...

  1. Implementation of electronic medical records requires more than new software: Lessons on integrating and managing health technologies from Mbarara, Uganda.

    Science.gov (United States)

    Madore, Amy; Rosenberg, Julie; Muyindike, Winnie R; Bangsberg, David R; Bwana, Mwebesa B; Martin, Jeffrey N; Kanyesigye, Michael; Weintraub, Rebecca

    2015-12-01

    Implementation lessons: • Technology alone does not necessarily lead to improvement in health service delivery, in contrast to the common assumption that advanced technology goes hand in hand with progress. • Implementation of electronic medical record (EMR) systems is a complex, resource-intensive process that, in addition to software, hardware, and human resource investments, requires careful planning, change management skills, adaptability, and continuous engagement of stakeholders. • Research requirements and goals must be balanced with service delivery needs when determining how much information is essential to collect and who should be interfacing with the EMR system. • EMR systems require ongoing monitoring and regular updates to ensure they are responsive to evolving clinical use cases and research questions. • High-quality data and analyses are essential for EMRs to deliver value to providers, researchers, and patients. Copyright © 2015 Elsevier Inc. All rights reserved.

  2. Quality Implementation in Transition: A Framework for Specialists and Administrators.

    Science.gov (United States)

    Wald, Judy L.; Repetto, Jeanne B.

    1995-01-01

    Quality Implementation in Transition is a framework designed to guide transition specialists and administrators in the implementation of total quality management. The framework uses the tenets set forth by W. Edwards Deming and is intended to help professionals facilitate change within transition programs. (Author/JOW)

  3. Characterizing the contribution of quality requirements to software sustainability

    NARCIS (Netherlands)

    Condori-Fernandez, Nelly; Lago, Patricia

    2018-01-01

    Most respondents considered modifiability as relevant for addressing both technical and environmental sustainability. Functional correctness, availability, modifiability, interoperability and recoverability favor positively the endurability of software systems. This study has also identified

  4. Software requirements elicitation to support internal monitoring of quality assurance system for higher education in Indonesia

    Science.gov (United States)

    Amalia, A.; Gunawan, D.; Hardi, S. M.; Rachmawati, D.

    2018-02-01

    The Internal Quality Assurance System (in Indonesian: SPMI (Sistem Penjaminan Mutu Internal) is a systemic activity of quality assurance of higher education in Indonesia. SPMI should be done by all higher education or universities in Indonesia based on the Regulation of the Minister of Research, Technology and Higher Education of the Republic of Indonesia Number 62 of 2016. Implementation of SPMI must refer to the principle of SPMI that is independent, standardize, accurate, well planned and sustainable, documented and systematic. To assist the SPMI cycle properly, universities need a supporting software to monitor all the activities of SPMI. But in reality, many universities are not optimal in building this SPMI monitoring system. One of the obstacles is the determination of system requirements in support of SPMI principles is difficult to achieve. In this paper, we observe the initial phase of the engineering requirements elicitation. Unlike other methods that collect system requirements from users and stakeholders, we find the system requirements of the SPMI principles from SPMI guideline book. The result of this paper can be used as a choice in determining SPMI software requirements. This paper can also be used by developers and users to understand the scenario of SPMI so that could overcome the problems of understanding between this two parties.

  5. Complete modeling and software implementation of a virtual solar hydrogen hybrid system

    International Nuclear Information System (INIS)

    Pedrazzi, S.; Zini, G.; Tartarini, P.

    2010-01-01

    A complete mathematical model and software implementation of a solar hydrogen hybrid system has been developed and applied to real data. The mathematical model has been derived from sub-models taken from literature with appropriate modifications and improvements. The model has been implemented as a stand-alone virtual energy system in a model-based, multi-domain software environment. A test run has then been performed on typical residential user data-sets over a year-long period. Results show that the virtual hybrid system can bring about complete grid independence; in particular, hydrogen production balance is positive (+1.25 kg) after a year's operation with a system efficiency of 7%.

  6. Round table discussion: Quality control and standardization of nuclear medicine software

    International Nuclear Information System (INIS)

    Anon.

    1988-01-01

    In summary the round table came to the following important conclusions: Nuclear medicine software systems need better documentation, especially regarding details of algorithms and limitations, and user friendliness could be considerably improved. Quality control of software is an integral part of quality assurance in nuclear medicine and should be performed at all levels of the software. Quality control of applications software should preferably be performed with assistance of generally accepted software phantoms. A basic form of standardization was welcomed and partly regarded as essential by all participants. Some areas such as patient study files could be standardized in the near future, whereas other areas such as the standardization of clinical applications programs or acquisition protocols still present major difficulties. An international cooperation in the field of standardization of software and other topics has already been started on the European level and should be continued and supported. (orig.)

  7. 77 FR 73320 - Approval of Air Quality Implementation Plans; California; South Coast Air Quality Management...

    Science.gov (United States)

    2012-12-10

    ... Quality Implementation Plans; California; South Coast Air Quality Management District; Prevention of... Implementation Plan (SIP) revision for the South Coast Air Quality Management District (SCAQMD or District... in a August 15, 2012 letter from the South Coast Air Quality Management District regarding specific...

  8. Implementation of LTE SC-FDMA on the USRP2 Software Defined Radio Platform

    DEFF Research Database (Denmark)

    Jørgensen, Peter Bjørn; Hansen, Thomas Lundgaard; Sørensen, Troels Bundgaard

    2011-01-01

    In this paper we discuss the implementation of a Single Carrier Frequency Division Multiple Access (SC-FDMA) transceiver running over the Universal Software Radio Peripheral 2 (USRP2). SC-FDMA is the air interface which has been selected for the uplink in the latest Long Term Evolution (LTE....../s. Experimental results on the Bit Error Rate (BER) versus Signal-to-Noise Ratio (SNR) are presented and compared to theoretical and simulated performance....

  9. MEDBASE: Strategic Planning and Implementation of an Army Medical Department Software Application

    Science.gov (United States)

    2003-07-17

    downtime procedures) Software/ hardware performance testing Contingency plans Celebration / people management Risk Zone 4 Planning for upgrades...goal to “Crush Adidas !” and Honda’s goal in 1970 to “destroy Yamaha!”. MEDBASE: Strategy and Implementation 57 (3) Role Model BHAGs suit up-and...Crush Adidas ! (Nike, 1960s) Yamaha so tsubusu! We will destroy Yamaha! (Honda, 1970s) Role-Model BHAGs suit up-and-coming organizations Become the

  10. Implementation of the INSPECT software package for statistical calculation in nuclear material accountability

    International Nuclear Information System (INIS)

    Marzo, M.A.S.

    1986-01-01

    The INSPECT software package was developed in the Pacific Northwest Laboratory for statistical calculations in nuclear material accountability. The programs apply the inspection and evaluation methodology described in Part of the Safeguards Technical Manual. In this paper the implementation of INSPECT at the Safeguards Division of CNEN, and the main characteristics of INSPECT are described. The potential applications of INSPECT to the nuclear material accountability is presented. (Author) [pt

  11. Software Implementation of a Secure Firmware Update Solution in an IOT Context

    Directory of Open Access Journals (Sweden)

    Lukas Kvarda

    2016-01-01

    Full Text Available The present paper is concerned with the secure delivery of firmware updates to Internet of Things (IoT devices. Additionally, it deals with the design of a safe and secure bootloader for a UHF RFID reader. A software implementation of a secure firmware update solution is performed. The results show there is space to integrate even more security features into existing devices.

  12. Caliko: An Inverse Kinematics Software Library Implementation of the FABRIK Algorithm

    Directory of Open Access Journals (Sweden)

    Alastair Lansley

    2016-09-01

    Full Text Available The Caliko library is an implementation of the FABRIK (Forward And Backward Reaching Inverse Kinematics algorithm written in Java. The inverse kinematics (IK algorithm is implemented in both 2D and 3D, and incorporates a variety of joint constraints as well as the ability to connect multiple IK chains together in a hierarchy. The library allows for the simple creation and solving of multiple IK chains as well as visualisation of these solutions. It is licensed under the MIT software license and the source code is freely available for use and modification at: https://github.com/feduni/caliko

  13. Design and implementation of a software tool intended for simulation and test of real time codes

    International Nuclear Information System (INIS)

    Le Louarn, C.

    1986-09-01

    The objective of real time software testing is to show off processing errors and unobserved functional requirements or timing constraints in a code. In the perspective of safety analysis of nuclear equipments of power plants testing should be carried independently from the physical process (which is not generally available), and because casual hardware failures must be considered. We propose here a simulation and test tool, integrally software, with large interactive possibilities for testing assembly code running on microprocessor. The OST (outil d'aide a la simulation et au Test de logiciels temps reel) simulates code execution and hardware or software environment behaviour. Test execution is closely monitored and many useful informations are automatically saved. The present thesis work details, after exposing methods and tools dedicated to real time software, the OST system. We show the internal mechanisms and objects of the system: particularly ''events'' (which describe evolutions of the system under test) and mnemonics (which describe the variables). Then, we detail the interactive means available to the user for constructing the test data and the environment of the tested software. Finally, a prototype implementation is presented along with the results of the tests carried out. This demonstrates the many advantages of the use of an automatic tool over a manual investigation. As a conclusion, further developments, nececessary to complete the final tool are rewieved [fr

  14. Evidence-based development of a diagnosis-dependent therapy planning system and its implementation in modern diagnostic software.

    Science.gov (United States)

    Ahlers, M O; Jakstat, H A

    2005-07-01

    The prerequisite for structured individual therapy of craniomandibular dysfunctions is differential diagnostics. Suggestions for the structured recording of findings and their structured evaluation beyond the global diagnosis of "craniomandibular disorders" have been published. Only this structured approach enables computerization of the diagnostic process. The respective software is available for use in practice (CMDcheck for CMD screening, CMDfact for the differential diagnostics). Based on this structured diagnostics, knowledge-based therapy planning is also conceivable. The prerequisite for this would be a model of achieving consensus on the indicated forms of therapy related to the diagnosis. Therefore, a procedure for evidence-based achievement of consensus on suitable forms of therapy in CMD was developed first in multicentric cooperation, and then implemented in corresponding software. The clinical knowledge of experienced specialists was included consciously for the consensus achievement process. At the same time, anonymized mathematical statistical evaluations were used for control and objectification. Different examiners form different departments of several universities working independently of one another assigned the theoretically conceiveable therapeutic alternatives to the already published diagnostic scheme. After anonymization, the correlation of these assignments was then calculated mathematically. For achieving consensus in those cases for which no agreement initally existed, agreement was subsequently arrived at in the course of a consensus conference on the basis of literature evaluations and the discussion of clinical case examples. This consensus in turn finally served as the basis of a therapy planner implemented in the above-mentioned diagnostic software CMDfact. Contributing to quality assurance, the principles of programming this assistant as well as the interface for linking into the diagnostic software are documented and also published

  15. The Implementation of Satellite Control System Software Using Object Oriented Design

    Science.gov (United States)

    Anderson, Mark O.; Reid, Mark; Drury, Derek; Hansell, William; Phillips, Tom

    1998-01-01

    NASA established the Small Explorer (SMEX) program in 1988 to provide frequent opportunities for highly focused and relatively inexpensive space science missions that can be launched into low earth orbit by small expendable vehicles. The development schedule for each SMEX spacecraft was three years from start to launch. The SMEX program has produced five satellites; Solar Anomalous and Magnetospheric Particle Explorer (SAMPEX), Fast Auroral Snapshot Explorer (FAST), Submillimeter Wave Astronomy Satellite (SWAS), Transition Region and Coronal Explorer (TRACE) and Wide-Field Infrared Explorer (WIRE). SAMPEX and FAST are on-orbit, TRACE is scheduled to be launched in April of 1998, WIRE is scheduled to be launched in September of 1998, and SWAS is scheduled to be launched in January of 1999. In each of these missions, the Attitude Control System (ACS) software was written using a modular procedural design. Current program goals require complete spacecraft development within 18 months. This requirement has increased pressure to write reusable flight software. Object-Oriented Design (OOD) offers the constructs for developing an application that only needs modification for mission unique requirements. This paper describes the OOD that was used to develop the SMEX-Lite ACS software. The SMEX-Lite ACS is three-axis controlled, momentum stabilized, and is capable of performing sub-arc-minute pointing. The paper first describes the high level requirements which governed the architecture of the SMEX-Lite ACS software. Next, the context in which the software resides is explained. The paper describes the benefits of encapsulation, inheritance and polymorphism with respect to the implementation of an ACS software system. This paper will discuss the design of several software components that comprise the ACS software. Specifically, Object-Oriented designs are presented for sensor data processing, attitude control, attitude determination and failure detection. The paper addresses

  16. Implementation of an integral program of quality assurance based on EPID to the IMRT

    International Nuclear Information System (INIS)

    Yannez Ruiz-Labrandera; Emilio; Gonzalez Perez, Y.

    2015-01-01

    We bring forward with this research the implementation of a procedure related to the assurance guaranty in the control of tue quality of IMRT treatment based on the technology of electronic portal images digital (EPID). For the sake of accomplishing quality controls, based in pylic digital images, we used like main tool the System of pylic digital images IviewGT TM with his application software. For the control of positioning of the multi-plates, we implemented a program in MATLAB, which yields the errors of positioning of the plates. For the dosimetric controls, the images obtained for the fields of treatment were climbed with the software ImageJ, and compared with the treatment planning systems (TPS) model Elekta's PrecisePlan ® for it we used the software Verisoft. We managed to implement a comprehensive program of quality control for IMRT. The positioning errors of the multiplates intervening bayouth's test younger errors of positioning under a 1m threw which the requisite is for the IMRT. The rest of the geometric proofs yielded favorable results inmail with them tolerance, same as the test Picket Fence. We verified 2 cases with the technique step and shoot, for it we verified 16 field, where gamma Index varied 85,8 - 98,9. It was checked the possibility to accomplish the quality controls for IMRT using pylic digital images, in our case checked itself himself I apply the Linac Elekta specify on the Ameijeiras. (Author)

  17. Implementation of the qualities of radiodiagnostic: mammography

    International Nuclear Information System (INIS)

    Pacífico, Leonardo; Magalhães, Luís A.G.; Fernandes, Elisabeth; Peixoto, José Guilherme P.

    2017-01-01

    The objective of the present study was to evaluate the expanded uncertainty and present the result of the internal audit performed at the Laboratory of Radiological Sciences (LCR). The qualities of the mammographic bundles that are references in the LCR calibrations had their uncertainties and conformities with the standard evaluated. The expanded uncertainty was 1.40%, and the result of the internal audit was satisfactory. We conclude that LCR can perform calibrations on mammography qualities for end users. (author)

  18. Implementation of the qualities of radiodiagnostic: mammography

    Energy Technology Data Exchange (ETDEWEB)

    Pacífico, Leonardo; Magalhães, Luís A.G.; Fernandes, Elisabeth, E-mail: leonardocpacifico@gmail.com [Universidade do Estado do Rio de Janeiro (UERJ), RJ (Brazil); Peixoto, José Guilherme P. [Instituto de Radioproteção e Dosimetria (IRD/CNEN-RJ), Rio de Janeiro, RJ (Brazil)

    2017-07-01

    The objective of the present study was to evaluate the expanded uncertainty and present the result of the internal audit performed at the Laboratory of Radiological Sciences (LCR). The qualities of the mammographic bundles that are references in the LCR calibrations had their uncertainties and conformities with the standard evaluated. The expanded uncertainty was 1.40%, and the result of the internal audit was satisfactory. We conclude that LCR can perform calibrations on mammography qualities for end users. (author)

  19. Improving Software Quality and Management Through Use of Service Level Agreements

    Science.gov (United States)

    2005-03-01

    many who believe that the quality of the development process is the best predictor of software product quality. ( Fenton ) Repeatable software processes...reduced errors per KLOC for small projects ( Fenton ), and the quality management metric (QMM) (Machniak, Osmundson). There are also numerous IEEE 14...attention to cosmetic user interface issues and any problems that may arise with the prototype. (Sawyer) The validation process is also another check

  20. Application of Domain Knowledge to Software Quality Assurance

    Science.gov (United States)

    Wild, Christian W.

    1997-01-01

    , and possible reassessments resulting in new policy are made. In this regard we implemented a machine learning algorithm which automatically defined business rules based on expert assessment of the quality of operational decisions as recorded during deployment.

  1. The Role and Quality of Software Safety in the NASA Constellation Program

    Science.gov (United States)

    Layman, Lucas; Basili, Victor R.; Zelkowitz, Marvin V.

    2010-01-01

    In this study, we examine software safety risk in the early design phase of the NASA Constellation spaceflight program. Obtaining an accurate, program-wide picture of software safety risk is difficult across multiple, independently-developing systems. We leverage one source of safety information, hazard analysis, to provide NASA quality assurance managers with information regarding the ongoing state of software safety across the program. The goal of this research is two-fold: 1) to quantify the relative importance of software with respect to system safety; and 2) to quantify the level of risk presented by software in the hazard analysis. We examined 154 hazard reports created during the preliminary design phase of three major flight hardware systems within the Constellation program. To quantify the importance of software, we collected metrics based on the number of software-related causes and controls of hazardous conditions. To quantify the level of risk presented by software, we created a metric scheme to measure the specificity of these software causes. We found that from 49-70% of hazardous conditions in the three systems could be caused by software or software was involved in the prevention of the hazardous condition. We also found that 12-17% of the 2013 hazard causes involved software, and that 23-29% of all causes had a software control. Furthermore, 10-12% of all controls were software-based. There is potential for inaccuracy in these counts, however, as software causes are not consistently scoped, and the presence of software in a cause or control is not always clear. The application of our software specificity metrics also identified risks in the hazard reporting process. In particular, we found a number of traceability risks in the hazard reports may impede verification of software and system safety.

  2. Software to model AXAF-I image quality

    Science.gov (United States)

    Ahmad, Anees; Feng, Chen

    1995-01-01

    A modular user-friendly computer program for the modeling of grazing-incidence type x-ray optical systems has been developed. This comprehensive computer software GRAZTRACE covers the manipulation of input data, ray tracing with reflectivity and surface deformation effects, convolution with x-ray source shape, and x-ray scattering. The program also includes the capabilities for image analysis, detector scan modeling, and graphical presentation of the results. A number of utilities have been developed to interface the predicted Advanced X-ray Astrophysics Facility-Imaging (AXAF-I) mirror structural and thermal distortions with the ray-trace. This software is written in FORTRAN 77 and runs on a SUN/SPARC station. An interactive command mode version and a batch mode version of the software have been developed.

  3. Design and implementation air quality monitoring robot

    Science.gov (United States)

    Chen, Yuanhua; Li, Jie; Qi, Chunxue

    2017-01-01

    Robot applied in environmental protection can break through the limitations in working environment, scope and mode of the existing environmental monitoring and pollution abatement equipments, which undertake the innovation and improvement in the basin, atmosphere, emergency and pollution treatment facilities. Actually, the relevant technology is backward with limited research and investment. Though the device companies have achieved some results in the study on the water quality monitoring, pipeline monitoring and sewage disposal, this technological progress on the whole is still much slow, and the mature product has not been formed. As a result, the market urges a demand of a new type of device which is more suitable for environmental protection on the basis of robot successfully applied in other fields. This paper designs and realizes a tracked mobile robot of air quality monitoring, which can be used to monitor air quality for the pollution accident in industrial parks and regular management.

  4. Implementation of the qualities of radiodiagnostic: mammography

    Science.gov (United States)

    Pacífico, L. C.; Magalhães, L. A. G.; Peixoto, J. G. P.; Fernandes, E.

    2018-03-01

    The objective of the present study was to evaluate the expanded uncertainty of the mammographic calibration process and present the result of the internal audit performed at the Laboratory of Radiological Sciences (LCR). The qualities of the mammographic beans that are references in the LCR, comprises two irradiation conditions: no-attenuated beam and attenuated beam. Both had satisfactory results, with an expanded uncertainty equals 2,1%. The internal audit was performed, and the degree of accordance with the ISO/IEC 17025 was evaluated. The result of the internal audit was satisfactory. We conclude that LCR can perform calibrations on mammography qualities for end users.

  5. Advanced quality prediction model for software architectural knowledge sharing

    NARCIS (Netherlands)

    Liang, Peng; Jansen, Anton; Avgeriou, Paris; Tang, Antony; Xu, Lai

    In the field of software architecture, a paradigm shift is occurring from describing the outcome of architecting process to describing the Architectural Knowledge (AK) created and used during architecting. Many AK models have been defined to represent domain concepts and their relationships, and

  6. A Knowledge Management Approach to Support Software Process Improvement Implementation Initiatives

    Science.gov (United States)

    Montoni, Mariano Angel; Cerdeiral, Cristina; Zanetti, David; Cavalcanti da Rocha, Ana Regina

    The success of software process improvement (SPI) implementation initiatives depends fundamentally of the strategies adopted to support the execution of such initiatives. Therefore, it is essential to define adequate SPI implementation strategies aiming to facilitate the achievement of organizational business goals and to increase the benefits of process improvements. The objective of this work is to present an approach to support the execution of SPI implementation initiatives. We also describe a methodology applied to capture knowledge related to critical success factors that influence SPI initiatives. This knowledge was used to define effective SPI strategies aiming to increase the success of SPI initiatives coordinated by a specific SPI consultancy organization. This work also presents the functionalities of a set of tools integrated in a process-centered knowledge management environment, named CORE-KM, customized to support the presented approach.

  7. Quality factors quantification/assurance for software related to safety in nuclear power plants

    International Nuclear Information System (INIS)

    Nunez McLeod, J.E.; Rivera, S.S.

    1997-01-01

    Quality assurance plan is needed to guarantee the software quality. The use of such a plan involves activities that should take place all along the life cycle, and which can be evaluated using the so called quality factors. This is due to the fact that the quality itself cannot be measured, but some of its manifestations can be used for this purpose. In the present work, a methodology to quantify a set of quality factors is proposed, for software based systems to be used in safety related areas in nuclear power plants. (author) [es

  8. Implementing Kanban for agile process management within the ALMA Software Operations Group

    Science.gov (United States)

    Reveco, Johnny; Mora, Matias; Shen, Tzu-Chiang; Soto, Ruben; Sepulveda, Jorge; Ibsen, Jorge

    2014-07-01

    After the inauguration of the Atacama Large Millimeter/submillimeter Array (ALMA), the Software Operations Group in Chile has refocused its objectives to: (1) providing software support to tasks related to System Integration, Scientific Commissioning and Verification, as well as Early Science observations; (2) testing the remaining software features, still under development by the Integrated Computing Team across the world; and (3) designing and developing processes to optimize and increase the level of automation of operational tasks. Due to their different stakeholders, each of these tasks presents a wide diversity of importances, lifespans and complexities. Aiming to provide the proper priority and traceability for every task without stressing our engineers, we introduced the Kanban methodology in our processes in order to balance the demand on the team against the throughput of the delivered work. The aim of this paper is to share experiences gained during the implementation of Kanban in our processes, describing the difficulties we have found, solutions and adaptations that led us to our current but still evolving implementation, which has greatly improved our throughput, prioritization and problem traceability.

  9. Automated Software Acceleration in Programmable Logic for an Efficient NFFT Algorithm Implementation: A Case Study.

    Science.gov (United States)

    Rodríguez, Manuel; Magdaleno, Eduardo; Pérez, Fernando; García, Cristhian

    2017-03-28

    Non-equispaced Fast Fourier transform (NFFT) is a very important algorithm in several technological and scientific areas such as synthetic aperture radar, computational photography, medical imaging, telecommunications, seismic analysis and so on. However, its computation complexity is high. In this paper, we describe an efficient NFFT implementation with a hardware coprocessor using an All-Programmable System-on-Chip (APSoC). This is a hybrid device that employs an Advanced RISC Machine (ARM) as Processing System with Programmable Logic for high-performance digital signal processing through parallelism and pipeline techniques. The algorithm has been coded in C language with pragma directives to optimize the architecture of the system. We have used the very novel Software Develop System-on-Chip (SDSoC) evelopment tool that simplifies the interface and partitioning between hardware and software. This provides shorter development cycles and iterative improvements by exploring several architectures of the global system. The computational results shows that hardware acceleration significantly outperformed the software based implementation.

  10. A proposed metamodel for the implementation of object oriented software through the automatic generation of source code

    Directory of Open Access Journals (Sweden)

    CARVALHO, J. S. C.

    2008-12-01

    Full Text Available During the development of software one of the most visible risks and perhaps the biggest implementation obstacle relates to the time management. All delivery deadlines software versions must be followed, but it is not always possible, sometimes due to delay in coding. This paper presents a metamodel for software implementation, which will rise to a development tool for automatic generation of source code, in order to make any development pattern transparent to the programmer, significantly reducing the time spent in coding artifacts that make up the software.

  11. Total Quality Management Implementation: Selected Readings

    Science.gov (United States)

    1989-04-01

    Investigation of the Method of Accumulation Analysis by George Box and Stephen Jones. A discussion of Taguchi’s method for analyzing ordered categorical data. 1...P., Quality Control for Management, Prentice-Hall, Englewood Grant E L and Leavenworth. R.S. Statistical Quahtu Control, Mc- Cliffs, NJ., 1967. Graw ... Stephen A. Maguire, Ingalis Shipbuiiding which, in turn, improve the ability to anticipate productio: It is ei-couraging to see full-scale application

  12. Development of innovative computer software to facilitate the setup and computation of water quality index.

    Science.gov (United States)

    Nabizadeh, Ramin; Valadi Amin, Maryam; Alimohammadi, Mahmood; Naddafi, Kazem; Mahvi, Amir Hossein; Yousefzadeh, Samira

    2013-04-26

    Developing a water quality index which is used to convert the water quality dataset into a single number is the most important task of most water quality monitoring programmes. As the water quality index setup is based on different local obstacles, it is not feasible to introduce a definite water quality index to reveal the water quality level. In this study, an innovative software application, the Iranian Water Quality Index Software (IWQIS), is presented in order to facilitate calculation of a water quality index based on dynamic weight factors, which will help users to compute the water quality index in cases where some parameters are missing from the datasets. A dataset containing 735 water samples of drinking water quality in different parts of the country was used to show the performance of this software using different criteria parameters. The software proved to be an efficient tool to facilitate the setup of water quality indices based on flexible use of variables and water quality databases.

  13. Identifying strengths and weaknesses of Quality Management Unit University of Sumatera Utara software using SCAMPI C

    Science.gov (United States)

    Gunawan, D.; Amalia, A.; Rahmat, R. F.; Muchtar, M. A.; Siregar, I.

    2018-02-01

    Identification of software maturity level is a technique to determine the quality of the software. By identifying the software maturity level, the weaknesses of the software can be observed. As a result, the recommendations might be a reference for future software maintenance and development. This paper discusses the software Capability Level (CL) with case studies on Quality Management Unit (Unit Manajemen Mutu) University of Sumatera Utara (UMM-USU). This research utilized Standard CMMI Appraisal Method for Process Improvement class C (SCAMPI C) model with continuous representation. This model focuses on activities for developing quality products and services. The observation is done in three process areas, such as Project Planning (PP), Project Monitoring and Control (PMC), and Requirements Management (REQM). According to the measurement of software capability level for UMM-USU software, turns out that the capability level for the observed process area is in the range of CL1 and CL2. Planning Project (PP) is the only process area which reaches capability level 2, meanwhile, PMC and REQM are still in CL 1 or in performed level. This research reveals several weaknesses of existing UMM-USU software. Therefore, this study proposes several recommendations for UMM-USU to improve capability level for observed process areas.

  14. Monitoring the software quality in FairRoot

    Energy Technology Data Exchange (ETDEWEB)

    Uhlig, Florian; Al-Turany, Mohammad [GSI, Darmstadt (Germany)

    2010-07-01

    Up-to-date informations about a software project helps to find problems as early as possible. This includes for example information if a software project can be build on all supported platforms without errors or if specified tests can be executed and deliver the correct results. We present the scheme which is used within the FairRoot framework to continuously monitor the status of the project. The tools used for these tasks are based on the open source tools CMake and CDash. CMake is used to generate standard build files for the different operating systems/compiler out of simple configuration files and to steer the build and test processes. The generated information is send to a central CDash server. From the generated web pages information about the status of the project at any given time can be obtained.

  15. 77 FR 52277 - Approval of Air Quality Implementation Plans; California; South Coast Air Quality Management...

    Science.gov (United States)

    2012-08-29

    ... Quality Implementation Plans; California; South Coast Air Quality Management District; Prevention of... rule. SUMMARY: EPA is proposing approval of a permitting rule submitted for the South Coast Air Quality Management District (District) portion of the California State Implementation Plan (SIP). The State is...

  16. 77 FR 12524 - Approval and Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality...

    Science.gov (United States)

    2012-03-01

    ... Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality Standards AGENCY... Indiana State Implementation Plan (SIP) for lead (Pb) under the Clean Air Act (CAA). This submittal incorporates the National Ambient Air Quality Standards (NAAQS) for Pb promulgated by EPA in 2008. DATES...

  17. 75 FR 65594 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2010-10-26

    ... Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality Standards AGENCY... the Ohio Administrative Code (OAC) relating to the consolidation of Ohio's Ambient Air Quality Standards (AAQS) into Ohio's State Implementation Plan (SIP) under the Clean Air Act. On April 8, 2009, and...

  18. Integrated State Estimation and Contingency Analysis Software Implementation using High Performance Computing Techniques

    Energy Technology Data Exchange (ETDEWEB)

    Chen, Yousu; Glaesemann, Kurt R.; Rice, Mark J.; Huang, Zhenyu

    2015-12-31

    Power system simulation tools are traditionally developed in sequential mode and codes are optimized for single core computing only. However, the increasing complexity in the power grid models requires more intensive computation. The traditional simulation tools will soon not be able to meet the grid operation requirements. Therefore, power system simulation tools need to evolve accordingly to provide faster and better results for grid operations. This paper presents an integrated state estimation and contingency analysis software implementation using high performance computing techniques. The software is able to solve large size state estimation problems within one second and achieve a near-linear speedup of 9,800 with 10,000 cores for contingency analysis application. The performance evaluation is presented to show its effectiveness.

  19. A Framework for Evaluating the Software Product Quality of Pregnancy Monitoring Mobile Personal Health Records.

    Science.gov (United States)

    Idri, Ali; Bachiri, Mariam; Fernández-Alemán, José Luis

    2016-03-01

    Stakeholders' needs and expectations are identified by means of software quality requirements, which have an impact on software product quality. In this paper, we present a set of requirements for mobile personal health records (mPHRs) for pregnancy monitoring, which have been extracted from literature and existing mobile apps on the market. We also use the ISO/IEC 25030 standard to suggest the requirements that should be considered during the quality evaluation of these mPHRs. We then go on to design a checklist in which we contrast the mPHRs for pregnancy monitoring requirements with software product quality characteristics and sub-characteristics in order to calculate the impact of these requirements on software product quality, using the ISO/IEC 25010 software product quality standard. The results obtained show that the requirements related to the user's actions and the app's features have the most impact on the external sub-characteristics of the software product quality model. The only sub-characteristic affected by all the requirements is Appropriateness of Functional suitability. The characteristic Operability is affected by 95% of the requirements while the lowest degrees of impact were identified for Compatibility (15%) and Transferability (6%). Lastly, the degrees of the impact of the mPHRs for pregnancy monitoring requirements are discussed in order to provide appropriate recommendations for the developers and stakeholders of mPHRs for pregnancy monitoring.

  20. A pioneering application of NQA-1 quality assurance standards in the development of software

    International Nuclear Information System (INIS)

    Weisbin, A.N.

    1988-01-01

    The application of NQA-1 Quality Assurance Standards to computer software programs has been recent at the Oak Ridge National Laboratory. One reason for systematically applying quality assurance to computer software is the extensive use of results from computer programs. to characterize potential sites for nuclear waste repositories leading ultimately to important policy making decisions. Because data from these programs characterize the likely radioactivity profile for many hundreds of years, experimental validation is not feasible. The Sensitivity and Uncertainty Analysis Methods Development Project (SUAMDP) was developed to formulate and utilize efficient and comprehensive methods for determining sensitivities of calculated results with respect to changes in all input parameters. The computerized methodology was embodied in the Gradient Enhanced Software System (GRESS). Due to the fact that GRESS was to be used in the site characterization for waste storage, stringent NQA-1 requirements were imposed by the sponsor. A working relationship between the Oak Ridge National Laboratory (ORNL) Quality Department and the research scientists developing GRESS was essential in achieving understanding and acceptance of the quality assurance requirements as applied to the SUAMDP. The relationship resulted in the SUAMDP becoming the first software project at ORNL to develop a comprehensive NQA-1 Quality Assurance Plan; this plan now serves as a model for software quality assurance at ORNL. This paper describes the evolution of this plan and its impact on the application of quality assurance procedures to software

  1. Effect of quality control implementation on image quality of radiographic films and irradiation doses to patients

    International Nuclear Information System (INIS)

    Cheng Yuxi; Zhou Qipu; Ge Lijuan; Hou Changsong; Qi Xuesong; Yue Baorong; Wang Zuoling; Wei Kedao

    1999-01-01

    Objective: To study the changes in the image quality of radiographic films and the irradiation doses to patients after quality control (QC) implementation. Methods: The entrance surface doses (ESD) to patients measured with TLD and the image quality of radiographic films were evaluated on the basis of CEC image quality criteria. Results: The ESD to patients were significantly reduced after QC implementation (P 0.05), but the post-QC image quality was significantly improved in chest PA, lumbar spine AP and pelvis AP(P0.01 or P<0.05). Conclusion: Significantly reduced irradiation dose with improved image quality can be obtained by QC implementation

  2. SU-F-P-04: Implementation of Dose Monitoring Software: Successes and Pitfalls

    Energy Technology Data Exchange (ETDEWEB)

    Och, J [Geisinger Medical Center, Danville, PA (United States)

    2016-06-15

    Purpose: to successfully install a dose monitoring software (DMS) application to assist in CT protocol and dose management. Methods: Upon selecting the DMS, we began our implementation of the application. A working group composed of Medical Physics, Radiology Administration, Information Technology, and CT technologists was formed. On-site training in the application was supplied by the vendor. The decision was made to apply the process for all the CT protocols on all platforms at all facilities. Protocols were painstakingly mapped to the correct masters, and the system went ‘live’. Results: We are routinely using DMS as a tool in our Clinical Performance CT QA program. It is useful in determining the effectiveness of revisions to existing protocols, and establishing performance baselines for new units. However, the implementation was not without difficulty. We identified several pitfalls and obstacles which frustrated progress. Including: Training deficiencies, Nomenclature problems, Communication, DICOM variability. Conclusion: Dose monitoring software can be a potent tool for QA. However, implementation of the program can be problematic and requires planning, organization and commitment.

  3. Software Design and Implementation of a Next Generation GNSS Data Logger for Radio Frequency Interference Detection

    Science.gov (United States)

    Rivera, Sean

    The aim of this thesis is to implement a new GNSS interference detection system that extends the functionality of previous systems. This technology will be evaluated on the effectiveness of data collection in GPS L1, GPS L2C and all software changes required to implement it. Additionally there is discussion about extending the technology to log GLONASS L1 and GLONASS L2 data. This thesis is broken up into three sections. First, it describes the current state of GPS ASIC systems, using the SiGe receiver. It covers the reasons that this ASIC requires modernization as well as what behavior the system seeks to emulate. Next, it describes the new hardware that will make up the base of the new system. It covers what each component has been added for and its differences from the previous version. The focus is on the new ability of the receiver to receiver L2C simultaneously with L1 as well as the increased data rate which allows for better positioning. Finally, the thesis describes new software development effort centered around the new system. The thesis describes three test modes used, continuous logging, triggered logging and periodic logging. Extra attention is paid to the implementation of a real time networked version of the code. The final goal of the thesis is to build a low cost distributed system that is capable of detecting and localizing GPS and GLONASS interference on both the L1 and L2 bands.

  4. Analisis Kriteria Pengembangan Penganggaran Elektronik Menggunakan Software Quality Function Deployment (SQFD dari Sudut Pandang Pengguna

    Directory of Open Access Journals (Sweden)

    Rahmi Kartika Jati

    2017-08-01

    Full Text Available Pesatnya perkembangan sistem informasi menjadi pemicu dalam pengembangan aplikasi pendukung operasional organisasi. Tak terkecuali pemerintah Indonesia. Sebagai organisasi yang besar, pemerintah Indonesia memandang perlu adanya perbaikan layanan kepada masyarakat. Hal tersebut dilakukan dengan menetapkan penerapan e-government. E-government merupakan salah satu alat yang digunakan dalam perbaikan birokrasi dan tata kelola pemerintahan. Pusat Penelitian X sebagai lembaga pemerintahan melakukan perbaikan terus menerus untuk meningkatkan tata kelola organisasi. Dari beberapa area yang akan menjadi objek pengembangan e-government, pihak manajemen Pusat Penelitian X memutuskan untuk memprioritaskan perbaikan pada proses keuangan, khususnya proses penganggaran. Dalam rangka perbaikan proses anggaran tersebut, perlu kajian terkait pengembangan penganggaran elektronik. Berdasarkan hal tersebut penelitian ini bertujuan untuk pengembangan penganggaran elektronik di Pusat Penelitian X. Pengembangan dilakukan dengan menggunakan Software Quality Function Deployment (SQFD yang merupakan pengembangan Quality Function Deployment (QFD tradisional yang umum digunakan pada pengembangan produk manufaktur. Dari penelitian ini diperoleh hasil bahwa dalam pengembangan penganggaran elektronik di Pusat Penelitian X, programmer perlu memperhatikan delapan kriteria prioritas yang menurut pengguna perlu ada di dalam penganggaran elektronik. Dimana delapan kriteria prioritas tersebut dapat digunakan sebagai dasar pengetesan sistem untuk finalisasi pengembangan penganggaran elektronik.*****The rapid development of information system is the trigger for the developing of organization operational support applications. This is also the case of Indonesian government. As a big organization, Indonesian government sees the need of improving public services. Those can be conducted by implementing e-government, as a tool to improve the bureaucracy and governance effectively and

  5. Attributes Of Quality Scenarios/Scenario Sets Used In Software Requirements Elicitation

    National Research Council Canada - National Science Library

    Braun, Kimberly

    1997-01-01

    .... This thesis examines scenarios used in software requirements elicitation. Many different definitions, formats, and ideas exist on scenarios, but no thorough work has been done on what makes a good, quality scenario and scenario set...

  6. Report on the working conference on requirements engineering: foundation for software quality (REFSQ'09)

    NARCIS (Netherlands)

    Glinz, Martin; Heymans, Patrick; Persson, Anne; Sindre, Guttorm; Aurum, Aybüke; Madhavji, Nazim; Madhavji, N.; Paech, Barbara; Regev, Gil; Wieringa, Roelf J.

    This report summarizes the presentations and discussions at REFSQ’09, the 15th International Working Conference on Requirements Engineering: Foundation for Software Quality which was held on June 8-9, 2009 in Amsterdam, The Netherlands.

  7. Improving Software Quality and Management Through Use of Service Level Agreements

    National Research Council Canada - National Science Library

    Gaines, Leonard T

    2005-01-01

    .... SLAs are typically used in outsourcing contracts for post-production support. We propose that SLAs be used in software acquisition to support quality and process control throughout the lifecycle...

  8. Implementation of quality management and quality measurements in education

    DEFF Research Database (Denmark)

    Madsen, Ole Nørgaard; Carlsson, Ricci

    Gennemgang af program for implementering af kvalitetsledelse på en uddannelsesinstitution. I 7 faser præsenteres en fremgangsmåde baseret på den danske/europæiske model for selvevaluering som grundlag for kvalitetsbeskrivelse af virksomheden, prioritering af kvalitetsindsats, planlægning af kvali...

  9. The Impact of Organization, Project and Governance Variables on Software Quality and Project Success

    OpenAIRE

    Abbas, Noura; Gravell, Andy; Wills, Gary

    2010-01-01

    In this paper we present a statistically tested evidence about how quality and success rate are correlated with variables reflecting the organization and aspects of its project’s governance, namely retrospectives and metrics. The results presented in this paper are based on the Agile Projects Governance Survey that collected 129 responses. This paper discuss the deep analysis of this survey, and the main findings suggest that when applying agile software development, the quality of software i...

  10. Handbook of software quality assurance techniques applicable to the nuclear industry

    Energy Technology Data Exchange (ETDEWEB)

    Bryant, J.L.; Wilburn, N.P.

    1987-08-01

    Pacific Northwest Laboratory is conducting a research project to recommend good engineering practices in the application of 10 CFR 50, Appendix B requirements to assure quality in the development and use of computer software for the design and operation of nuclear power plants for NRC and industry. This handbook defines the content of a software quality assurance program by enumerating the techniques applicable. Definitions, descriptions, and references where further information may be obtained are provided for each topic.

  11. Handbook of software quality assurance techniques applicable to the nuclear industry

    International Nuclear Information System (INIS)

    Bryant, J.L.; Wilburn, N.P.

    1987-08-01

    Pacific Northwest Laboratory is conducting a research project to recommend good engineering practices in the application of 10 CFR 50, Appendix B requirements to assure quality in the development and use of computer software for the design and operation of nuclear power plants for NRC and industry. This handbook defines the content of a software quality assurance program by enumerating the techniques applicable. Definitions, descriptions, and references where further information may be obtained are provided for each topic

  12. Determinants of quality management systems implementation in hospitals.

    Science.gov (United States)

    Wardhani, Viera; Utarini, Adi; van Dijk, Jitse Pieter; Post, Doeke; Groothoff, Johan Willem

    2009-03-01

    To identify the problems and facilitating factors in the implementation of quality management system (QMS) in hospitals through a systematic review. A search strategy was performed on the Medline database for articles written in English published between 1992 and early 2006. Using the thesaurus terms 'Total Quality Management' and 'Quality Assurance Health Care', combined with the term 'hospital' and 'implement*', we identified 533 publications. The screening process was based on empirical articles describing organization-wide QMS implementation. Fourteen empirical articles fulfilled the inclusion criteria and were reviewed in this paper. An organization culture emphasizing standards and values associated with affiliation, teamwork and innovation, assumption of change and risk taking, play as the key success factor in QMS implementation. This culture needs to be supported by sufficient technical competence to apply a scientific problem-solving approach. A clear distribution of QMS function within the organizational structure is more important than establishing a formal quality structure. In addition to management leadership, physician involvement also plays an important role in implementing QMS. Six supporting and limiting factors determining QMS implementation are identified in this review. These are the organization culture, design, leadership for quality, physician involvement, quality structure and technical competence.

  13. Implementing Total Quality Management in a University Setting.

    Science.gov (United States)

    Coate, L. Edwin

    1991-01-01

    Oregon State University implemented Total Quality Management in nine phases: exploration; establishing a pilot study team; defining customer needs; adopting the breakthrough planning process; performing breakthrough planning in divisions; forming daily management teams; initiating cross-functional pilot projects; implementing cross-functional…

  14. Malaysian B5 Implementation and Its Quality

    International Nuclear Information System (INIS)

    Liang, Y.C.; Loh, S.K.; Soon, L.W.; May, C.Y.

    2016-01-01

    A quality survey of 80 samples of diesel fuels (B5) from 80 retail stations throughout Peninsular Malaysia was performed. The bio diesel contents in these B5 samples were in the range of 4.61 vol. % to 5.33 vol. %. The water contents of the samples were < 200 mg kg -1 , meeting the stringent water requirements recommended by the car manufacturers. In addition, all the samples exhibited superior lubricity as indicated by a much shorter wear scar diameter compared to neat diesel. More importantly, 98 % of the samples passed the oxidation stability test (PetroOXY) with an induction period of > 65 min. Overall, the B5 diesel fuel sold was in full compliance with the Malaysian diesel fuel standard specification. (author)

  15. Fast and predictable video compression in software design and implementation of an H.261 codec

    Science.gov (United States)

    Geske, Dagmar; Hess, Robert

    1998-09-01

    The use of software codecs for video compression becomes commonplace in several videoconferencing applications. In order to reduce conflicts with other applications used at the same time, mechanisms for resource reservation on endsystems need to determine an upper bound for computing time used by the codec. This leads to the demand for predictable execution times of compression/decompression. Since compression schemes as H.261 inherently depend on the motion contained in the video, an adaptive admission control is required. This paper presents a data driven approach based on dynamical reduction of the number of processed macroblocks in peak situations. Besides the absolute speed is a point of interest. The question, whether and how software compression of high quality video is feasible on today's desktop computers, is examined.

  16. Implementation guide for Hanford Analytical Services Quality Assurance Plan

    International Nuclear Information System (INIS)

    1994-09-01

    This implementation guide for the Hanford Analytical Services Quality Assurance Plan (HASQAP) was developed by the US Department of Energy, Richland Operations Office (RL) Waste Management Division, Analytical Services Branch. This plan formally presents RL's direction for Hanford Sitewide implementation of the HASQAP. The HASQAP establishes a uniform standard for quality requirements to meet US Department of Energy Order 5700.6C, Quality Assurance (10 CFR 830.120, ''Quality Assurance Requirements''), and is intended to satisfy the Hanford Federal Facility Agreement and Consent Order (Tri-Party Agreement) requirements for ''Guidance on Preparation of Laboratory Quality Assurance Plans''. The quality assurance criteria specified in the HASQAP shall serve as a baseline for implementing quality management systems for the laboratories that provide analytical services, for data requesters and users, and for oversight organizations that monitor the data-generation process. Affected organizations shall implement the HASQAP requirements that are applicable to their work scope. Full implementation of the HASQAP is scheduled to occur by August 1995. RL will work with the US Environmental Protection Agency (EPA) and Washington State Department of Ecology (Ecology) to have the HASQAP document incorporated into Appendix F of the Tri-Party Agreement by early Fiscal Year 1996

  17. Quality Assurance of Software Used In Aircraft Or Related Products

    Science.gov (United States)

    1993-02-01

    This advisory circular (AC) provides an acceptable means, but not the only means, to show compliance with the quality assurance requirements of Federal Aviation Regulations (FAR) Part 21, Certification Procedures for Products and Parts, as applicable...

  18. Implementation of software-based sensor linearization algorithms on low-cost microcontrollers.

    Science.gov (United States)

    Erdem, Hamit

    2010-10-01

    Nonlinear sensors and microcontrollers are used in many embedded system designs. As the input-output characteristic of most sensors is nonlinear in nature, obtaining data from a nonlinear sensor by using an integer microcontroller has always been a design challenge. This paper discusses the implementation of six software-based sensor linearization algorithms for low-cost microcontrollers. The comparative study of the linearization algorithms is performed by using a nonlinear optical distance-measuring sensor. The performance of the algorithms is examined with respect to memory space usage, linearization accuracy and algorithm execution time. The implementation and comparison results can be used for selection of a linearization algorithm based on the sensor transfer function, expected linearization accuracy and microcontroller capacity. Copyright © 2010 ISA. Published by Elsevier Ltd. All rights reserved.

  19. Exploring Issues of Implementation, Equity, and Student Achievement With Educational Software in the DC Public Schools

    Directory of Open Access Journals (Sweden)

    June Ahn

    2016-08-01

    Full Text Available In this article, we present analyses from a researcher-practitioner partnership in the District of Columbia Public Schools, where we are exploring the impact of educational software on students’ academic achievement. We analyze a unique data set that combines student-level information from the district with data of student usage of a mathematics game platform: First in Math (FIM. These data offer a window into long-standing issues in the educational technology literature around implementation, equity, and student achievement. We show that time spent in FIM was correlated with improved future performance on standardized math assessments for students in Grades 4–8. However, student time spent using FIM was highly related to factors such as race, gender, and prior achievement. Such observations from data are helpful for school districts and researchers to inform equitable implementation of new technologies and maximize benefits to learners.

  20. A Case of Engineering Quality for Mobile Healthcare Applications Using Augmented Personal Software Process Improvement

    Directory of Open Access Journals (Sweden)

    Shahbaz Ahmed Khan Ghayyur

    2016-01-01

    Full Text Available Mobile healthcare systems are currently considered as key research areas in the domain of software engineering. The adoption of modern technologies, for mobile healthcare systems, is a quick option for industry professionals. Software architecture is a key feature that contributes towards a software product, solution, or services. Software architecture helps in better communication, documentation of design decisions, risks identification, basis for reusability, scalability, scheduling, and reduced maintenance cost and lastly it helps to avoid software failures. Hence, in order to solve the abovementioned issues in mobile healthcare, the software architecture is integrated with personal software process. Personal software process has been applied successfully but it is unable to address the issues related to architectural design and evaluation capabilities. Hence, a new technique architecture augmented personal process is presented in order to enhance the quality of the mobile healthcare systems through the use of architectural design with integration of personal software process. The proposed process was validated by case studies. It was found that the proposed process helped in reducing the overall costs and effort. Moreover, an improved architectural design helped in development of high quality mobile healthcare system.

  1. A Comparison of Four Software Programs for Implementing Decision Analytic Cost-Effectiveness Models.

    Science.gov (United States)

    Hollman, Chase; Paulden, Mike; Pechlivanoglou, Petros; McCabe, Christopher

    2017-08-01

    The volume and technical complexity of both academic and commercial research using decision analytic modelling has increased rapidly over the last two decades. The range of software programs used for their implementation has also increased, but it remains true that a small number of programs account for the vast majority of cost-effectiveness modelling work. We report a comparison of four software programs: TreeAge Pro, Microsoft Excel, R and MATLAB. Our focus is on software commonly used for building Markov models and decision trees to conduct cohort simulations, given their predominance in the published literature around cost-effectiveness modelling. Our comparison uses three qualitative criteria as proposed by Eddy et al.: "transparency and validation", "learning curve" and "capability". In addition, we introduce the quantitative criterion of processing speed. We also consider the cost of each program to academic users and commercial users. We rank the programs based on each of these criteria. We find that, whilst Microsoft Excel and TreeAge Pro are good programs for educational purposes and for producing the types of analyses typically required by health technology assessment agencies, the efficiency and transparency advantages of programming languages such as MATLAB and R become increasingly valuable when more complex analyses are required.

  2. Fuzzy system for risk analysis in software projects through the attributes of quality standards iso 25000

    Directory of Open Access Journals (Sweden)

    Chau Sen Shia

    2014-02-01

    Full Text Available With the growth in demand for products and services in the IT area, companies encounter difficulties in establishing a metric or measure of quality of services to address measurably qualitative values in their planning. In this work fuzzy logic, standard SQuaRE (measurement of the quality of software products, Likertscale, GQM method (Goal-Question-Metric -indicator of quality of Software and the project risk analysis model of Boehm were used to assess the quality of services and decision-making, according to your demand and requests for software development. With the aim of improving the quality in the provision of services, the application is used to integrate the team and follow the life cycle of a project from its initial phase, and to assist in the comparison with the proposed schedule during the requirements elicitation.

  3. 78 FR 63934 - Approval of Air Quality Implementation Plans; California; El Dorado County Air Quality Management...

    Science.gov (United States)

    2013-10-25

    ...] Approval of Air Quality Implementation Plans; California; El Dorado County Air Quality Management District... California for the El Dorado County Air Quality Management District (EDAQMD) portion of the California SIP... 24, 1987 Federal Register, May 25, 1988, U.S. EPA, Air Quality Management Division, Office of Air...

  4. 78 FR 19990 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2013-04-03

    ... Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality Standards; Correction AGENCY... approved revisions to Ohio regulations that consolidated air quality standards in a new chapter of rules... State's air quality standards into Ohio Administrative Code (OAC) 3745-25 and modifying an assortment of...

  5. 75 FR 65572 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2010-10-26

    ... Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality Standards AGENCY... Ohio Administrative Code (OAC) relating to the consolidation of Ohio's Ambient Air Quality Standards... apply to Ohio's SIP. Incorporating the air quality standards into Ohio's SIP helps assure that...

  6. 77 FR 12482 - Approval and Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality...

    Science.gov (United States)

    2012-03-01

    ... Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality Standards AGENCY... incorporates the National Ambient Air Quality Standards (NAAQS) for Pb promulgated by EPA in 2008. DATES: This... FR 66964) and codified at 40 CFR 50.16, ``National primary and secondary ambient air quality...

  7. 78 FR 10589 - Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2013-02-14

    ... Quality Implementation Plan; California; Sacramento Metropolitan Air Quality Management District... Sacramento Metropolitan Air Quality Management District (SMAQMD or District) portion of the California State... sources within the areas covered by the plan as necessary to assure that the National Ambient Air Quality...

  8. Global Crisis as Enterprise Software Motivator: from Lifecycle Optimization to Efficient Implementation Series

    Directory of Open Access Journals (Sweden)

    Sergey V. Zykov

    2012-04-01

    Full Text Available It is generally known that software system development lifecycle (SSDL should be managed adequately. The global economy crisis and subsequent depression have taught us certain lessons on the subject, which is so vital for enterprises. The paper presents the adaptive methodology of enterprise SSDL, which allows to avoid "local crises" while producing large-scale software. The methodology is based on extracting common ERP module level patterns and applying them to series of heterogeneous implementations. The approach includes a lifecycle model, which extends conventional spiral model by formal data representation/management models and DSL-based "low-level" CASE tools supporting the formalisms. The methodology has been successfully implemented as a series of portal-based ERP systems in ITERA oil-and-gas corporation, and in a number of trading/banking enterprise applications for other enterprises. Semantic network-based airline dispatch system, and a 6D-model-driven nuclear power plant construction support system are currently in progress. Combining various SSDL models is discussed. Terms-and-cost reduction factors are examined. Correcting SSDL according to project size and scope is overviewed. The so-called “human factor errors” resulting from non-systematic SSDL approach, and their influencing crisis and depression, are analyzed. The ways to systematic and efficient SSDL are outlined. Troubleshooting advises are given for the problems concerned.

  9. Value of Investment as a Key Driver for Prioritization and Implementation of Healthcare Software.

    Science.gov (United States)

    Bata, Seth A; Richardson, Terry

    2018-01-01

    Health systems across the nation are recovering from massive financial and resource investments in electronic health record applications. In the midst of these recovery efforts, implementations of new care models, including accountable care organizations and population health initiatives, are underway. The shift from fee-for-service to fee-for-outcomes and fee-for-value payment models calls for care providers to work in new ways. It also changes how physicians are compensated and reimbursed. These changes necessitate that healthcare systems further invest in information technology solutions. Selecting which information technology (IT) projects are of most value is vital, especially in light of recent expenditures. Return-on-investment analysis is a powerful tool used in various industries to select the most appropriate IT investments. It has proven vital in selecting, justifying, and implementing software projects. Other financial metrics, such as net present value, economic value added, and total economic impact, also quantify the success of expenditures on information systems. This paper extends the concept of quantifying project value to include clinical outcomes and nonfinancial value as investment returns, applying a systematic approach to healthcare software projects. We term this inclusive approach Value of Investment. It offers a necessary extension for application in clinical settings where a strictly financial view may fall short in providing a complete picture of important benefits. This paper outlines the Value of Investment process and its attributes, and uses illustrative examples to explore the efficacy of this methodology within a midsized health system.

  10. Development of Software to Model AXAF-I Image Quality

    Science.gov (United States)

    Geary, Joseph; Hawkins, Lamar; Ahmad, Anees; Gong, Qian

    1997-01-01

    This report describes work conducted on Delivery Order 181 between October 1996 through June 1997. During this period software was written to: compute axial PSD's from RDOS AXAF-I mirror surface maps; plot axial surface errors and compute PSD's from HDOS "Big 8" axial scans; plot PSD's from FITS format PSD files; plot band-limited RMS vs axial and azimuthal position for multiple PSD files; combine and organize PSD's from multiple mirror surface measurements formatted as input to GRAZTRACE; modify GRAZTRACE to read FITS formatted PSD files; evaluate AXAF-I test results; improve and expand the capabilities of the GT x-ray mirror analysis package. During this period work began on a more user-friendly manual for the GT program, and improvements were made to the on-line help manual.

  11. C++ software quality in the ATLAS experiment: tools and experience

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00236968; The ATLAS collaboration; Kluth, Stefan; Seuster, Rolf; Snyder, Scott; Obreshkov, Emil; Roe, Shaun; Sherwood, Peter; Stewart, Graeme

    2017-01-01

    In this paper we explain how the C++ code quality is managed in ATLAS using a range of tools from compile-time through to run time testing and reflect on the substantial progress made in the last two years largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other available code analysis tools are also discussed, as is the role of unit testing with an example of how the GoogleTest framework can be applied to our codebase.

  12. C++ software quality in the ATLAS experiment: tools and experience

    Science.gov (United States)

    Martin-Haugh, S.; Kluth, S.; Seuster, R.; Snyder, S.; Obreshkov, E.; Roe, S.; Sherwood, P.; Stewart, G. A.

    2017-10-01

    In this paper we explain how the C++ code quality is managed in ATLAS using a range of tools from compile-time through to run time testing and reflect on the substantial progress made in the last two years largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other available code analysis tools are also discussed, as is the role of unit testing with an example of how the GoogleTest framework can be applied to our codebase.

  13. C++ Software Quality in the ATLAS Experiment: Tools and Experience

    CERN Document Server

    Kluth, Stefan; The ATLAS collaboration; Obreshkov, Emil; Roe, Shaun; Seuster, Rolf; Snyder, Scott; Stewart, Graeme

    2016-01-01

    The ATLAS experiment at CERN uses about six million lines of code and currently has about 420 developers whose background is largely from physics. In this paper we explain how the C++ code quality is managed using a range of tools from compile-time through to run time testing and reflect on the great progress made in the last year largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other tools including cppcheck, Include-What-You-Use and run-time 'sanitizers' are also discussed.

  14. New tools for digital medical image processing implemented in DIP software

    International Nuclear Information System (INIS)

    Araujo, Erica A.C.; Santana, Ivan E.; Lima, Fernando R.A.; Viera, Jose W.

    2011-01-01

    The anthropomorphic models used in computational dosimetry, also called phantoms, are mostly built from stacks of images CT (Computed Tomography) or MRI (Magnetic Resonance Imaging) obtained from scans of patients or volunteers. The construction of voxel phantoms requires computational processing for transforming image formats, dimensional image compression (2D) to form three-dimensional arrays (3D), quantization, resampling, enhancement, restoration and image segmentation, among others. Hardly the computational dosimetry researcher finds all these skills into a single software and often it results in a decreased development of their research or inadequate use of alternative tools. The need to integrate the various tasks of the original digital image processing to obtain an image that can be used in a computational model of exposure led to the development of software DIP (Digital Image Processing). This software reads, writes and edits binary files containing the 3D matrix corresponding to a stack of cross-sectional images of a given geometry that can be a human body or other volume of interest. It can also read any type of computer image and do conversions. When the task involves only one output image, it is saved in the JPEG standard Windows. When it involves a stack of images, the binary output file is called SGI (Interactive Graphic Simulations, a symbol already used in other publications of the Research Group in Numerical Dosimetry). The following paper presents the third version of the DIP software and emphasizes the new tools it implemented. Currently it has the menus Basics, Views, Spatial Domain, Frequency Domain, Segmentations and Study. Each menu contains items and subitems with features that generally require an image as input and produce an image or an attribute in the output. (author)

  15. Contribution of computer software to quality assurance process in radiopharmacy service at Grenoble University Hospital

    International Nuclear Information System (INIS)

    Sylvoz, N.; Desruet, M.-D.; Hustacheb, C.; Foroni, L.; Allenet, B.

    2009-01-01

    The optimization of the drug use process chain has become a priority for the improvement of quality of the care and for the logistic and economic rationalization. The aim was to assess the impact of the computerization, in radiopharmacy and nuclear medicine services, on safety of the supply chain and on the costs. We performed a before after survey: 50 weeks before and 50 weeks after introduction of the software. Indicators were selected to assess the safety and costs. Regarding the safety of the drug supply chain, computerization leads to a significant increase in the rate of nominative drug delivering satisfying legal requirements and significant decrease in delivering delay, especially those due to problems in radiopharmacy organization. Regarding the costs, the rate of drugs ordered but not used is significantly lower after computerization. This study made possible to identify critical control points and to implement corrective actions. From a legal point of view, computerization of drug prescribing and dispensing led to quantitative and qualitative improvements in the quality, safety and traceability of the drug supply chain. Lastly, computerization would be also a potential tool for cost saving. (authors) [fr

  16. Management assessments of Quality Assurance Program implementation effectiveness

    International Nuclear Information System (INIS)

    Snyder, D.A.

    1984-01-01

    This paper describes a method currently being used by UNC Nuclear Industries, Richland, Washington, to help assure the effectiveness of Quality Assurance (QA) Program implementation. Assessments are conducted annually by management in each department, and the results summarized to the president and his staff. The purpose of these assessments is to review the adequacy of the department's implementing procedures, training/instruction on implementing procedures, and procedure implementation effectiveness. The primary purpose is to assess effectiveness and take improvement action where the need is indicated. The QA organization provides only general guidance in conducting the assessments

  17. Implementation of Programmatic Quality and the Impact on Safety

    Science.gov (United States)

    Huls, Dale Thomas; Meehan, Kevin

    2005-01-01

    The purpose of this paper is to discuss the implementation of a programmatic quality assurance discipline within the International Space Station Program and the resulting impact on safety. NASA culture has continued to stress safety at the expense of quality when both are extremely important and both can equally influence the success or failure of a Program or Mission. Although safety was heavily criticized in the media after Colimbiaa, strong case can be made that it was the failure of quality processes and quality assurance in all processes that eventually led to the Columbia accident. Consequently, it is possible to have good quality processes without safety, but it is impossible to have good safety processes without quality. The ISS Program quality assurance function was analyzed as representative of the long-term manned missions that are consistent with the President s Vision for Space Exploration. Background topics are as follows: The quality assurance organizational structure within the ISS Program and the interrelationships between various internal and external organizations. ISS Program quality roles and responsibilities with respect to internal Program Offices and other external organizations such as the Shuttle Program, JSC Directorates, NASA Headquarters, NASA Contractors, other NASA Centers, and International Partner/participants will be addressed. A detailed analysis of implemented quality assurance responsibilities and functions with respect to NASA Headquarters, the JSC S&MA Directorate, and the ISS Program will be presented. Discussions topics are as follows: A comparison of quality and safety resources in terms of staffing, training, experience, and certifications. A benchmark assessment of the lessons learned from the Columbia Accident Investigation (CAB) Report (and follow-up reports and assessments), NASA Benchmarking, and traditional quality assurance activities against ISS quality procedures and practices. The lack of a coherent operational

  18. Software quality and process improvement in scientific simulation codes

    Energy Technology Data Exchange (ETDEWEB)

    Ambrosiano, J.; Webster, R. [Los Alamos National Lab., NM (United States)

    1997-11-01

    This report contains viewgraphs on the quest to develope better simulation code quality through process modeling and improvement. This study is based on the experience of the authors and interviews with ten subjects chosen from simulation code development teams at LANL. This study is descriptive rather than scientific.

  19. Achieving dependable software through Continuous Delivery and Quality Monitoring

    CERN Multimedia

    CERN. Geneva

    2015-01-01

    The idea for the presentation is to present our implementation of the CI/CD paradigms and explain on real live examples advantages and drawbacks of the current solution. During the presentation we will try to cover all the required steps which should automatically triggered by a developer’s commit. The presentation should give users a good hands-on experience on basic CI/CD principles and allow them to design and i...

  20. THE EXISTING BARRIERS IN IMPLEMENTING TOTAL QUALITY MANAGEMENT

    Directory of Open Access Journals (Sweden)

    Salagean Horatiu Catalin

    2014-07-01

    Full Text Available In the current market economy, companies are constantly struggling to achieve a sustained competitive advantage that will enable them to improve performance, which results in increased competitiveness, and of course, profit. Among the few competitive advantages that can become sustainable competitive advantages, quality plays a crucial role. Recent research shows that about 90% of buyers in the international market, consider quality as having at least equal importance with price in making the decision to purchase. In the opinion of some specialists in economic theory and practice, total quality refers to the holistic approach of quality, which actually means, addressing all aspects of economic and social development and technical of quality. Thus, the holistic approach of quality at organisation-wide involves procedural approach of quality, in this respect, the study focuses on this type of quality approach, i.e. the procedural approach, taking into account the strategic aspects of the continuous improvement of quality, which means in fact, the quality management. Total Quality Management is seen as a way to transform the economies of some countries to be more competitive than others. However, Total Quality Management brings not and will not produce results overnight, it is not a panacea for all the problems facing the organization. Total Quality Management requires a change in organizational culture, which must focus on meeting customer expectations and increasing the involvement of all employees to meet this objective, as an expression of the ethics of continuous improvement. In general, research on quality aiming identify why an organization should adopt the principles of total quality management, but attempts to identify the failing companies' attempts to implement total quality management principles are not so visible. Concerns companies to introduce quality management systems are becoming more pronounced, therefore, in this study we try to

  1. A pioneering application of NQA-1 quality assurance standards in the development of software

    International Nuclear Information System (INIS)

    Weisbin, A.N.

    1988-01-01

    One reason for systematically applying quality assurance to computer software is the extensive use of results from computer programs to characterize potential sited for nuclear waste repositories leading ultimately to important policy making decisions. Because data from these programs characterize the likely radioactivity profile for many hundreds of years, experimental validation is not feasible. The Sensitivity and Uncertainty Analysis Methods Development Project (SUAMDP) was developed to formulate and utilize efficient and comprehensive methods for determining sensitivities of calculated results with respect to changes in all input parameters. The computerized methodology was embodied in the Gradient Enhanced Software System (GRESS). Due to the fact that GRESS was to be used in the site characterization for waste storage, stringent NQA-1 requirements were imposed by the sponsor. A working relationship between the Oak Ridge National Laboratory (ORNL) Quality Department and the research scientists developing GRESS was essential in achieving understanding and acceptance of the quality assurance requirements as applied to the SUAMDP. The relationship resulted in the SUAMDP becoming the first software project at ORNL to develop a comprehensive NQA-1 Quality Assurance Plan; this plan now serves as a model for software quality assurance at ORNL. This paper describes the evolution of this plan and its impact on the application of quality assurance procedures to software. 2 refs

  2. Quality system implementation in the Radiotoxicology Laboratory of IPEN, Brazil

    International Nuclear Information System (INIS)

    Gaburo, J.C.; Caldeira Filho, J.S.; Todo, A.S.; Sanches, M.; Campos, L.L.

    2002-01-01

    The perception of assured quality is getting more transparence in the research and development areas. The Radiotoxicology Laboratory, LRT, of IPEN, operating since 1978, has as main attribution the development and implantation of analytical techniques for the measurements of different radionuclides in biological samples. Thus, the LRT considers being of extreme importance to have a management of the quality system to guarantee the reliability of the results and to obtain Accreditation Certificate of the National Institute of Metrology, INMETRO. With this objective the LRT has participated of projects of quality assurance for analytical laboratories since 1997, promoted by International Atomic Energy Agency, IAEA. Currently the quality system of the LRT is in implementation phase, operating in compliance with the quality system of IPEN (consistent pair ISO 9001-9004:2000) and with NBR ISO/IEC 17025. The quality system implemented in the LRT is described in its Quality Manual, MQ-LRT and in complementary procedures that are in their first revision. The participation of the laboratory in intercomparison programs among national as international laboratories and the analysis of the results of internal as well as external audits has demonstrated that the LRT laboratory presents good performance and with suitable methodology and accurate and precise results. With the implementation of the quality system it was possible to verify the effectiveness and efficiency of the tests carried out in the Radiotoxicology Laboratory. The project ARCAL XXVI was concluded in November 2001 and the LRT earned the Certificate of Recognition by IAEA. (author)

  3. A perspective on software quality management using microcomputers in safety-related activities

    International Nuclear Information System (INIS)

    Braudt, T.E.; Pratl, M.J.

    1992-01-01

    Software Quality Management, often referred to as Software Quality Assurance or SQA, is a belief or mindset in establishing and protecting the value of software as a corporate asset. It is often expressed in terms of a basic methodology for ensuring adequate controls to maintain the integrity of the configuration of a software system. SQA applies to all activities germane to the acquisition, installation, operation and maintenance of software systems and is key to calculational accuracy and completeness in an Engineering and/or Scientific arena. Simply, it is a vital management tool for ensuring cost-effective utilization of information management resources. The basis principles of SQA apply equally to software applications in microcomputer environments and mainframe environments alike. Regardless of the nature of the computing environment, divisions of responsibilities or logistical difficulties, quality measures must be established to ensure accuracy, completeness, reliability, and reproducibility of the results of the software application. The extent to which these measures are applied should be based upon regulation, economics and practicality

  4. Implementation of a Quality Management System in regulatory inspection activities

    International Nuclear Information System (INIS)

    Pires do Rio, Monica; Ferreira, Paulo Roberto; Cunha, Paulo G. da; Acar, Maria Elizabeth

    2005-01-01

    The Institute for Radioprotection and Dosimetry - IRD -, of the Brazilian National Nuclear Energy Commission, CNEN, started in 2001, the implementation of a quality management system (SGQ), in the inspection, testing and calibration activities. The SGQ was an institutional guideline and is inserted in a larger system of management of the IRD started in 1999, with the adoption of the National Quality Award criteria - PNQ, within the Project for Excellence in Technological Research of Associacao Brasileira das Instituicoes de Pesquisas Tecnologicas - ABIPTI (Brazilian Association of Technological Research institutions). The proposed quality management system and adopted at the IRD was developed and implemented in accordance with the requirements of NBR ISO/IEC 17025 - General requirements for the competence of testing and calibration laboratories, and ISO/IEC 17020 - General criteria for operation of various types of bodies performing inspections. For regulatory inspection activities, the quality system was implemented on three program inspection services of radiological protection led, respectively, by clinics and hospitals that operate radiotherapy services; industries that use nuclear gauges in their control or productive processes and power reactor operators (CNAAA) - just the environmental part. It was formed a pioneering team of inspectors for standardizing the processes, procedures and starting the implementation of the system in the areas. This work describes the implementation process steps, including difficulties, learning and advantages of the adoption of a quality management system in inspection activities

  5. Design of common software for quality control of SPECT

    International Nuclear Information System (INIS)

    Li Xiaohua; Gao Ruzhen; Chen Shengzu

    1993-01-01

    The goal of this study is to design a common testing system for SPECT quality control according to NEMA standard. Using the system, the performances of different types of SPECT can be tested, so that the acceptance testing, performance comparing and routine quality control for SPECT can be normalized. The system was based on IBM PC series of microcomputer. Testing data are acquired from various types of SPECT, then transferred into IBM PC through interface and tested with an unique testing program. Two parts were included: interface and SPECT testing program. It emphatically studied the managing program of RS232 interface, designing skills and the mathematic patterns of SPECT testing program. The system which was composed of 11 subroutines can be used to measure the performances for both gamma camera and SPECT. The system was tested on OMEGA 500/MCS 560 SPECT and the results showed that it is effective, accurate and easy to use

  6. IEEE Std 730 Software Quality Assurance: Supporting CMMI-DEV v1.3, Product and Process Quality Assurance

    Science.gov (United States)

    2011-05-27

    frameworks 4 CMMI-DEV IEEE / ISO / IEC 15288 / 12207 Quality Assurance ©2011 Walz IEEE Life Cycle Processes & Artifacts • Systems Life Cycle Processes...TAG to ISO TC 176 Quality Management • Quality: ASQ, work experience • Software: three books, consulting, work experience • Systems: Telecom & DoD...and IEEE 730 SQA need to align. The P730 IEEE standards working group has expanded the scope of the SQA process standard to align with IS 12207

  7. THE QUALITY MANAGEMENT SYSTEM AT THE UNIVERSITY. PROBLEMS IN IMPLEMENTING

    Directory of Open Access Journals (Sweden)

    L. I. Nazina

    2014-01-01

    Full Text Available Summary. Beginning of the XXI century marked the active integration of Russia and the world community, including entry into the world educational space. The international activities of the leading universities of the country depends on the level of cooperation with foreign countries in the field of scientific and educational activities. Today topical issue recognition of the quality of Russian education abroad. One way to confirm this is the certification of the quality management system of the university. As long-term, mostly foreign experience, achieving the required level of quality and continue to improve the most productive is accomplished by implementing a quality management system in accordance with ISO 9000. In the modern system of higher education a lot of attention paid to quality, to ensure the competitiveness of the institution in the education market by training qualified specialists. Implementation of the quality management system is to move the university to process management prepared in accordance with approved and enacted by the QMS documents. Workplaces should determine the duties, powers and responsibilities of each employee within the quality management system, develop and implement the job descriptions, staff verify documented procedures. Implementation of the certification must be accompanied by specific and persistent administration by management in conjunction with the training and retraining of all managers and employees involved in the work on the implementation of the QMS in the organization. This article describes the goals, objectives in the formation and implementation of QMS in higher education arising from this problem and barriers (information, methodological, psychological, multiple ways to address them.

  8. Identification of Patient Safety Risks Associated with Electronic Health Records: A Software Quality Perspective.

    Science.gov (United States)

    Virginio, Luiz A; Ricarte, Ivan Luiz Marques

    2015-01-01

    Although Electronic Health Records (EHR) can offer benefits to the health care process, there is a growing body of evidence that these systems can also incur risks to patient safety when developed or used improperly. This work is a literature review to identify these risks from a software quality perspective. Therefore, the risks were classified based on the ISO/IEC 25010 software quality model. The risks identified were related mainly to the characteristics of "functional suitability" (i.e., software bugs) and "usability" (i.e., interface prone to user error). This work elucidates the fact that EHR quality problems can adversely affect patient safety, resulting in errors such as incorrect patient identification, incorrect calculation of medication dosages, and lack of access to patient data. Therefore, the risks presented here provide the basis for developers and EHR regulating bodies to pay attention to the quality aspects of these systems that can result in patient harm.

  9. The impact of software quality characteristics on healthcare outcome: a literature review.

    Science.gov (United States)

    Aghazadeh, Sakineh; Pirnejad, Habibollah; Moradkhani, Alireza; Aliev, Alvosat

    2014-01-01

    The aim of this study was to discover the effect of software quality characteristics on healthcare quality and efficiency indicators. Through a systematic literature review, we selected and analyzed 37 original research papers to investigate the impact of the software indicators (coming from the standard ISO 9126 quality characteristics and sub-characteristics) on some of healthcare important outcome indicators and finally ranked these software indicators. The results showed that the software characteristics usability, reliability and efficiency were mostly favored in the studies, indicating their importance. On the other hand, user satisfaction, quality of patient care, clinical workflow efficiency, providers' communication and information exchange, patient satisfaction and care costs were among the healthcare outcome indicators frequently evaluated in relation to the mentioned software characteristics. Regression Logistic Method was the most common assessment methodology, and Confirmatory Factor Analysis and Structural Equation Modeling were performed to test the structural model's fit. The software characteristics were considered to impact the healthcare outcome indicators through other intermediate factors (variables).

  10. Software de balanced scorecard: proposta de um roteiro de implantação Balanced scorecard software: proposal of a guideline implementation

    Directory of Open Access Journals (Sweden)

    Oswaldo Keiji Hikage

    2006-04-01

    Full Text Available Em 1992, o conceito de Balanced Scorecard (BSC foi apresentado por Robert Kaplan e David Norton e em razão de sua disseminação muitas empresas que o adotaram e passaram por processos de fusões e aquisições tiveram suas informações aumentadas consideravelmente em seus bancos de dados. Esta série de informações e a necessidade de gerenciar com eficiência os indicadores estratégicos, de disponibilizar rapidamente relatórios gerenciais, analisar e simular cenários levaram as empresas a buscar um sistema automatizado. Dessa forma, algumas preocupações passaram a ter importância: como selecionar um software de BSC? Como implantar um software de BSC? Além das dificuldades inerentes à aquisição de qualquer software, a situação especificamente tratada neste artigo apresenta algumas peculiaridades, diante da intensa interação entre a sistemática do BSC e os softwares que o suportam. Este artigo, por meio de um estudo de caso realizado em uma empresa do setor de telecomunicações, enfoca a implantação de um software de BSC, visando ao desenvolvimento de um roteiro que possa sistematizar o processo.Since 1992, when Robert Kaplan and David Norton developed the concepts of Balanced Scorecard (BSC, many companies that adopted these concepts and survived in a higher competitive environment charicterized of high number of mergers, purchases and shutdowns increased the volume of information in their database. So, in consequence of the adoption of BSC, there is the needed for implementing a computerized control system, due to the work with the BSC and the management of a great volume of information. In this context, when the companies decide to implement a BSC software, they faced two problems: how to choose the BSC software? How to implement the selected software? This study proposes to develop a guideline for the BSC software introduction based on a case study in a telecommunication company and the experience from the practice on

  11. [Construction and implementation of two quality indicators in nursing services].

    Science.gov (United States)

    de Moura, Gisela Maria Schebela Souto; Juchem, Beatriz Cavalcanti; Falk, Maria Lucia Rodrigues; de Magalhães, Ana Maria Muller; Suzuki, Lyliam Midori

    2009-03-01

    Indicators monitor the quality of services and improve the attention offered to the patients. Hospital de Clinicas de Porto Alegre, Rio Grande do Sul, Brazil, has been developing strategies to assess its services according to its institutional management policy of quality The present study aims to report the experience at this university hospital with the construction and implementation of quality indicators in its nursing services. In 2006, four indicators were established: incidence of pressure ulcer, with a goal of quality of nursing services, for this is a large hospital.

  12. A tool to include gamma analysis software into a quality assurance program.

    Science.gov (United States)

    Agnew, Christina E; McGarry, Conor K

    2016-03-01

    To provide a tool to enable gamma analysis software algorithms to be included in a quality assurance (QA) program. Four image sets were created comprising two geometric images to independently test the distance to agreement (DTA) and dose difference (DD) elements of the gamma algorithm, a clinical step and shoot IMRT field and a clinical VMAT arc. The images were analysed using global and local gamma analysis with 2 in-house and 8 commercially available software encompassing 15 software versions. The effect of image resolution on gamma pass rates was also investigated. All but one software accurately calculated the gamma passing rate for the geometric images. Variation in global gamma passing rates of 1% at 3%/3mm and over 2% at 1%/1mm was measured between software and software versions with analysis of appropriately sampled images. This study provides a suite of test images and the gamma pass rates achieved for a selection of commercially available software. This image suite will enable validation of gamma analysis software within a QA program and provide a frame of reference by which to compare results reported in the literature from various manufacturers and software versions. Copyright © 2015. Published by Elsevier Ireland Ltd.

  13. Design and Implementation of a Mobile Phone Locator Using Software Defined Radio

    National Research Council Canada - National Science Library

    Larsen, Ian P

    2007-01-01

    ...) signal using software defined radio and commodity computer hardware. Using software designed by the GNU free software project as a base, standard GSM packets were transmitted and received over the air, and their arrival times detected...

  14. Software for creating quality control database in diagnostic radiology

    International Nuclear Information System (INIS)

    Stoeva, M.; Spassov, G.; Tabakov, S.

    2000-01-01

    The paper describes a PC based program with database for quality control (QC). It keeps information about all surveyed equipment and measured parameters. The first function of the program is to extract information from old (existing) MS Excel spreadsheets with QC surveys. The second function is used for input of measurements which are automatically organized in MS Excel spreadsheets and built into the database. The spreadsheets are based on the protocols described in the EMERALD Training Scheme. In addition, the program can make statistics of all measured parameters, both in absolute term and in time

  15. World Workshop on Oral Medicine VI: Utilization of Oral Medicine-specific software for support of clinical care, research, and education: current status and strategy for broader implementation.

    Science.gov (United States)

    Brailo, Vlaho; Firriolo, Francis John; Tanaka, Takako Imai; Varoni, Elena; Sykes, Rosemary; McCullough, Michael; Hua, Hong; Sklavounou, Alexandra; Jensen, Siri Beier; Lockhart, Peter B; Mattsson, Ulf; Jontell, Mats

    2015-08-01

    To assess the current scope and status of Oral Medicine-specific software (OMSS) utilized to support clinical care, research, and education in Oral Medicine and to propose a strategy for broader implementation of OMSS within the global Oral Medicine community. An invitation letter explaining the objectives was sent to the global Oral Medicine community. Respondents were interviewed to obtain information about different aspects of OMSS functionality. Ten OMSS tools were identified. Four were being used for clinical care, one was being used for research, two were being used for education, and three were multipurpose. Clinical software was being utilized as databases developed to integrate of different type of clinical information. Research software was designed to facilitate multicenter research. Educational software represented interactive, case-orientated technology designed for clinical training in Oral Medicine. Easy access to patient data was the most commonly reported advantage. Difficulty of use and poor integration with other software was the most commonly reported disadvantage. The OMSS presented in this paper demonstrate how information technology (IT) can have an impact on the quality of patient care, research, and education in the field of Oral Medicine. A strategy for broader implementation of OMSS is proposed. Copyright © 2015 Elsevier Inc. All rights reserved.

  16. Improving quality through effective implementation of information technology in healthcare.

    Science.gov (United States)

    Øvretveit, John; Scott, Tim; Rundall, Thomas G; Shortell, Stephen M; Brommels, Mats

    2007-10-01

    To describe an implementation of one information technology system (electronic medical record, EMR) in one hospital, the perceived impact, the factors thought to help and hinder implementation and the success of the system and compare this with theories of effective IT implementation. To draw on previous research, empirical data from this study is used to develop IT implementation theory. Qualitative case study, replicating the methods and questions of a previously published USA EMR implementation study using semi-structured interviews and documentation. Large Swedish teaching hospital shortly after a merger of two hospital sites. Thirty senior clinicians, managers, project team members, doctors and nurses. The Swedish implementation was achieved within a year and for under half the budget, with a generally popular EMR which was thought to save time and improve the quality of patient care. Evidence from this study and findings from the more problematic USA implementation case suggests that key factors for cost effective implementation and operation were features of the system itself, the implementation process and the conditions under which the implementation was carried out. There is empirical support for the IT implementation theory developed in this study, which provides a sound basis for future research and successful implementation. Successful implementation of an EMR is likely with an intuitive system, requiring little training, already well developed for clinical work but allowing flexibility for development, where clinicians are involved in selection and in modification for their department needs and where a realistic timetable is made using an assessment of the change-capability of the organization. Once a system decision is made, the implementation should be driven by top and departmental leaders assisted by competent project teams involving information technology specialists and users. Corrections for unforeseen eventualities will be needed, especially

  17. Introductory Molecular Orbital Theory: An Honors General Chemistry Computational Lab as Implemented Using Three-Dimensional Modeling Software

    Science.gov (United States)

    Ruddick, Kristie R.; Parrill, Abby L.; Petersen, Richard L.

    2012-01-01

    In this study, a computational molecular orbital theory experiment was implemented in a first-semester honors general chemistry course. Students used the GAMESS (General Atomic and Molecular Electronic Structure System) quantum mechanical software (as implemented in ChemBio3D) to optimize the geometry for various small molecules. Extended Huckel…

  18. NEMA NU-1 2007 based and independent quality control software for gamma cameras and SPECT

    International Nuclear Information System (INIS)

    Vickery, A; Joergensen, T; De Nijs, R

    2011-01-01

    A thorough quality assurance of gamma and SPECT cameras requires a careful handling of the measured quality control (QC) data. Most gamma camera manufacturers provide the users with camera specific QC Software. This QC software is indeed a useful tool for the following of day-to-day performance of a single camera. However, when it comes to objective performance comparison of different gamma cameras and a deeper understanding of the calculated numbers, the use of camera specific QC software without access to the source code is rather avoided. Calculations and definitions might differ, and manufacturer independent standardized results are preferred. Based upon the NEMA Standards Publication NU 1-2007, we have developed a suite of easy-to-use data handling software for processing acquired QC data providing the user with instructive images and text files with the results.

  19. PLANNING QUALITY ASSURANCE PROCESSES IN A LARGE SCALE GEOGRAPHICALLY SPREAD HYBRID SOFTWARE DEVELOPMENT PROJECT

    Directory of Open Access Journals (Sweden)

    Святослав Аркадійович МУРАВЕЦЬКИЙ

    2016-02-01

    Full Text Available There have been discussed key points of operational activates in a large scale geographically spread software development projects. A look taken at required QA processes structure in such project. There have been given up to date methods of integration quality assurance processes into software development processes. There have been reviewed existing groups of software development methodologies. Such as sequential, agile and based on RPINCE2. There have been given a condensed overview of quality assurance processes in each group. There have been given a review of common challenges that sequential and agile models are having in case of large geographically spread hybrid software development project. Recommendations were given in order to tackle those challenges.  The conclusions about the best methodology choice and appliance to the particular project have been made.

  20. SOFTWARE QUALITY ASSURANCE FOR EMERGENCY RESPONSE CONSEQUENCE ASSESSMENT MODELS AT DOE'S SAVANNAH RIVER SITE

    International Nuclear Information System (INIS)

    Hunter, C

    2007-01-01

    The Savannah River National Laboratory's (SRNL) Atmospheric Technologies Group develops, maintains, and operates computer-based software applications for use in emergency response consequence assessment at DOE's Savannah River Site. These applications range from straightforward, stand-alone Gaussian dispersion models run with simple meteorological input to complex computational software systems with supporting scripts that simulate highly dynamic atmospheric processes. A software quality assurance program has been developed to ensure appropriate lifecycle management of these software applications. This program was designed to meet fully the overall structure and intent of SRNL's institutional software QA programs, yet remain sufficiently practical to achieve the necessary level of control in a cost-effective manner. A general overview of this program is described

  1. SSCL magnet systems quality program implementation for laboratory and industry

    International Nuclear Information System (INIS)

    Warner, D.G.; Bever, D.L.

    1992-01-01

    The development and delivery of reliable and producible magnets for the Superconducting Super Collider Laboratory (SSCL) require the teamwork of a large and diverse workforce composed of personnel with backgrounds in laboratory research, defense, and energy. The SSCL Magnet Quality Program is being implemented with focus on three definitive objectives: (1) communication of requirements, (2) teamwork, and (3) verification. Examination of the SSCL Magnet Systems Division's (MSD) current and planned approach to implementation of the SSCL Magnet Quality Program utilizing these objectives is discussed

  2. Implementation Of Code And Carrier Tracking Loops For Software GPS Receivers

    Directory of Open Access Journals (Sweden)

    Win Kay Khaing

    2015-06-01

    Full Text Available Abstract GPS is playing in very important role in our modern mobile societies. Software approach is very flexible rather than the traditional hardware receivers. The soft-GPS receiver includes two portions hardware and software. In hardware portion an antenna filter down-converter from RF Radio Frequency to IF Intermediate Frequency and an ADC Analog to Digital Converter are included. In software portion signal processing such as acquisition tracking and navigation that runs on general purpose processor is included. The GPS signal is taken from N-FUELS Full Educational Library of Signals for Navigation signal simulator. The heart of soft-GPS receiver is the synchronization processes such as acquisition and tracking. In tracking there are two main loops for code and carrier tracking. The objective of this paper is to analyse and find the optimum discriminator function for the code tracking loop in soft-GPS receivers. The delay lock loop DLL is a well-known technique to track the codes for GNSS spread spectrum systems. This paper also presents non-coherent square law DLLs and the impacts of some parameters on DLL discriminators such as number of samples per chip early-late spacing different C No values where C denotes the signal power and No is the noise spectral density and the impact of with or without front-end device. The results of discriminator outputs are illustrated by using S-curves. Testing results with the real GPS signal are also described. This optimized discriminator functions can be implemented in any soft-GPS receivers.

  3. Space Shuttle Program Primary Avionics Software System (PASS) Success Legacy - Quality and Reliability Date

    Science.gov (United States)

    Orr, James K.; Peltier, Daryl

    2010-01-01

    Thsi slide presentation reviews the avionics software system on board the space shuttle, with particular emphasis on the quality and reliability. The Primary Avionics Software System (PASS) provides automatic and fly-by-wire control of critical shuttle systems which executes in redundant computers. Charts given show the number of space shuttle flights vs time, PASS's development history, and other charts that point to the reliability of the system's development. The reliability of the system is also compared to predicted reliability.

  4. Designing an Adaptive Nuero-Fuzzy Inference System for Evaluating the Business Intelligence System Implementation in Software Industry

    Directory of Open Access Journals (Sweden)

    Iman Raeesi Vanani

    2015-03-01

    Full Text Available The main goal of research is designing an adaptive nuero-fuzzy inference system for evaluating the implementation of business intelligence systems in software industry. Iranian software development organizations have been facing a lot of problems in case of implementing business intelligence systems. This system would be helpful in recognizing the conditions and prerequisites of success or failure. Organizations can recalculate the neuro-fuzzy system outputs with some considerations on various inputs to figure out which inputs have the most effect on the implementation outputs. By resolving the problems on inputs, organizations can achieve a better level of implementation success. The designed system has been trained by a data set and afterwards, it has been evaluated. The trained system has reached the error value of 0.08. Eventually, some recommendations have been provided for software development firms on the areas that might need more considerations and improvements.

  5. Considerations for control system software verification and validation specific to implementations using distributed processor architectures

    International Nuclear Information System (INIS)

    Munro, J.K. Jr.

    1993-01-01

    Until recently, digital control systems have been implemented on centralized processing systems to function in one of several ways: (1) as a single processor control system; (2) as a supervisor at the top of a hierarchical network of multiple processors; or (3) in a client-server mode. Each of these architectures uses a very different set of communication protocols. The latter two architectures also belong to the category of distributed control systems. Distributed control systems can have a central focus, as in the cases just cited, or be quite decentralized in a loosely coupled, shared responsibility arrangement. This last architecture is analogous to autonomous hosts on a local area network. Each of the architectures identified above will have a different set of architecture-associated issues to be addressed in the verification and validation activities during software development. This paper summarizes results of efforts to identify, describe, contrast, and compare these issues

  6. Implementation of a data management software system for SSME test history data

    Science.gov (United States)

    Abernethy, Kenneth

    1986-01-01

    The implementation of a software system for managing Space Shuttle Main Engine (SSME) test/flight historical data is presented. The software system uses the database management system RIM7 for primary data storage and routine data management, but includes several FORTRAN programs, described here, which provide customized access to the RIM7 database. The consolidation, modification, and transfer of data from the database THIST, to the RIM7 database THISRM is discussed. The RIM7 utility modules for generating some standard reports from THISRM and performing some routine updating and maintenance are briefly described. The FORTRAN accessing programs described include programs for initial loading of large data sets into the database, capturing data from files for database inclusion, and producing specialized statistical reports which cannot be provided by the RIM7 report generator utility. An expert system tutorial, constructed using the expert system shell product INSIGHT2, is described. Finally, a potential expert system, which would analyze data in the database, is outlined. This system could use INSIGHT2 as well and would take advantage of RIM7's compatibility with the microcomputer database system RBase 5000.

  7. Design and Implementation of Integrated Software Research and Community Service at State Polytechnic of Manado

    Science.gov (United States)

    Saroinsong, T.; A. S Kondoj, M.; Kandiyoh, G.; Pontoh, G.

    2018-01-01

    The State Polytechnic of Manado (Polimdo) is one of the reliable institutions in North Sulawesi that first implemented ISO 9001. But the accreditation of the institution has not been satisfactory, it means there is still much to be prepared to achieve the expected target. One of the criteria of assessment of institutional accreditation is related to research activities and social work in accordance with the standard seven. Data documentation systems related to research activities and social work are not well integrated and well documented in all existing work units. This causes the process of gathering information related to the activities and the results of research and social work in order to support the accreditation activities of the institution is still not efficient. This study aims to build an integrated software in all work units in Polimdo to obtain documentation and data synchronization in support of activities or reporting of documents accreditation institution in accordance with standard seven specifically in terms of submission of research proposal and dedication. The software will be developed using RUP method with analysis using data flow diagram and ERM so that the result of this research is documentation and synchronization of data and information of research activity and community service which can be used in preparing documents report for accreditation institution.

  8. Implementation of highly parallel and large scale GW calculations within the OpenAtom software

    Science.gov (United States)

    Ismail-Beigi, Sohrab

    The need to describe electronic excitations with better accuracy than provided by band structures produced by Density Functional Theory (DFT) has been a long-term enterprise for the computational condensed matter and materials theory communities. In some cases, appropriate theoretical frameworks have existed for some time but have been difficult to apply widely due to computational cost. For example, the GW approximation incorporates a great deal of important non-local and dynamical electronic interaction effects but has been too computationally expensive for routine use in large materials simulations. OpenAtom is an open source massively parallel ab initiodensity functional software package based on plane waves and pseudopotentials (http://charm.cs.uiuc.edu/OpenAtom/) that takes advantage of the Charm + + parallel framework. At present, it is developed via a three-way collaboration, funded by an NSF SI2-SSI grant (ACI-1339804), between Yale (Ismail-Beigi), IBM T. J. Watson (Glenn Martyna) and the University of Illinois at Urbana Champaign (Laxmikant Kale). We will describe the project and our current approach towards implementing large scale GW calculations with OpenAtom. Potential applications of large scale parallel GW software for problems involving electronic excitations in semiconductor and/or metal oxide systems will be also be pointed out.

  9. Inter-comparison and Quality Assurance of acquisition and processing software for MUGA studies in Cuba

    International Nuclear Information System (INIS)

    Lopez, A.; Ponce, F.; Peix, A.; Gonzalez, J.; Perez, M.; Diaz, M.

    2002-01-01

    With the purpose of create the bases for quality control and quality assurance of the acquisition and processing program of gated cardiac blood-pool (MUGA) studies, we used the VENSTRA cardiac function phantom in 7 cameras (4 SOPHA- DSX-1000, 2 GE- IMAGAMMA-2001 and 1 SIEMENS- HERMES) and made 3 acquisition for each Global Left Ventricular Ejection Fraction (LVEF 30%, 60% and 80%) and for each Heart Rate (HR 40, 80 and 160 heart beat/min). The planar resolution and the planar uniformity were proper in all the equipment. Differences less than 5% were found between the acquisition and processing program. To evaluate the processing program without the acquisition parameter's influence, we used one group of these image like software phantom and test the semi-automatic software in all cameras. The semi-automatic protocol showed difference less than 3% between software. The automatic processing software of gated cardiac studies were checked with the COST-B2 software phantom; the difference between the Left Ventricle Ejection Fraction calculated by these software was less than 5% and the regional wall motion analysis was complete coincident in the 93% of the cases. The use of VENSTRA and COST- B2 phantom confirm the correct functioning of the acquisition and the LVEF calculus software of MUGA studies in the 83% of cuban nuclear medicine centers

  10. On the characterization and software implementation of general protein lattice models.

    Directory of Open Access Journals (Sweden)

    Alessio Bechini

    Full Text Available models of proteins have been widely used as a practical means to computationally investigate general properties of the system. In lattice models any sterically feasible conformation is represented as a self-avoiding walk on a lattice, and residue types are limited in number. So far, only two- or three-dimensional lattices have been used. The inspection of the neighborhood of alpha carbons in the core of real proteins reveals that also lattices with higher coordination numbers, possibly in higher dimensional spaces, can be adopted. In this paper, a new general parametric lattice model for simplified protein conformations is proposed and investigated. It is shown how the supporting software can be consistently designed to let algorithms that operate on protein structures be implemented in a lattice-agnostic way. The necessary theoretical foundations are developed and organically presented, pinpointing the role of the concept of main directions in lattice-agnostic model handling. Subsequently, the model features across dimensions and lattice types are explored in tests performed on benchmark protein sequences, using a Python implementation. Simulations give insights on the use of square and triangular lattices in a range of dimensions. The trend of potential minimum for sequences of different lengths, varying the lattice dimension, is uncovered. Moreover, an extensive quantitative characterization of the usage of the so-called "move types" is reported for the first time. The proposed general framework for the development of lattice models is simple yet complete, and an object-oriented architecture can be proficiently employed for the supporting software, by designing ad-hoc classes. The proposed framework represents a new general viewpoint that potentially subsumes a number of solutions previously studied. The adoption of the described model pushes to look at protein structure issues from a more general and essential perspective, making

  11. Software for evaluating magnetic induction field generated by power lines: implementation of a new algorithm

    International Nuclear Information System (INIS)

    Comelli, M.; Benes, M.; Bampo, A.; Villalta, R.

    2006-01-01

    The Regional Environment Protection Agency of Friuli Venezia Giulia (A.R.P.A. F.V.G., Italy) has performed an analysis on existing software designed to calculate magnetic induction field generated by power lines. As far as the agency requirements are concerned the tested programs display some difficulties in the immediate processing of electrical and geometrical data supplied by plant owners, and in certain cases turn out to be inadequate in representing complex configurations of power lines. Furthermore, none of them is preset for cyclic calculus to determine the time evolution of induction in a certain exposure area. Finally, the output data are not immediately importable by ArcView, the G.I.S. used by A.R.P.A. F.V.G., and it is not always possible to implement the territory orography to determine the field at specified heights above the ground. P.h.i.d.e.l., an innovative software, tackles and works out al l the above mentioned problems. The power line wires interested in its implementation are represented by poly lines, and the field is analytically calculated, with no further approximation, not even when more power lines are concerned. Therefore, the obtained results, when compared with those of other programs, are the closest to experimental measurements. The output data can be employed both in G.I.S. and Excel environments, allowing the immediate overlaying of digital cartography and the determining of the 3 and 10 μT bands, in compliance with the Italian Decree of the President of the Council of Ministers of 8 July 2003. (authors)

  12. Architecture and Implementation of OpenPET Firmware and Embedded Software

    International Nuclear Information System (INIS)

    Abu-Nimeh, Faisal T.; Ito, Jennifer; Moses, William W.; Peng, Qiyu; Choong, Woon-Seng

    2016-01-01

    OpenPET is an open source, modular, extendible, and high-performance platform suitable for multi-channel data acquisition and analysis. Due to the versatility of the hardware, firmware, and software architectures, the platform is capable of interfacing with a wide variety of detector modules not only in medical imaging but also in homeland security applications. Analog signals from radiation detectors share similar characteristics-a pulse whose area is proportional to the deposited energy and whose leading edge is used to extract a timing signal. As a result, a generic design method of the platform is adopted for the hardware, firmware, and software architectures and implementations. The analog front-end is hosted on a module called a Detector Board, where each board can filter, combine, timestamp, and process multiple channels independently. The processed data is formatted and sent through a backplane bus to a module called Support Board, where 1 Support Board can host up to eight Detector Board modules. The data in the Support Board, coming from 8 Detector Board modules, can be aggregated or correlated (if needed) depending on the algorithm implemented or runtime mode selected. It is then sent out to a computer workstation for further processing. The number of channels (detector modules), to be processed, mandates the overall OpenPET System Configuration, which is designed to handle up to 1,024 channels using 16-channel Detector Boards in the Standard System Configuration and 16,384 channels using 32-channel Detector Boards in the Large System Configuration.

  13. Measuring the impact of computer resource quality on the software development process and product

    Science.gov (United States)

    Mcgarry, Frank; Valett, Jon; Hall, Dana

    1985-01-01

    The availability and quality of computer resources during the software development process was speculated to have measurable, significant impact on the efficiency of the development process and the quality of the resulting product. Environment components such as the types of tools, machine responsiveness, and quantity of direct access storage may play a major role in the effort to produce the product and in its subsequent quality as measured by factors such as reliability and ease of maintenance. During the past six years, the NASA Goddard Space Flight Center has conducted experiments with software projects in an attempt to better understand the impact of software development methodologies, environments, and general technologies on the software process and product. Data was extracted and examined from nearly 50 software development projects. All were related to support of satellite flight dynamics ground-based computations. The relationship between computer resources and the software development process and product as exemplified by the subject NASA data was examined. Based upon the results, a number of computer resource-related implications are provided.

  14. ALGORITHM OF PLACEMENT OF VIDEO SURVEILLANCE CAMERAS AND ITS SOFTWARE IMPLEMENTATION

    Directory of Open Access Journals (Sweden)

    Loktev Alexey Alexeevich

    2012-10-01

    Full Text Available Comprehensive distributed safety, control, and monitoring systems applied by companies and organizations of different ownership structure play a substantial role in the present-day society. Video surveillance elements that ensure image processing and decision making in automated or automatic modes are the essential components of new systems. This paper covers the modeling of video surveillance systems installed in buildings, and the algorithm, or pattern, of video camera placement with due account for nearly all characteristics of buildings, detection and recognition facilities, and cameras themselves. This algorithm will be subsequently implemented as a user application. The project contemplates a comprehensive approach to the automatic placement of cameras that take account of their mutual positioning and compatibility of tasks. The project objective is to develop the principal elements of the algorithm of recognition of a moving object to be detected by several cameras. The image obtained by different cameras will be processed. Parameters of motion are to be identified to develop a table of possible options of routes. The implementation of the recognition algorithm represents an independent research project to be covered by a different article. This project consists in the assessment of the degree of complexity of an algorithm of camera placement designated for identification of cases of inaccurate algorithm implementation, as well as in the formulation of supplementary requirements and input data by means of intercrossing sectors covered by neighbouring cameras. The project also contemplates identification of potential problems in the course of development of a physical security and monitoring system at the stage of the project design development and testing. The camera placement algorithm has been implemented as a software application that has already been pilot tested on buildings and inside premises that have irregular dimensions. The

  15. Design, implementation and verification of software code for radiation dose assessment based on simple generic environmental model

    International Nuclear Information System (INIS)

    I Putu Susila; Arif Yuniarto

    2017-01-01

    Radiation dose assessment to determine the potential of radiological impacts of various installations within nuclear facility complex is necessary to ensure environmental and public safety. A simple generic model-based method for calculating radiation doses caused by the release of radioactive substances into the environment has been published by the International Atomic Energy Agency (IAEA) as the Safety Report Series No. 19 (SRS-19). In order to assist the application of the assessment method and a basis for the development of more complex assessment methods, an open-source based software code has been designed and implemented. The software comes with maps and is very easy to be used because assessment scenarios can be done through diagrams. Software verification was performed by comparing its result to SRS-19 and CROM software calculation results. Dose estimated by SRS-19 are higher compared to the result of developed software. However, these are still acceptable since dose estimation in SRS-19 is based on conservative approach. On the other hand, compared to CROM software, the same results for three scenarios and a non-significant difference of 2.25 % in another scenario were obtained. These results indicate the correctness of our implementation and implies that the developed software is ready for use in real scenario. In the future, the addition of various features and development of new model need to be done to improve the capability of software that has been developed. (author)

  16. Use of a quality improvement tool, the prioritization matrix, to identify and prioritize triage software algorithm enhancement.

    Science.gov (United States)

    North, Frederick; Varkey, Prathiba; Caraballo, Pedro; Vsetecka, Darlene; Bartel, Greg

    2007-10-11

    Complex decision support software can require significant effort in maintenance and enhancement. A quality improvement tool, the prioritization matrix, was successfully used to guide software enhancement of algorithms in a symptom assessment call center.

  17. Manual on quality assurance for computer software related to the safety of nuclear power plants

    International Nuclear Information System (INIS)

    1988-01-01

    The objective of the Manual is to provide guidance in the assurance of quality of specification, design, maintenance and use of computer software related to items and activities important to safety (hereinafter referred to as safety related) in nuclear power plants. This guidance is consistent with, and supplements, the requirements and recommendations of Quality Assurance for Safety in Nuclear Power Plants: A Code of Practice, 50-C-QA, and related Safety Guides on quality assurance for nuclear power plants. Annex A identifies the IAEA documents referenced in the Manual. The Manual is intended to be of use to all those who, in any way, are involved with software for safety related applications for nuclear power plants, including auditors who may be called upon to audit management systems and product software. Figs

  18. Improving Code Quality of the Compact Muon Solenoid Electromagnetic Calorimeter Control Software to Increase System Maintainability

    CERN Multimedia

    Holme, Oliver; Dissertori, Günther; Djambazov, Lubomir; Lustermann, Werner; Zelepoukine, Serguei

    2013-01-01

    The Detector Control System (DCS) software of the Electromagnetic Calorimeter (ECAL) of the Compact Muon Solenoid (CMS) experiment at CERN is designed primarily to enable safe and efficient operation of the detector during Large Hadron Collider (LHC) data-taking periods. Through a manual analysis of the code and the adoption of ConQAT [1], a software quality assessment toolkit, the CMS ECAL DCS team has made significant progress in reducing complexity and improving code quality, with observable results in terms of a reduction in the effort dedicated to software maintenance. This paper explains the methodology followed, including the motivation to adopt ConQAT, the specific details of how this toolkit was used and the outcomes that have been achieved. [1] ConQAT, Continuous Quality Assessment Toolkit; https://www.conqat.org/

  19. When Spreadsheets Become Software - Quality Control Challenges and Approaches - 13360

    International Nuclear Information System (INIS)

    Fountain, Stefanie A.; Chen, Emmie G.; Beech, John F.; Wyatt, Elizabeth E.; Quinn, Tanya B.; Seifert, Robert W.; Bonczek, Richard R.

    2013-01-01

    As part of a preliminary waste acceptance criteria (PWAC) development, several commercial models were employed, including the Hydrologic Evaluation of Landfill Performance model (HELP) [1], the Disposal Unit Source Term - Multiple Species model (DUSTMS) [2], and the Analytical Transient One, Two, and Three-Dimensional model (AT123D) [3]. The results of these models were post-processed in MS Excel spreadsheets to convert the model results to alternate units, compare the groundwater concentrations to the groundwater concentration thresholds, and then to adjust the waste contaminant masses (based on average concentration over the waste volume) as needed in an attempt to achieve groundwater concentrations at the limiting point of assessment that would meet the compliance concentrations while maximizing the potential use of the landfill (i.e., maximizing the volume of projected waste being generated that could be placed in the landfill). During the course of the PWAC calculation development, one of the Microsoft (MS) Excel spreadsheets used to post-process the results of the commercial model packages grew to include more than 575,000 formulas across 18 worksheets. This spreadsheet was used to assess six base scenarios as well as nine uncertainty/sensitivity scenarios. The complexity of the spreadsheet resulted in the need for a rigorous quality control (QC) procedure to verify data entry and confirm the accuracy of formulas. (authors)

  20. When Spreadsheets Become Software - Quality Control Challenges and Approaches - 13360

    Energy Technology Data Exchange (ETDEWEB)

    Fountain, Stefanie A.; Chen, Emmie G.; Beech, John F. [Geosyntec Consultants, Inc., 1255 Roberts Boulevard NW, Suite 200, Kennesaw, GA 30144 (United States); Wyatt, Elizabeth E. [LATA Environmental Services of Kentucky, LLC, 761 Veterans Ave, Kevil, KY 42053 (United States); Quinn, Tanya B. [Geosyntec Consultants, Inc., 2002 Summit Boulevard NE, Suite 885, Atlanta, GA 30319 (United States); Seifert, Robert W. [Portsmouth/Paducah Project Office, United States Department of Energy, 5600 Hobbs Rd, Kevil, KY 42053 (United States); Bonczek, Richard R. [Portsmouth/Paducah Project Office, United States Department of Energy, 1017 Majestic Drive, Lexington, KY 40513 (United States)

    2013-07-01

    As part of a preliminary waste acceptance criteria (PWAC) development, several commercial models were employed, including the Hydrologic Evaluation of Landfill Performance model (HELP) [1], the Disposal Unit Source Term - Multiple Species model (DUSTMS) [2], and the Analytical Transient One, Two, and Three-Dimensional model (AT123D) [3]. The results of these models were post-processed in MS Excel spreadsheets to convert the model results to alternate units, compare the groundwater concentrations to the groundwater concentration thresholds, and then to adjust the waste contaminant masses (based on average concentration over the waste volume) as needed in an attempt to achieve groundwater concentrations at the limiting point of assessment that would meet the compliance concentrations while maximizing the potential use of the landfill (i.e., maximizing the volume of projected waste being generated that could be placed in the landfill). During the course of the PWAC calculation development, one of the Microsoft (MS) Excel spreadsheets used to post-process the results of the commercial model packages grew to include more than 575,000 formulas across 18 worksheets. This spreadsheet was used to assess six base scenarios as well as nine uncertainty/sensitivity scenarios. The complexity of the spreadsheet resulted in the need for a rigorous quality control (QC) procedure to verify data entry and confirm the accuracy of formulas. (authors)

  1. Challenges of Implementing Free and Open Source Software (FOSS): Evidence from the Indian Educational Setting

    Science.gov (United States)

    Thankachan, Briju; Moore, David Richard

    2017-01-01

    The use of Free and Open Source Software (FOSS), a subset of Information and Communication Technology (ICT), can reduce the cost of purchasing software. Despite the benefit in the initial purchase price of software, deploying software requires total cost that goes beyond the initial purchase price. Total cost is a silent issue of FOSS and can only…

  2. Analyst Tools and Quality Control Software for the ARM Data System

    Energy Technology Data Exchange (ETDEWEB)

    Moore, S.T.

    2004-12-14

    ATK Mission Research develops analyst tools and automated quality control software in order to assist the Atmospheric Radiation Measurement (ARM) Data Quality Office with their data inspection tasks. We have developed a web-based data analysis and visualization tool, called NCVweb, that allows for easy viewing of ARM NetCDF files. NCVweb, along with our library of sharable Interactive Data Language procedures and functions, allows even novice ARM researchers to be productive with ARM data with only minimal effort. We also contribute to the ARM Data Quality Office by analyzing ARM data streams, developing new quality control metrics, new diagnostic plots, and integrating this information into DQ HandS - the Data Quality Health and Status web-based explorer. We have developed several ways to detect outliers in ARM data streams and have written software to run in an automated fashion to flag these outliers.

  3. Model reliability and software quality assurance in simulation of nuclear fuel waste management systems

    International Nuclear Information System (INIS)

    Oeren, T.I.; Elzas, M.S.; Sheng, G.; Wageningen Agricultural Univ., Netherlands; McMaster Univ., Hamilton, Ontario)

    1985-01-01

    As is the case with all scientific simulation studies, computerized simulation of nuclear fuel waste management systems can introduce and hide various types of errors. Frameworks to clarify issues of model reliability and software quality assurance are offered. Potential problems with reference to the main areas of concern for reliability and quality are discussed; e.g., experimental issues, decomposition, scope, fidelity, verification, requirements, testing, correctness, robustness are treated with reference to the experience gained in the past. A list comprising over 80 most common computerization errors is provided. Software tools and techniques used to detect and to correct computerization errors are discussed

  4. Lightweight and Continuous Architectural Software Quality Assurance using the aSQA Technique

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak; Hansen, Klaus Marius; Lindstrøm, Bo

    2010-01-01

    In this paper, we present a novel technique for assessing and prioritizing architectural quality in large-scale software development projects. The technique can be applied with relatively little effort by software architects and thus suited for agile development in which quality attributes can...... be assessed and prioritized, e.g., within each development sprint. We outline the processes and metrics embodied in the technique, and report initial experiences on the benefits and liabilities. In conclusion, the technique is considered valuable and a viable tool, and has benefits in an architectural...

  5. Implementing Quality Service-Learning Programs in Community Colleges

    Science.gov (United States)

    Vaknin, Lauren Weiner; Bresciani, Marilee J.

    2013-01-01

    This cross-case comparative study at Western Community College and the University of the Coast explored through a constructive lens the characteristics that lead to sustainable, high quality service-learning programs and how they are implemented at institutions of higher education. The researchers determined that both Western Community College and…

  6. Implementation of quality control systematics for personnel monitoring services

    International Nuclear Information System (INIS)

    Franco, J.O.A.

    1984-01-01

    The implementation of statistical quality control techniques used in industrial practise is proposed to dosimetric services. 'Control charts' and 'sampling inspection' are adapted respectively for control of measuring process and of dose results produced in routine. A chapter on Radiation Protection and Personnel Monitoring was included. (M.A.C.) [pt

  7. School Indoor Air Quality Assessment and Program Implementation.

    Science.gov (United States)

    Prill, R.; Blake, D.; Hales, D.

    This paper describes the effectiveness of a three-step indoor air quality (IAQ) program implemented by 156 schools in the states of Washington and Idaho during the 2000-2001 school year. An experienced IAQ/building science specialist conducted walk-through assessments at each school. These assessments documented deficiencies and served as an…

  8. Determinants of quality management systems implementation in hospitals

    NARCIS (Netherlands)

    Wardhani, Viera; Utarini, Adi; van Dijk, Jitse Pieter; Post, Doeke; Groothoff, Johan Willem

    Objective: To identify the problems and facilitating factors in the implementation of quality management system (QMS) in hospitals through a systematic review. Method: A search strategy was pet-formed on the Medline database for articles written in English published between 1992 and early 2006.

  9. Some key principles in implementing knowledge management: the views of employees in a small software company

    Directory of Open Access Journals (Sweden)

    Clair Wagner

    2002-09-01

    Full Text Available As the Knowledge Era replaces the Information Age the challenge is to transform information into knowledge and use this knowledge to create and sustain the competitive advantage of organisations. In this article four key principles for implementing knowledge management are discussed. These principles were derived from a survey and discussions with employees at a small software company. The four principles are: the need for a common definition of and vision for knowledge management, the interdependence between the human and technology track, the way employees punctuate ‘knowing’ and the structure of a reward system for sharing knowledge. Opsomming Organisasies word gekonfronteer met unieke uitdagings soos wat die Kennisera die Inligtingsera vervang. Een van die grootste uitdagings is om inligting te verander na betekenisvolle kennis en om die kennis te gebruik om ‘n kompeterende voorsprong te skep en te bou. In hierdie artikel word vier kern beginsels vir die implementering van kennisbestuur bespreek. Die beginsels is geskoei op ‘n opname van en onderhoude met werknemers in ‘n klein sagteware besigheid. Die vier beginsels is: die behoefte aan ‘n definisie en visie van kennisbestuur, die interafhanklikheid tussen mense en tegnologie, die wyse waarop werknemers kennis punktueer en die skep van ‘n beloningstruktuur vir die deel van kennis.

  10. Implementation of New Process Models for Tailored Polymer Composite Structures into Processing Software Packages

    International Nuclear Information System (INIS)

    Nguyen, Ba Nghiep; Jin, Xiaoshi; Wang, Jin; Phelps, Jay; Tucker, Charles L. III; Kunc, Vlastimil; Bapanapalli, Satish K.; Smith, Mark T.

    2010-01-01

    This report describes the work conducted under the Cooperative Research and Development Agreement (CRADA) (Nr. 260) between the Pacific Northwest National Laboratory (PNNL) and Autodesk, Inc. to develop and implement process models for injection-molded long-fiber thermoplastics (LFTs) in processing software packages. The structure of this report is organized as follows. After the Introduction Section (Section 1), Section 2 summarizes the current fiber orientation models developed for injection-molded short-fiber thermoplastics (SFTs). Section 3 provides an assessment of these models to determine their capabilities and limitations, and the developments needed for injection-molded LFTs. Section 4 then focuses on the development of a new fiber orientation model for LFTs. This model is termed the anisotropic rotary diffusion - reduced strain closure (ARD-RSC) model as it explores the concept of anisotropic rotary diffusion to capture the fiber-fiber interaction in long-fiber suspensions and uses the reduced strain closure method of Wang et al. to slow down the orientation kinetics in concentrated suspensions. In contrast to fiber orientation modeling, before this project, no standard model was developed to predict the fiber length distribution in molded fiber composites. Section 5 is therefore devoted to the development of a fiber length attrition model in the mold. Sections 6 and 7 address the implementations of the models in AMI, and the conclusions drawn from this work is presented in Section 8.

  11. SME's Implementing a Quality Management System- Risks and Opportunities

    Directory of Open Access Journals (Sweden)

    Luise ZEININGER

    2017-12-01

    Full Text Available Part of our daily lives, risk ceased to be just something that managers are desperate to alleviate if not eliminate. It is now rightfully considered a path to explore new opportunities. The last version of the ISO 9001, the 2015 edition, is asking from organization implementing it to determine the risk and opportunities associated with their main processes, the impact over relevant third parties and the context, both internal and external. With this duality as starting point, we have investigated among SME's implementing a quality management system what would they consider to be the risk and opportunities associated to this action. Our paper presents hereinafter the partial results of a wider study on Romanian SME's management, especially the findings related to what managers consider as risk or opportunity when implementing a quality management system.

  12. Quality assurance for CORAL and COOL within the LCG software stack for the LHC experiments

    CERN Multimedia

    CERN. Geneva

    2015-01-01

    CORAL and COOL are software packages used by the LHC experiments for managing different categories of physics data using a variety of relational database technologies. The core components are written in C++, but Python bindings are also provided. CORAL is a generic relational access layer, while COOL includes the implementation of a specific relational data model and optimization of SQL queries for "conditions data". The software is the result of more than 10 years of development in colaboration between the IT department and the LHC experiments. The packages are built and released within the LCG software stack, for which automatic nightly builds and release installations are provided by PH-SFT (cmake, jenkins, cdash) for many different platforms, compilers and software version configurations. Test-driven development and functional tests of both C++ and Python components (CppUnit, unittest) have been key elements in the success of the projects. Dedicated test suites have also been prepared to commission and ma...

  13. Implementation of quality standards in an individual monitoring service

    International Nuclear Information System (INIS)

    Hyvoenen, H.; Vartiainen, E.

    2001-01-01

    In this paper the implementation of a quality system to the procedures of an individual monitoring service (IMS) is described from the practical perspective. The IMS of the Radiation and Nuclear Safety Authority (STUK) is used as an example. The IMS of STUK monitors about 8500 persons mainly working in hospitals, industry and research centres. The current thermoluminescence dosemeter (TLD) system was introduced in 1992 and the whole service changed to TLDs in 1995. The quality system compatible with the quality standards was introduced in 1999. An application for accreditation to full EN45001 and ISO/IEC Guide 25 was made in December 1999, accreditation was achieved in August 2000 by the Finnish Accreditation Service (FINAS). The considerations needed for the quality system to fulfil the requirements of the quality standards are reported. (author)

  14. Implementation of quality standards in an individual monitoring service

    Energy Technology Data Exchange (ETDEWEB)

    Hyvoenen, H.; Vartiainen, E

    2001-07-01

    In this paper the implementation of a quality system to the procedures of an individual monitoring service (IMS) is described from the practical perspective. The IMS of the Radiation and Nuclear Safety Authority (STUK) is used as an example. The IMS of STUK monitors about 8500 persons mainly working in hospitals, industry and research centres. The current thermoluminescence dosemeter (TLD) system was introduced in 1992 and the whole service changed to TLDs in 1995. The quality system compatible with the quality standards was introduced in 1999. An application for accreditation to full EN45001 and ISO/IEC Guide 25 was made in December 1999, accreditation was achieved in August 2000 by the Finnish Accreditation Service (FINAS). The considerations needed for the quality system to fulfil the requirements of the quality standards are reported. (author)

  15. Implementation of quality control program in radiodiagnostic services

    International Nuclear Information System (INIS)

    Herrera S, A.; Roas Z, N.

    1995-01-01

    This monograph is the first version of the implementation of the quality control programme in radiology diagnostic services. Here all information related to diagnostic quality to better radiation protection to patients and personnel was collected. The programme was implemented on the X-ray equipment at three hospitals (named hospital A, hospital B and hospital C) and included the evaluation of technical parameters such as kilovolts, exposition time, filtration, fields. In addition, dark room, chassis and image intensifiers were also evaluated. The procedures to carry out the quality control and the manner in which the observations, conclusions and recommendations should be formulated are based on documents issued by the International Commission on Radiological Protection (I.C.R.P.), International Atomic Energy Agency (I.A.E.A.) and World Health Organization (W.H.O.)

  16. Implementation in free software of the PWR type university nucleo electric simulator (SU-PWR)

    International Nuclear Information System (INIS)

    Valle H, J.; Hidago H, F.; Morales S, J.B.

    2007-01-01

    Presently work is shown like was carried out the implementation of the University Simulator of Nucleo-electric type PWR (SU-PWR). The implementation of the simulator was carried out in a free software simulation platform, as it is Scilab, what offers big advantages that go from the free use and without cost of the product, until the codes modification so much of the system like of the program with the purpose of to improve it or to adapt it to future routines and/or more advanced graphic interfaces. The SU-PWR shows the general behavior of a PWR nuclear plant (Pressurized Water Reactor) describing the dynamics of the plant from the generation process of thermal energy in the nuclear fuel, going by the process of energy transport toward the coolant of the primary circuit the one which in turn transfers this energy to the vapor generators of the secondary circuit where the vapor is expanded by means of turbines that in turn move the electric generator producing in this way the electricity. The pressurizer that is indispensable for the process is also modeled. Each one of these stages were implemented in scicos that is the Scilab tool specialized in the simulation. The simulation was carried out by means of modules that contain the differential equation that mathematically models each stage or equipment of the PWR plant. The result is a series of modules that based on certain entrances and characteristic of the system they generate exits that in turn are the entrance to other module. Because the SU-PWR is an experimental project in early phase, it is even work and modifications to carry out, for what the models that are presented in this work can vary a little the being integrated to the whole system to simulate, but however they already show clearly the operation and the conformation of the plant. (Author)

  17. Digital radiography: optimization of image quality and dose using multi-frequency software.

    Science.gov (United States)

    Precht, H; Gerke, O; Rosendahl, K; Tingberg, A; Waaler, D

    2012-09-01

    New developments in processing of digital radiographs (DR), including multi-frequency processing (MFP), allow optimization of image quality and radiation dose. This is particularly promising in children as they are believed to be more sensitive to ionizing radiation than adults. To examine whether the use of MFP software reduces the radiation dose without compromising quality at DR of the femur in 5-year-old-equivalent anthropomorphic and technical phantoms. A total of 110 images of an anthropomorphic phantom were imaged on a DR system (Canon DR with CXDI-50 C detector and MLT[S] software) and analyzed by three pediatric radiologists using Visual Grading Analysis. In addition, 3,500 images taken of a technical contrast-detail phantom (CDRAD 2.0) provide an objective image-quality assessment. Optimal image-quality was maintained at a dose reduction of 61% with MLT(S) optimized images. Even for images of diagnostic quality, MLT(S) provided a dose reduction of 88% as compared to the reference image. Software impact on image quality was found significant for dose (mAs), dynamic range dark region and frequency band. By optimizing image processing parameters, a significant dose reduction is possible without significant loss of image quality.

  18. Implementation Of Quality Management System For Irradiation Processing Services

    Science.gov (United States)

    Lungu, Ion-Bogdan; Manea, Maria-Mihaela

    2015-07-01

    In today's market, due to an increasing competitiveness, quality management has set itself as an indispensable tool and a reference point for every business. It is ultimately focused on customer satisfaction which is a stringent factor for every business. Implementing and maintaining a QMS is a rather difficult, time consuming and expensive process which must be done with respect of many factors. The aim of this paper is to present a case study for implementing QMS ISO 9001 in a gamma irradiation treatment service provider. The research goals are the identification of key benefits, reasons, advantages, disadvantages, drawbacks etc for a successful QMS implementation and use. Finally, the expected results focus on creating a general framework for implementing an efficient QMS plan that can be easily adapted to other kind of services and markets.

  19. SOFTWARE IMPLEMENTATION FOR SCORM CONTENT MIGRATION IN THE LEARNING MANAGEMENT SYSTEM

    Directory of Open Access Journals (Sweden)

    Y. B. Popova

    2017-01-01

    Full Text Available Using of learning management systems increases the possibility of teachers and students in achieving their goals in education. Such systems provide learning content, help to organize and to monitor training progress, help to collect statistics. However, the transition from one LMS to another there is a problem of content migration, because all training materials and tests should either be recreated, or somehow be migrated to the new system. Content migration by hand is a very time-consuming process, so the leading developers of the learning management systems developed a standard for the organization and storage of content, called SCORM (Eng., Sharable Content Object Reference Model. Created by this standard, the content must migrate to the learning management system provided its support for these systems. SCORM standard allows you to create training content that is not dependent on the learning management system, but the loosely embedded in it. This approach enables teachers to develop unique courses and put them free available or for sale in the Internet for all interested persons, and to use educational content created by the best specialists around the world to carry out their activities. The content on the SCORM standard imposes certain requirements on the learning management systems, as they do not distort the training content and properly interact with the tests. The aim of this article is a software implementation of a content migration by SCORM standard from other learning management systems in its own development used at the Software Department of the Faculty of Information Technology and Robotics of the Belarusian National Technical University.

  20. COST OF QUALITY MODELS AND THEIR IMPLEMENTATION IN MANUFACTURING FIRMS

    Directory of Open Access Journals (Sweden)

    N.M. Vaxevanidis

    2009-03-01

    Full Text Available In order to improve quality, an organization must take into account the costs associated with achieving quality since the objective of continuous improvement programs is not only to meet customer requirements, but also to do it at the lowest, possible, cost. This can only obtained by reducing the costs needed to achieve quality, and the reduction of these costs is only possible if they are identified and measured. Therefore, measuring and reporting the cost of quality (CoQ should be considered an important issue for achieving quality excellence. To collect quality costs an organization needs to adopt a framework to classify costs; however, there is no general agreement on a single broad definition of quality costs. CoQ is usually understood as the sum of conformance plus non-conformance costs, where cost of conformance is the price paid for prevention of poor quality (for example, inspection and quality appraisal and cost of non-conformance is the cost of poor quality caused by product and service failure (for example, rework and returns. The objective of this paper is to give a survey of research articles on the topic of CoQ; it opens with a literature review focused on existing CoQ models; then, it briefly presents the most common CoQ parameters and the metrics (indices used for monitoring CoQ. Finally, the use of CoQ models in practice, i.e., the implementation of a quality costing system and cost of quality reporting in companies is discussed, with emphasis in cases concerning manufacturing firms.

  1. A systematic literature review of open source software quality assessment models.

    Science.gov (United States)

    Adewumi, Adewole; Misra, Sanjay; Omoregbe, Nicholas; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    Many open source software (OSS) quality assessment models are proposed and available in the literature. However, there is little or no adoption of these models in practice. In order to guide the formulation of newer models so they can be acceptable by practitioners, there is need for clear discrimination of the existing models based on their specific properties. Based on this, the aim of this study is to perform a systematic literature review to investigate the properties of the existing OSS quality assessment models by classifying them with respect to their quality characteristics, the methodology they use for assessment, and their domain of application so as to guide the formulation and development of newer models. Searches in IEEE Xplore, ACM, Science Direct, Springer and Google Search is performed so as to retrieve all relevant primary studies in this regard. Journal and conference papers between the year 2003 and 2015 were considered since the first known OSS quality model emerged in 2003. A total of 19 OSS quality assessment model papers were selected. To select these models we have developed assessment criteria to evaluate the quality of the existing studies. Quality assessment models are classified into five categories based on the quality characteristics they possess namely: single-attribute, rounded category, community-only attribute, non-community attribute as well as the non-quality in use models. Our study reflects that software selection based on hierarchical structures is found to be the most popular selection method in the existing OSS quality assessment models. Furthermore, we found that majority (47%) of the existing models do not specify any domain of application. In conclusion, our study will be a valuable contribution to the community and helps the quality assessment model developers in formulating newer models and also to the practitioners (software evaluators) in selecting suitable OSS in the midst of alternatives.

  2. Taking advantage of ground data systems attributes to achieve quality results in testing software

    Science.gov (United States)

    Sigman, Clayton B.; Koslosky, John T.; Hageman, Barbara H.

    1994-01-01

    During the software development life cycle process, basic testing starts with the development team. At the end of the development process, an acceptance test is performed for the user to ensure that the deliverable is acceptable. Ideally, the delivery is an operational product with zero defects. However, the goal of zero defects is normally not achieved but is successful to various degrees. With the emphasis on building low cost ground support systems while maintaining a quality product, a key element in the test process is simulator capability. This paper reviews the Transportable Payload Operations Control Center (TPOCC) Advanced Spacecraft Simulator (TASS) test tool that is used in the acceptance test process for unmanned satellite operations control centers. The TASS is designed to support the development, test and operational environments of the Goddard Space Flight Center (GSFC) operations control centers. The TASS uses the same basic architecture as the operations control center. This architecture is characterized by its use of distributed processing, industry standards, commercial off-the-shelf (COTS) hardware and software components, and reusable software. The TASS uses much of the same TPOCC architecture and reusable software that the operations control center developer uses. The TASS also makes use of reusable simulator software in the mission specific versions of the TASS. Very little new software needs to be developed, mainly mission specific telemetry communication and command processing software. By taking advantage of the ground data system attributes, successful software reuse for operational systems provides the opportunity to extend the reuse concept into the test area. Consistency in test approach is a major step in achieving quality results.

  3. Software: our quest for excellence. Honoring 50 years of software history, progress, and process

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-06-01

    The Software Quality Forum was established by the Software Quality Assurance (SQA) Subcommittee, which serves as a technical advisory group on software engineering and quality initiatives and issues for DOE`s quality managers. The forum serves as an opportunity for all those involved in implementing SQA programs to meet and share ideas and concerns. Participation from managers, quality engineers, and software professionals provides an ideal environment for identifying and discussing issues and concerns. The interaction provided by the forum contributes to the realization of a shared goal--high quality software product. Topics include: testing, software measurement, software surety, software reliability, SQA practices, assessments, software process improvement, certification and licensing of software professionals, CASE tools, software project management, inspections, and management`s role in ensuring SQA. The bulk of this document consists of vugraphs. Selected papers have been indexed separately for inclusion in the Energy Science and Technology Database.

  4. Assessing the impact of continuous quality improvement/total quality management: concept versus implementation.

    Science.gov (United States)

    Shortell, S M; O'Brien, J L; Carman, J M; Foster, R W; Hughes, E F; Boerstler, H; O'Connor, E J

    1995-01-01

    OBJECTIVE: This study examines the relationships among organizational culture, quality improvement processes and selected outcomes for a sample of up to 61 U. S. hospitals. DATA SOURCES AND STUDY SETTING: Primary data were collected from 61 U. S. hospitals (located primarily in the midwest and the west) on measures related to continuous quality improvement/total quality management (CQI/TQM), organizational culture, implementation approaches, and degree of quality improvement implementation based on the Baldrige Award criteria. These data were combined with independently collected data on perceived impact and objective measures of clinical efficiency (i.e., charges and length of stay) for six clinical conditions. STUDY DESIGN: The study involved cross-sectional examination of the named relationships. DATA COLLECTION/EXTRACTION METHODS: Reliable and valid scales for the organizational culture and quality improvement implementation measures were developed based on responses from over 7,000 individuals across the 61 hospitals with an overall completion rate of 72 percent. Independent data on perceived impact were collected from a national survey and independent data on clinical efficiency from a companion study of managed care. PRINCIPAL FINDINGS: A participative, flexible, risk-taking organizational culture was significantly related to quality improvement implementation. Quality improvement implementation, in turn, was positively associated with greater perceived patient outcomes and human resource development. Larger-size hospitals experienced lower clinical efficiency with regard to higher charges and higher length of stay, due in part to having more bureaucratic and hierarchical cultures that serve as a barrier to quality improvement implementation. CONCLUSIONS: What really matters is whether or not a hospital has a culture that supports quality improvement work and an approach that encourages flexible implementation. Larger-size hospitals face more difficult

  5. Total Quality Management Implementation and Guest Satisfaction in Hospitality

    Directory of Open Access Journals (Sweden)

    Miroslav Knežević

    2017-02-01

    Full Text Available Total quality management (TQM has become a modern system of constant improvement of the quality of all company activities. The purpose of this study is to measure the expectations and satisfaction of the guests concerning the attribute quality of the hotel product. Furthermore obtained results were compared in such a way as to analyse particularly the reviews of hotels which have implemented TQM and have the ISO 9001 certificates with reviews from hotels which have not implemented TQM and do not have the ISO 9001 certificates. The conducted analysis included 55 hotels in Serbia belonging to the 4- and 5-star categories, i.e. 1308 guests who have stayed in them. The results show that between the observed groups of guests there are fewer differences in expectations than in perception, and that generally speaking guests who have stayed in the hotels that have implemented TQM are more satisfied. The biggest difference concerning the guest satisfaction with the quality of service in the observed hotels is noticeable in relation to the employees and the value-for-money.

  6. Software quality assurance and information management, October 1986 to October 1992

    International Nuclear Information System (INIS)

    Hill, I.E.

    1993-01-01

    This report describes the work carried out by Cedar Design Systems Limited under contract PECD 7/9/384. The brief for the contract was initially to provide advice on Software Quality Assurance (SQA) as part of the CEC PACOMA project. This was later extended to include further SQA and information management tasks specific to the HMIP Radioactive Waste Disposal Assessments Research Programme. (Author)

  7. 76 FR 54800 - International Business Machines (IBM), Software Group Business Unit, Quality Assurance Group, San...

    Science.gov (United States)

    2011-09-02

    ... DEPARTMENT OF LABOR Employment and Training Administration [TA-W-74,554] International Business Machines (IBM), Software Group Business Unit, Quality Assurance Group, San Jose, California; Notice of Negative Determination on Reconsideration On January 21, 2011, the Department of Labor (Department) issued an Affirmative Determination Regarding...

  8. Implementation of Quality Management System for Historical Building Conservation

    Directory of Open Access Journals (Sweden)

    Zahari N.F.

    2014-01-01

    Full Text Available The main objectives of this study are twofold. Firstly, to identify the implementation of ISO 9001 procedures being used as references for conservation works and the development of Quality Management System (QMS guidelines. Data were solicited from three (3 conservation areas. The analysis involved of descriptive approach and statistical methods. The findings revealed that QMS is not structurally established, implemented and enforced as part of conservation practice in Malaysia. From the findings, the authors hope to give clear perception to the reader on current preservation practice and the existence of QMS with reference to ISO 9001 for future conservation mechanism.

  9. Implementations of service oriented architecture and agile software development: What works and what are the challenges?

    NARCIS (Netherlands)

    Schramm, Milan; Daneva, Maia

    2016-01-01

    Today many organizations use service-oriented architecture and agile software development as their software paradigms. While both certainly have their advantages, in the fields of Empirical Software Engineering and Information Systems these have been treated in relative isolation and their impact on

  10. Sandia software guidelines, Volume 4: Configuration management

    Energy Technology Data Exchange (ETDEWEB)

    1992-06-01

    This volume is one in a series of Sandia Software Guidelines for use in producing quality software within Sandia National Laboratories. This volume is based on the IEEE standard and guide for software configuration management. The basic concepts and detailed guidance on implementation of these concepts are discussed for several software project types. Example planning documents for both projects and organizations are included.

  11. 76 FR 44535 - Revisions to the California State Implementation Plan, Northern Sierra Air Quality Management...

    Science.gov (United States)

    2011-07-26

    ... the California State Implementation Plan, Northern Sierra Air Quality Management District, Sacramento Metropolitan Air Quality Management District, and South Coast Air Quality Management District AGENCY... the Northern Sierra Air Quality Management District (NSAQMD), Sacramento Metropolitan Air Quality...

  12. An empirical evaluation of software quality assurance practices and challenges in a developing country: a comparison of Nigeria and Turkey.

    Science.gov (United States)

    Sowunmi, Olaperi Yeside; Misra, Sanjay; Fernandez-Sanz, Luis; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    The importance of quality assurance in the software development process cannot be overemphasized because its adoption results in high reliability and easy maintenance of the software system and other software products. Software quality assurance includes different activities such as quality control, quality management, quality standards, quality planning, process standardization and improvement amongst others. The aim of this work is to further investigate the software quality assurance practices of practitioners in Nigeria. While our previous work covered areas on quality planning, adherence to standardized processes and the inherent challenges, this work has been extended to include quality control, software process improvement and international quality standard organization membership. It also makes comparison based on a similar study carried out in Turkey. The goal is to generate more robust findings that can properly support decision making by the software community. The qualitative research approach, specifically, the use of questionnaire research instruments was applied to acquire data from software practitioners. In addition to the previous results, it was observed that quality assurance practices are quite neglected and this can be the cause of low patronage. Moreover, software practitioners are neither aware of international standards organizations or the required process improvement techniques; as such their claimed standards are not aligned to those of accredited bodies, and are only limited to their local experience and knowledge, which makes it questionable. The comparison with Turkey also yielded similar findings, making the results typical of developing countries. The research instrument used was tested for internal consistency using the Cronbach's alpha, and it was proved reliable. For the software industry in developing countries to grow strong and be a viable source of external revenue, software assurance practices have to be taken seriously

  13. Image Quality Improvement after Implementation of a CT Accreditation Program

    International Nuclear Information System (INIS)

    Kim, You Sung; Jung, Seung Eun; Choi, Byung Gil; Shin, Yu Ri; Hwang, Seong Su; Ku, Young Mi; Lim, Yeon Soo; Lee, Jae Mun

    2010-01-01

    The purpose of this study was to evaluate any improvement in the quality of abdominal CTs after the utilization of the nationally based accreditation program. Approval was obtained from the Institutional Review Board, and informed consent was waived. We retrospectively analyzed 1,011 outside abdominal CTs, from 2003 to 2007. We evaluated images using a fill-up sheet form of the national accreditation program, and subjectively by grading for the overall CT image quality. CT scans were divided into two categories according to time periods; before and after the implementation of the accreditation program. We compared CT scans between two periods according to parameters pertaining to the evaluation of images. We determined whether there was a correlation between the results of a subjective assessment of the image quality and the evaluation scores of the clinical image. The following parameters were significantly different after the implementation of the accreditation program: identifying data, display parameters, scan length, spatial and contrast resolution, window width and level, optimal contrast enhancement, slice thickness, and total score. The remaining parameters were not significantly different between scans obtained from the two different periods: scan parameters, film quality, and artifacts. After performing the CT accreditation program, the quality of the outside abdominal CTs show marked improvement, especially for the parameters related to the scanning protocol

  14. ThermoData Engine (TDE): software implementation of the dynamic data evaluation concept. 9. Extensible thermodynamic constraints for pure compounds and new model developments.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Muzny, Chris D; Kazakov, Andrei F; Kroenlein, Kenneth; Magee, Joseph W; Abdulagatov, Ilmutdin; Frenkel, Michael

    2013-12-23

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported in this journal. The present article describes the background and implementation for new additions in latest release of TDE. Advances are in the areas of program architecture and quality improvement for automatic property evaluations, particularly for pure compounds. It is shown that selection of appropriate program architecture supports improvement of the quality of the on-demand property evaluations through application of a readily extensible collection of constraints. The basis and implementation for other enhancements to TDE are described briefly. Other enhancements include the following: (1) implementation of model-validity enforcement for specific equations that can provide unphysical results if unconstrained, (2) newly refined group-contribution parameters for estimation of enthalpies of formation for pure compounds containing carbon, hydrogen, and oxygen, (3) implementation of an enhanced group-contribution method (NIST-Modified UNIFAC) in TDE for improved estimation of phase-equilibrium properties for binary mixtures, (4) tools for mutual validation of ideal-gas properties derived through statistical calculations and those derived independently through combination of experimental thermodynamic results, (5) improvements in program reliability and function that stem directly from the recent redesign of the TRC-SOURCE Data Archival System for experimental property values, and (6) implementation of the Peng-Robinson equation of state for binary mixtures, which allows for critical evaluation of mixtures involving supercritical components. Planned future developments are summarized.

  15. Development, implementation and quality assurance of biokinetic models within CONRAD

    International Nuclear Information System (INIS)

    Nosske, D.; Birchall, A.; Blanchardon, E.; Breustedt, B.; Giussani, A.; Luciani, A.; Oeh, U.; Lopez, M. A.

    2008-01-01

    The work of the Task Group 5.2 'Research Studies on Biokinetic Models' of the CONRAD project is presented. New biokinetic models have been implemented by several European institutions. Quality assurance procedures included intercomparison of the results as well as quality assurance of model formulation. Additionally, the use of the models was examined leading to proposals of tuning parameters. Stable isotope studies were evaluated with respect to their implications to the new models, and new biokinetic models were proposed on the basis of their results. Furthermore, the development of a biokinetic model describing the effects of decorporation of actinides by diethylenetriaminepentaacetic acid treatment was initiated. (authors)

  16. Clinical audit and quality systems - practical implementation in Finland

    International Nuclear Information System (INIS)

    Jaervinen, H.

    2003-01-01

    Clinical audit is a new concept of significant importance for the quality of radiological practices, introduced by the EC Medical Exposure Directive (MED, 97/43/EURATOM). By definition, clinical audit means 'a systematic examination or review of medical radiological procedures which seeks to improve the quality and the outcome of patient care, through structured review whereby radiological practices, procedures, and results are examined against agreed standards for good medical radiological procedures, with modifications of the practices where indicated and the application of new standards if necessary'. In its most profound meaning, being introduced in the medical exposure directive, clinical audit can be seen as a review of the success in implementing the justification and optimization principles, and therefore, it is to a large extent an issue of radiation safety for the patient. According to the directive, clinical audits shall be 'carried out in accordance with national procedures'. For the last few years, parallel to the development of the MED in Europe, there has been a worldwide tendency to implement appropriate quality systems (QS) in the health care organizations, in accordance with the international quality standards (ISO 9000 series etc). Such quality systems have been applied for a long time and very widely by the industry. It is a strong belief that the development of quality systems for health care would result in equal benefits as trusted in industry, in terms of efficiency and safety of health care services. For radiological practices, the quality systems are expected to become a framework for improving the optimization of practices and for maintaining good radiation safety, as well as providing a mechanism to prevent mistakes and accidents. In some countries, like the UK and The Netherlands, there are legal requirements to establish and maintain quality systems at certain type of radiological units. In some countries and some radiological units

  17. Implementing a Quality Management System in the Medical Microbiology Laboratory.

    Science.gov (United States)

    Carey, Roberta B; Bhattacharyya, Sanjib; Kehl, Sue C; Matukas, Larissa M; Pentella, Michael A; Salfinger, Max; Schuetz, Audrey N

    2018-07-01

    This document outlines a comprehensive practical approach to a laboratory quality management system (QMS) by describing how to operationalize the management and technical requirements described in the ISO 15189 international standard. It provides a crosswalk of the ISO requirements for quality and competence for medical laboratories to the 12 quality system essentials delineated by the Clinical and Laboratory Standards Institute. The quality principles are organized under three main categories: quality infrastructure, laboratory operations, and quality assurance and continual improvement. The roles and responsibilities to establish and sustain a QMS are outlined for microbiology laboratory staff, laboratory management personnel, and the institution's leadership. Examples and forms are included to assist in the real-world implementation of this system and to allow the adaptation of the system for each laboratory's unique environment. Errors and nonconforming events are acknowledged and embraced as an opportunity to improve the quality of the laboratory, a culture shift from blaming individuals. An effective QMS encourages "systems thinking" by providing a process to think globally of the effects of any type of change. Ultimately, a successful QMS is achieved when its principles are adopted as part of daily practice throughout the total testing process continuum. Copyright © 2018 American Society for Microbiology.

  18. Implementing a DVB-T/H Receiver on a Software-Defined Radio Platform

    Directory of Open Access Journals (Sweden)

    Yong Jiang

    2009-01-01

    Full Text Available Digital multimedia broadcasting is available in more and more countries with various forms. One of the most successful forms is Digital Video Broadcasting for Terrestrial (DVB-T, which has been deployed in most countries of the world for years. In order to bring the digital multimedia broadcasting services to battery-powered handheld receivers in a mobile environment, Digital Video Broadcasting for Handheld (DVB-H has been formally adopted by ETSI. More advanced and complex digital multimedia broadcasting systems are under development, for example, the next generation of DVB-T, a.k.a. DVB-T2. Current commercial DVB-T/H receivers are usually built upon dedicated application-specific integrated circuits (ASICs. However, ASICs are not flexible for incoming evolved standards and less overall-area efficient since they cannot be efficiently reused and shared among different radio standards, when we integrate a DVB-T/H receiver into a mobile phone. This paper presents an example implementation of a DVB-T/H receiver on the prototype of Infineon Technologies' Software-Defined Radio (SDR platform called MuSIC (Multiple SIMD Cores, which is a DSP-centered and accelerator-assisted architecture and aims at battery-powered mass-market handheld terminals.

  19. Software implementation of a high speed interface between a PDP-10 and several PDP-11s

    International Nuclear Information System (INIS)

    De Mesa, N.P. III.

    1975-01-01

    The DMA10 is a high speed link between a PDP-10 and up to eight PDP-11s; specifically, the PDP-10 shares sections of its memory with the PDP-11s. The two segment concept on the PDP-10 of shared/reentrant code and non-shared code is implemented. The inclusion of read only memory on the PDP-11s allows for the development of ''PROM'' software which all the PDP-11s may share. The principal difference between the DMA10 and other communications interfaces is that it is not a block transfer device. Because of the shared memory concept the features of the DMA10 are high data bandwidth and minimal processor intervention between data transfers. Communication programs between the PDP-10 and the PDP-11 may be tested wholly in either processor, independent of the DMA10 interface. In the current mode of operation the PDP-11's simply act as device controllers. Future plans include separate operating systems in various PDP-11s

  20. Designing and Implementing a Distributed System Architecture for the Mars Rover Mission Planning Software (Maestro)

    Science.gov (United States)

    Goldgof, Gregory M.

    2005-01-01

    Distributed systems allow scientists from around the world to plan missions concurrently, while being updated on the revisions of their colleagues in real time. However, permitting multiple clients to simultaneously modify a single data repository can quickly lead to data corruption or inconsistent states between users. Since our message broker, the Java Message Service, does not ensure that messages will be received in the order they were published, we must implement our own numbering scheme to guarantee that changes to mission plans are performed in the correct sequence. Furthermore, distributed architectures must ensure that as new users connect to the system, they synchronize with the database without missing any messages or falling into an inconsistent state. Robust systems must also guarantee that all clients will remain synchronized with the database even in the case of multiple client failure, which can occur at any time due to lost network connections or a user's own system instability. The final design for the distributed system behind the Mars rover mission planning software fulfills all of these requirements and upon completion will be deployed to MER at the end of 2005 as well as Phoenix (2007) and MSL (2009).

  1. Implementation of the ATLAS trigger within the multi-threaded software framework AthenaMT

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00225867; The ATLAS collaboration

    2017-01-01

    We present an implementation of the ATLAS High Level Trigger, HLT, that provides parallel execution of trigger algorithms within the ATLAS multithreaded software framework, AthenaMT. This development will enable the ATLAS HLT to meet future challenges due to the evolution of computing hardware and upgrades of the Large Hadron Collider, LHC, and ATLAS Detector. During the LHC data-taking period starting in 2021, luminosity will reach up to three times the original design value. Luminosity will increase further, to up to 7.5 times the design value, in 2026 following LHC and ATLAS upgrades. This includes an upgrade of the ATLAS trigger architecture that will result in an increase in the HLT input rate by a factor of 4 to 10 compared to the current maximum rate of 100 kHz. The current ATLAS multiprocess framework, AthenaMP, manages a number of processes that each execute algorithms sequentially for different events. AthenaMT will provide a fully multi-threaded environment that will additionally enable concurrent ...

  2. Microsatellite Attitude Determination and Control Subsystem Design and Implementation: Software-in-the-Loop Approach

    Directory of Open Access Journals (Sweden)

    Ho-Nien Shou

    2014-01-01

    Full Text Available The paper describes the development of a microsatellite attitude determination and control subsystem (ADCS and verification of its functionality by software-in-the-loop (SIL method. The role of ADCS is to provide attitude control functions, including the de-tumbling and stabilizing the satellite angular velocity, and as well as estimating the orbit and attitude information during the satellite operation. In Taiwan, Air Force Institute of Technology (AFIT, dedicating for students to design experimental low earth orbit micro-satellite, called AFITsat. For AFITsat, the operation of the ADCS consists of three modes which are initialization mode, detumbling mode, and normal mode, respectively. During the initialization mode, ADCS collects the early orbit measurement data from various sensors so that the data can be downlinked to the ground station for further analysis. As particularly emphasized in this paper, during the detumbling mode, ADCS implements the thrusters in plus-wide modulation control method to decrease the satellite angular velocity. ADCS provides the attitude determination function for the estimation of the satellite state, during normal mode. The three modes of microsatellite adopted Kalman filter algorithm estimate microsatellite attitude. This paper will discuss using the SIL validation ADCS function and verify its feasibility.

  3. Implementing quality assurance with multiple contractors under changing regulations

    International Nuclear Information System (INIS)

    Hanrahan, T.

    1989-01-01

    This paper discusses the tools and techniques used to establish the responsibilities for quality implementation on the California Low-Level Radioactive Waste Disposal Project. The project is structured to use the traditional nuclear-oriented quality assurance criteria in the non-traditional application of scientific investigations required for site characterization activities. It has required a careful blending of approaches and mentalities from backgrounds in both nuclear QA and EPA QA as well as the selective application of guidance from both of the agencies. As a significant portion of the work was subcontracted to organizations which ranged from a small group of college professors to a multi-national corporation, the program needed to be versatile, easy to implement but yet definitive. The author describes some of the misconceptions encountered and areas of weakness to be identified and strengthened

  4. IMPLEMENTATION OF INTEGRAL SYSTEM OF QUALITY MANAGEMENT IN TOURISM

    Directory of Open Access Journals (Sweden)

    Vidoje Vujic

    2007-12-01

    Full Text Available Theory and practice have corroborated the need and usefulness of an integral approach to quality management in tourist destinations, since the destination and its touristic offer define a number of disparate participants and interested parties. An integral system of quality management is one of the models of touristic management that with the implementation of contemporary principles strives to achieve business excellence and competitory advantage. The paper determines and projects the need and importance of implementing an integral system and accordingly seeks to form a model for its development. By studying the oneness of the whole we have established the dependence and firm connections between particular norms and elements, and by analyzing them in this paper we describe the structure and associated characteristics of the whole.

  5. Report of AAPM Task Group 162: Software for planar image quality metrology.

    Science.gov (United States)

    Samei, Ehsan; Ikejimba, Lynda C; Harrawood, Brian P; Rong, John; Cunningham, Ian A; Flynn, Michael J

    2018-02-01

    The AAPM Task Group 162 aimed to provide a standardized approach for the assessment of image quality in planar imaging systems. This report offers a description of the approach as well as the details of the resultant software bundle to measure detective quantum efficiency (DQE) as well as its basis components and derivatives. The methodology and the associated software include the characterization of the noise power spectrum (NPS) from planar images acquired under specific acquisition conditions, modulation transfer function (MTF) using an edge test object, the DQE, and effective DQE (eDQE). First, a methodological framework is provided to highlight the theoretical basis of the work. Then, a step-by-step guide is included to assist in proper execution of each component of the code. Lastly, an evaluation of the method is included to validate its accuracy against model-based and experimental data. The code was built using a Macintosh OSX operating system. The software package contains all the source codes to permit an experienced user to build the suite on a Linux or other *nix type system. The package further includes manuals and sample images and scripts to demonstrate use of the software for new users. The results of the code are in close alignment with theoretical expectations and published results of experimental data. The methodology and the software package offered in AAPM TG162 can be used as baseline for characterization of inherent image quality attributes of planar imaging systems. © 2017 American Association of Physicists in Medicine.

  6. Balancing compliance and cost when implementing a Quality Assurance program

    International Nuclear Information System (INIS)

    Pickering, S.Y.

    1997-12-01

    When implementing a Quality Assurance (QA) program, compliance and cost must be balanced. A QA program must be developed that hits the mark in terms of adequate control and documentation, but does not unnecessarily expand resources. As the Waste Isolation Pilot Plant (WIPP) has moved towards certification, Sandia National Laboratories has learned much about balancing compliance and costs. Some of these lessons are summarized here

  7. The six critical attributes of the next generation of quality management software systems.

    Science.gov (United States)

    Clark, Kathleen

    2011-07-01

    Driven by both the need to meet regulatory requirements and a genuine desire to drive improved quality, quality management systems encompassing standard operating procedure, corrective and preventative actions and related processes have existed for many years, both in paper and electronic form. The impact of quality management systems on 'actual' quality, however, is often reported as far less than desired. A quality management software system that moves beyond formal forms-driven processes to include a true closed loop design, manage disparate processes across the enterprise, provide support for collaborative processes and deliver insight into the overall state of control has the potential to close the gap between simply accomplishing regulatory compliance and delivering measurable improvements in quality and efficiency.

  8. Regulatory inspection of the implementation of quality assurance programmes

    International Nuclear Information System (INIS)

    1989-01-01

    This Manual provides guidance to Member States in the organization and performance of their regulatory inspection functions regarding the implementation of nuclear power plant quality assurance programmes. It addresses the interface between, and is consistent with, the IAEA Nuclear Safety Standards (NUSS programme) documents on quality assurance and governmental organization. The Manual offers a practical model and examples for performing regulatory inspections to ensure that the quality assurance programme is operating satisfactorily in the siting, design, manufacturing, construction, commissioning, operation and decommissioning of nuclear power plants. The primary objective is to confirm that the licensee has the capability to manage and control the effective performance of all quality assurance responsibilities during all phases of a nuclear power project. The guidance provided through this Manual for proper establishment and execution of the regulatory inspections helps to enforce the effective implementation of the quality assurance programme as a management control system that the nuclear industry should establish and use in attaining the safety and reliability objectives for nuclear installations. This enforcement action by national regulatory bodies and the emphasis on the purposes and advantages of quality assurance as an important management tool integrated within the total project task have been recommended by the IAEA International Nuclear Safety Advisory Group (INSAG). The primary intended users of this Manual are the management personnel and high level staff from regulatory bodies but it will also be helpful to management personnel from nuclear utilities and vendors. They all are inevitable partners in a nuclear power project and this document offers all of them valuable information on the better accomplishment of quality assurance activities to ensure the common objective of safe and reliable nuclear power production

  9. Implementing hospital quality assurance policies in Iran: balancing licensing, annual evaluation, inspections and quality management systems.

    Science.gov (United States)

    Aghaei Hashjin, Asgar; Delgoshaei, Bahram; Kringos, Dionne S; Tabibi, Seyed Jamaladin; Manouchehri, Jila; Klazinga, Niek S

    2015-01-01

    The purpose of this paper is to provide an overview of applied hospital quality assurance (QA) policies in Iran. A mixed method (quantitative data and qualitative document analysis) study was carried out between 1996 and 2010. The QA policy cycle forms a tight monitoring system to assure hospital quality by combining mandatory and voluntary methods in Iran. The licensing, annual evaluation and grading, and regulatory inspections statutorily implemented by the government as a national package to assure and improve hospital care quality, while implementing quality management systems (QMS) was voluntary for hospitals. The government's strong QA policy legislation role and support has been an important factor for successful QA implementation in Iran, though it may affected QA assessment independency and validity. Increased hospital evaluation independency and repositioning, updating standards, professional involvement and effectiveness studies could increase QA policy impact and maturity. The study highlights the current QA policy implementation cycle in Iranian hospitals. It provides a basis for further quality strategy development in Iranian hospitals and elsewhere. It also raises attention about finding the optimal balance between different QA policies, which is topical for many countries. This paper describes experiences when implementing a unique approach, combining mandatory and voluntary QA policies simultaneously in a developing country, which has invested considerably over time to improve hospital quality. The experiences with a mixed obligatory/voluntary approach and comprehensive policies in Iran may contain lessons for policy makers in developing and developed countries.

  10. Quality Assurance Framework Implementation Guide for Isolated Community Power Systems

    Energy Technology Data Exchange (ETDEWEB)

    Esterly, Sean R. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Baring-Gould, Edward I. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Burman, Kari A. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Greacen, Chris [Independent Consultant (United States)

    2017-08-15

    This implementation guide is a companion document to the 'Quality Assurance Framework for Mini-Grids' technical report. This document is intended to be used by one of the many stakeholder groups that take part in the implementation of isolated power systems. Although the QAF could be applied to a single system, it was designed primarily to be used within the context of a larger national or regional rural electrification program in which many individual systems are being installed. This guide includes a detailed overview of the Quality Assurance Framework and provides guidance focused on the implementation of the Framework from the perspective of the different stakeholders that are commonly involved in expanding energy development within specific communities or regions. For the successful long-term implementation of a specific rural electrification program using mini-grid systems, six key stakeholders have been identified that are typically engaged, each with a different set of priorities 1. Regulatory agency 2. Governmental ministry 3. System developers 4. Mini-utility 5. Investors 6. Customers/consumers. This document is broken into two distinct sections. The first focuses on the administrative processes in the development and operation of community-based mini-grid programs, while the second focuses on the process around the installation of the mini-grid project itself.

  11. WE-D-9A-06: Open Source Monitor Calibration and Quality Control Software for Enterprise Display Management

    Energy Technology Data Exchange (ETDEWEB)

    Bevins, N; Vanderhoek, M; Lang, S; Flynn, M [Henry Ford Health System, Detroit, MI (United States)

    2014-06-15

    Purpose: Medical display monitor calibration and quality control present challenges to medical physicists. The purpose of this work is to demonstrate and share experiences with an open source package that allows for both initial monitor setup and routine performance evaluation. Methods: A software package, pacsDisplay, has been developed over the last decade to aid in the calibration of all monitors within the radiology group in our health system. The software is used to calibrate monitors to follow the DICOM Grayscale Standard Display Function (GSDF) via lookup tables installed on the workstation. Additional functionality facilitates periodic evaluations of both primary and secondary medical monitors to ensure satisfactory performance. This software is installed on all radiology workstations, and can also be run as a stand-alone tool from a USB disk. Recently, a database has been developed to store and centralize the monitor performance data and to provide long-term trends for compliance with internal standards and various accrediting organizations. Results: Implementation and utilization of pacsDisplay has resulted in improved monitor performance across the health system. Monitor testing is now performed at regular intervals and the software is being used across multiple imaging modalities. Monitor performance characteristics such as maximum and minimum luminance, ambient luminance and illuminance, color tracking, and GSDF conformity are loaded into a centralized database for system performance comparisons. Compliance reports for organizations such as MQSA, ACR, and TJC are generated automatically and stored in the same database. Conclusion: An open source software solution has simplified and improved the standardization of displays within our health system. This work serves as an example method for calibrating and testing monitors within an enterprise health system.

  12. WE-D-9A-06: Open Source Monitor Calibration and Quality Control Software for Enterprise Display Management

    International Nuclear Information System (INIS)

    Bevins, N; Vanderhoek, M; Lang, S; Flynn, M

    2014-01-01

    Purpose: Medical display monitor calibration and quality control present challenges to medical physicists. The purpose of this work is to demonstrate and share experiences with an open source package that allows for both initial monitor setup and routine performance evaluation. Methods: A software package, pacsDisplay, has been developed over the last decade to aid in the calibration of all monitors within the radiology group in our health system. The software is used to calibrate monitors to follow the DICOM Grayscale Standard Display Function (GSDF) via lookup tables installed on the workstation. Additional functionality facilitates periodic evaluations of both primary and secondary medical monitors to ensure satisfactory performance. This software is installed on all radiology workstations, and can also be run as a stand-alone tool from a USB disk. Recently, a database has been developed to store and centralize the monitor performance data and to provide long-term trends for compliance with internal standards and various accrediting organizations. Results: Implementation and utilization of pacsDisplay has resulted in improved monitor performance across the health system. Monitor testing is now performed at regular intervals and the software is being used across multiple imaging modalities. Monitor performance characteristics such as maximum and minimum luminance, ambient luminance and illuminance, color tracking, and GSDF conformity are loaded into a centralized database for system performance comparisons. Compliance reports for organizations such as MQSA, ACR, and TJC are generated automatically and stored in the same database. Conclusion: An open source software solution has simplified and improved the standardization of displays within our health system. This work serves as an example method for calibrating and testing monitors within an enterprise health system

  13. Quality control of the software in the JT-60 computer control system

    International Nuclear Information System (INIS)

    Isaji, Nobuaki; Kurihara, Kenichi; Kimura, Toyoaki

    1990-07-01

    The JT-60 Control System should be improved corresponding to the experimental requirements. In order to keep the integrity of the system even in the modification the concept of quality control (QC) was introduced in the software development. What we have done for QC activity are (1) to establish standard procedures of the software development, (2) to develop support tools for grasping the present status of the program structure, and (3) to develop a document system, and a source program management system. This paper reports these QC activities and their problems for the JT-60 control system. (author)

  14. A proposed defect tracking model for classifying the inserted defect reports to enhance software quality control.

    Science.gov (United States)

    Sultan, Torky; Khedr, Ayman E; Sayed, Mostafa

    2013-01-01

    NONE DECLARED Defect tracking systems play an important role in the software development organizations as they can store historical information about defects. There are many research in defect tracking models and systems to enhance their capabilities to be more specifically tracking, and were adopted with new technology. Furthermore, there are different studies in classifying bugs in a step by step method to have clear perception and applicable method in detecting such bugs. This paper shows a new proposed defect tracking model for the purpose of classifying the inserted defects reports in a step by step method for more enhancement of the software quality.

  15. Efficient FPGA Implementation of a STBC-OFDM Combiner for an IEEE 802.16 Software Radio Receiver

    DEFF Research Database (Denmark)

    Cattoni, Andrea Fabio; Le Moullec, Yannick; Sacchi, Claudio

    2014-01-01

    In this paper, an efficient FPGA implementation of a 4x4 Space-Time Block Coding (STBC) combiner for MIMO-OFDM software radio receivers is considered. The proposed combiner is based on a low-complexity algorithm which reduces the interference due to the Quasi-Orthogonality of the STBC decoding...

  16. Investigating the Practical Impact of Agile Practices on the Quality of Software Projects in Continuous Delivery

    OpenAIRE

    Olumide Akerele; Muthu Ramachandran; Mark Dixon

    2014-01-01

    Various factors affect the impact of agile factors on the continuous delivery of software projects. This is a major reason why projects perform differently- some failing and some succeeding- when they implement some agile practices in various environments. This is not helped by the fact that many projects work within limited budget while project plans also change-- making them to fall into some sort of pressure to meet deadline when they fall behind in their planned work. This study investiga...

  17. Quality control in urodynamics and the role of software support in the QC procedure.

    Science.gov (United States)

    Hogan, S; Jarvis, P; Gammie, A; Abrams, P

    2011-11-01

    This article aims to identify quality control (QC) best practice, to review published QC audits in order to identify how closely good practice is followed, and to carry out a market survey of the software features that support QC offered by urodynamics machines available in the UK. All UK distributors of urodynamic systems were contacted and asked to provide information on the software features relating to data quality of the products they supply. The results of the market survey show that the features offered by manufacturers differ greatly. Automated features, which can be turned off in most cases, include: cough recognition, detrusor contraction detection, and high pressure alerts. There are currently no systems that assess data quality based on published guidelines. A literature review of current QC guidelines for urodynamics was carried out; QC audits were included in the literature review to see how closely guidelines were being followed. This review highlights the fact that basic QC is not being carried out effectively by urodynamicists. Based on the software features currently available and the results of the literature review there is both the need and capacity for a greater degree of automation in relation to urodynamic data quality and accuracy assessment. Some progress has been made in this area and certain manufacturers have already developed automated cough detection. Copyright © 2011 Wiley Periodicals, Inc.

  18. QCScreen: a software tool for data quality control in LC-HRMS based metabolomics.

    Science.gov (United States)

    Simader, Alexandra Maria; Kluger, Bernhard; Neumann, Nora Katharina Nicole; Bueschl, Christoph; Lemmens, Marc; Lirk, Gerald; Krska, Rudolf; Schuhmacher, Rainer

    2015-10-24

    Metabolomics experiments often comprise large numbers of biological samples resulting in huge amounts of data. This data needs to be inspected for plausibility before data evaluation to detect putative sources of error e.g. retention time or mass accuracy shifts. Especially in liquid chromatography-high resolution mass spectrometry (LC-HRMS) based metabolomics research, proper quality control checks (e.g. for precision, signal drifts or offsets) are crucial prerequisites to achieve reliable and comparable results within and across experimental measurement sequences. Software tools can support this process. The software tool QCScreen was developed to offer a quick and easy data quality check of LC-HRMS derived data. It allows a flexible investigation and comparison of basic quality-related parameters within user-defined target features and the possibility to automatically evaluate multiple sample types within or across different measurement sequences in a short time. It offers a user-friendly interface that allows an easy selection of processing steps and parameter settings. The generated results include a coloured overview plot of data quality across all analysed samples and targets and, in addition, detailed illustrations of the stability and precision of the chromatographic separation, the mass accuracy and the detector sensitivity. The use of QCScreen is demonstrated with experimental data from metabolomics experiments using selected standard compounds in pure solvent. The application of the software identified problematic features, samples and analytical parameters and suggested which data files or compounds required closer manual inspection. QCScreen is an open source software tool which provides a useful basis for assessing the suitability of LC-HRMS data prior to time consuming, detailed data processing and subsequent statistical analysis. It accepts the generic mzXML format and thus can be used with many different LC-HRMS platforms to process both multiple

  19. European Interoperability Assets Register and Quality Framework Implementation.

    Science.gov (United States)

    Moreno-Conde, Alberto; Thienpont, Geert; Lamote, Inge; Coorevits, Pascal; Parra, Carlos; Kalra, Dipak

    2016-01-01

    Interoperability assets is the term applied to refer to any resource that can support the design, implementation and successful adoption of eHealth services that can exchange data meaningfully. Some examples may include functional requirements, specifications, standards, clinical models and term lists, guidance on how standards may be used concurrently, implementation guides, educational resources, and other resources. Unfortunately, these are largely accessible in ad hoc ways and result in scattered fragments of a solution space that urgently need to be brought together. At present, it is well known that new initiatives and projects will reinvent assets of which they were unaware, while those assets which were potentially of great value are forgotten, not maintained and eventually fall into disuse. This research has defined a quality in use model and assessed the suitability of this quality framework based on the feedback and opinion of a representative sample of potential end users. This quality framework covers the following domains of asset development and adoption: (i) Development process, (ii) Maturity level, (iii) Trustworthiness, (iv) Support & skills, (v) Sustainability, (vi) Semantic interoperability, (vii) Cost & effort of adoption (viii) Maintenance. When participants were requested to evaluate how the overall quality in use framework, 70% would recommend using the register to their colleagues, 70% felt that it could provide relevant benefits for discovering new assets, and 50% responded that it would support their decision making about the recommended asset to adopt or implement in their organisation. Several European projects have expressed interest in using the register, which will now be sustained and promoted by the the European Institute for Innovation through Health Data.

  20. Process-based quality management for clinical implementation of adaptive radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Noel, Camille E.; Santanam, Lakshmi; Parikh, Parag J.; Mutic, Sasa, E-mail: smutic@radonc.wustl.edu [Department of Radiation Oncology, Washington University School of Medicine, St. Louis, Missouri 63110 (United States)

    2014-08-15

    Purpose: Intensity-modulated adaptive radiotherapy (ART) has been the focus of considerable research and developmental work due to its potential therapeutic benefits. However, in light of its unique quality assurance (QA) challenges, no one has described a robust framework for its clinical implementation. In fact, recent position papers by ASTRO and AAPM have firmly endorsed pretreatment patient-specific IMRT QA, which limits the feasibility of online ART. The authors aim to address these obstacles by applying failure mode and effects analysis (FMEA) to identify high-priority errors and appropriate risk-mitigation strategies for clinical implementation of intensity-modulated ART. Methods: An experienced team of two clinical medical physicists, one clinical engineer, and one radiation oncologist was assembled to perform a standard FMEA for intensity-modulated ART. A set of 216 potential radiotherapy failures composed by the forthcoming AAPM task group 100 (TG-100) was used as the basis. Of the 216 failures, 127 were identified as most relevant to an ART scheme. Using the associated TG-100 FMEA values as a baseline, the team considered how the likeliness of occurrence (O), outcome severity (S), and likeliness of failure being undetected (D) would change for ART. New risk priority numbers (RPN) were calculated. Failures characterized by RPN ≥ 200 were identified as potentially critical. Results: FMEA revealed that ART RPN increased for 38% (n = 48/127) of potential failures, with 75% (n = 36/48) attributed to failures in the segmentation and treatment planning processes. Forty-three of 127 failures were identified as potentially critical. Risk-mitigation strategies include implementing a suite of quality control and decision support software, specialty QA software/hardware tools, and an increase in specially trained personnel. Conclusions: Results of the FMEA-based risk assessment demonstrate that intensity-modulated ART introduces different (but not necessarily