WorldWideScience

Sample records for implementing software quality

  1. Program for implementing software quality metrics

    Energy Technology Data Exchange (ETDEWEB)

    Yule, H.P.; Riemer, C.A.

    1992-04-01

    This report describes a program by which the Veterans Benefit Administration (VBA) can implement metrics to measure the performance of automated data systems and demonstrate that they are improving over time. It provides a definition of quality, particularly with regard to software. Requirements for management and staff to achieve a successful metrics program are discussed. It lists the attributes of high-quality software, then describes the metrics or calculations that can be used to measure these attributes in a particular system. Case studies of some successful metrics programs used by business are presented. The report ends with suggestions on which metrics the VBA should use and the order in which they should be implemented.

  2. Implementing Quality Assurance Features in Component-based Software System

    National Research Council Canada - National Science Library

    Navdeep Batolar; Parminder Kaur

    2016-01-01

    The increasing demand of component-based development approach (CBDA) gives opportunity to the software developers to increase the speed of the software development process and lower its production cost...

  3. An Approach for the Implementation of Software Quality Models Adpoting CERTICS and CMMI-DEV

    Directory of Open Access Journals (Sweden)

    GARCIA, F.W.

    2015-12-01

    Full Text Available This paper proposes a mapping between two product quality and software processes models used in the industry, the CERTICS national model and the CMMI-DEV international model. The stages of mapping are presented step by step, as well as the mapping review, which had the cooperation of one specialist in CERTICS and CMMI-DEV models. It aims to correlate the structures of the two models in order to facilitate and reduce the implementation time and costs, and to stimulate the execution of multi-model implementations in software developers companies.

  4. A General Approach of Quality Cost Management Suitable for Effective Implementation in Software Systems

    Directory of Open Access Journals (Sweden)

    Stelian BRAD

    2010-01-01

    Full Text Available Investments in quality are best quantified by implementing and managing quality cost systems. A review of various opinions coming from practitioners and researchers about the existent quality cost models reveals a set of drawbacks (e.g. too theoretical and too close to ideal cases; too academic, with less practical impact; too much personalized to particular business processes, with difficulties in extrapolating to other cases; not comprising all dimensions of a business system. Using concepts and tools in quality management theory and practice and algorithms of innovative problem solving, this paper formulates a novel approach to improve practical usability, comprehensiveness, flexibility and customizability of a quality cost management system (QCMS when implementing it in a specific software application. Conclusions arising from the implementation in real industrial cases are also highlighted.

  5. Software quality assurance handbook

    Energy Technology Data Exchange (ETDEWEB)

    1990-09-01

    There are two important reasons for Software Quality Assurance (SQA) at Allied-Signal Inc., Kansas City Division (KCD): First, the benefits from SQA make good business sense. Second, the Department of Energy has requested SQA. This handbook is one of the first steps in a plant-wide implementation of Software Quality Assurance at KCD. The handbook has two main purposes. The first is to provide information that you will need to perform software quality assurance activities. The second is to provide a common thread to unify the approach to SQA at KCD. 2 figs.

  6. SU-E-T-103: Development and Implementation of Web Based Quality Control Software

    Energy Technology Data Exchange (ETDEWEB)

    Studinski, R; Taylor, R; Angers, C; La Russa, D; Clark, B [The Ottawa Hospital Regional Cancer Ctr., Ottawa, ON (Canada)

    2014-06-01

    Purpose: Historically many radiation medicine programs have maintained their Quality Control (QC) test results in paper records or Microsoft Excel worksheets. Both these approaches represent significant logistical challenges, and are not predisposed to data review and approval. It has been our group's aim to develop and implement web based software designed not just to record and store QC data in a centralized database, but to provide scheduling and data review tools to help manage a radiation therapy clinics Equipment Quality control program. Methods: The software was written in the Python programming language using the Django web framework. In order to promote collaboration and validation from other centres the code was made open source and is freely available to the public via an online source code repository. The code was written to provide a common user interface for data entry, formalize the review and approval process, and offer automated data trending and process control analysis of test results. Results: As of February 2014, our installation of QAtrack+ has 180 tests defined in its database and has collected ∼22 000 test results, all of which have been reviewed and approved by a physicist via QATrack+'s review tools. These results include records for quality control of Elekta accelerators, CT simulators, our brachytherapy programme, TomoTherapy and Cyberknife units. Currently at least 5 other centres are known to be running QAtrack+ clinically, forming the start of an international user community. Conclusion: QAtrack+ has proven to be an effective tool for collecting radiation therapy QC data, allowing for rapid review and trending of data for a wide variety of treatment units. As free and open source software, all source code, documentation and a bug tracker are available to the public at https://bitbucket.org/tohccmedphys/qatrackplus/.

  7. Software Quality Assurance in Software Projects: A Study of Pakistan

    Directory of Open Access Journals (Sweden)

    Faisal Shafique Butt

    2013-05-01

    Full Text Available Software quality is specific property which tells what kind of standard software should have. In a software project, quality is the key factor of success and decline of software related organization. Many researches have been done regarding software quality. Software related organization follows standards introduced by Capability Maturity Model Integration (CMMI to achieve good quality software. Quality is divided into three main layers which are Software Quality Assurance (SQA, Software Quality Plan (SQP and Software Quality Control (SQC. So In this study, we are discussing the quality standards and principles of software projects in Pakistan software Industry and how these implemented quality standards are measured and managed. In this study, we will see how many software firms are following the rules of CMMI to create software. How many are reaching international standards and how many firms are measuring the quality of their projects. The results show some of the companies are using software quality assurance techniques in Pakstan.

  8. Implementing Software Defined Radio

    CERN Document Server

    Grayver, Eugene

    2013-01-01

    Software Defined Radio makes wireless communications easier, more efficient, and more reliable. This book bridges the gap between academic research and practical implementation. When beginning a project, practicing engineers, technical managers, and graduate students can save countless hours by considering the concepts presented in these pages. The author covers the myriad options and trade-offs available when selecting an appropriate hardware architecture. As demonstrated here, the choice between hardware- and software-centric architecture can mean the difference between meeting an aggressive schedule and bogging down in endless design iterations. Because of the author’s experience overseeing dozens of failed and successful developments, he is able to present many real-life examples. Some of the key concepts covered are: Choosing the right architecture for the market – laboratory, military, or commercial Hardware platforms – FPGAs, GPPs, specialized and hybrid devices Standardization efforts to ens...

  9. Software quality in 1997

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Inc., Burlington, MA (United States)

    1997-11-01

    For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Deployment (QFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels. Since software is on the critical path for many engineered products, and for internal business systems as well, the new approaches are starting to affect global competition and attract widespread international interest. It can be hypothesized that success in mastering software quality will be a key strategy for dominating global software markets in the 21st century.

  10. Sandia software guidelines: Software quality planning

    Energy Technology Data Exchange (ETDEWEB)

    1987-08-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standard for Software Quality Assurance Plans, this volume identifies procedures to follow in producing a Software Quality Assurance Plan for an organization or a project, and provides an example project SQA plan. 2 figs., 4 tabs.

  11. Software product quality control

    CERN Document Server

    Wagner, Stefan

    2013-01-01

    Quality is not a fixed or universal property of software; it depends on the context and goals of its stakeholders. Hence, when you want to develop a high-quality software system, the first step must be a clear and precise specification of quality. Yet even if you get it right and complete, you can be sure that it will become invalid over time. So the only solution is continuous quality control: the steady and explicit evaluation of a product's properties with respect to its updated quality goals.This book guides you in setting up and running continuous quality control in your environment. Star

  12. Sandia National Laboratories Advanced Simulation and Computing (ASC) : appraisal method for the implementation of the ASC software quality engineering practices: Version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Turgeon, Jennifer; Minana, Molly A.

    2008-02-01

    This document provides a guide to the process of conducting software appraisals under the Sandia National Laboratories (SNL) ASC Program. The goal of this document is to describe a common methodology for planning, conducting, and reporting results of software appraisals thereby enabling: development of an objective baseline on implementation of the software quality engineering (SQE) practices identified in the ASC Software Quality Plan across the ASC Program; feedback from project teams on SQE opportunities for improvement; identification of strengths and opportunities for improvement for individual project teams; guidance to the ASC Program on the focus of future SQE activities Document contents include process descriptions, templates to promote consistent conduct of appraisals, and an explanation of the relationship of this procedure to the SNL ASC software program.

  13. ERP Software Implementation Best Practices.

    Science.gov (United States)

    Frantz, Pollyanne S.; Southerland, Arthur R.; Johnson, James T.

    2002-01-01

    Studied the perceptions of chief financial and information officers of enterprise resource planning (ERP) software implementation best practices. Usable responses from 159 respondents show consensus for the most part between the perceptions of the two groups and describe some best practices that represent common ground. (SLD)

  14. Modernization of software quality assurance

    Science.gov (United States)

    Bhaumik, Gokul

    1988-01-01

    The customers satisfaction depends not only on functional performance, it also depends on the quality characteristics of the software products. An examination of this quality aspect of software products will provide a clear, well defined framework for quality assurance functions, which improve the life-cycle activities of software development. Software developers must be aware of the following aspects which have been expressed by many quality experts: quality cannot be added on; the level of quality built into a program is a function of the quality attributes employed during the development process; and finally, quality must be managed. These concepts have guided our development of the following definition for a Software Quality Assurance function: Software Quality Assurance is a formal, planned approach of actions designed to evaluate the degree of an identifiable set of quality attributes present in all software systems and their products. This paper is an explanation of how this definition was developed and how it is used.

  15. Roadmap for Peridynamic Software Implementation

    Energy Technology Data Exchange (ETDEWEB)

    Littlewood, David John [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-10-01

    The application of peridynamics for engineering analysis requires an efficient and robust software implementation. Key elements include processing of the discretization, the proximity search for identification of pairwise interactions, evaluation of the con- stitutive model, application of a bond-damage law, and contact modeling. Additional requirements may arise from the choice of time integration scheme, for example esti- mation of the maximum stable time step for explicit schemes, and construction of the tangent stiffness matrix for many implicit approaches. This report summaries progress to date on the software implementation of the peridynamic theory of solid mechanics. Discussion is focused on parallel implementation of the meshfree discretization scheme of Silling and Askari [33] in three dimensions, although much of the discussion applies to computational peridynamics in general.

  16. Software Metrics to Estimate Software Quality using Software Component Reusability

    Directory of Open Access Journals (Sweden)

    Prakriti Trivedi

    2012-03-01

    Full Text Available Today most of the applications developed using some existing libraries, codes, open sources etc. As a code is accessed in a program, it is represented as the software component. Such as in java beans and in .net ActiveX controls are the software components. These components are ready to use programming code or controls that excel the code development. A component based software system defines the concept of software reusability. While using these components the main question arise is whether to use such components is beneficial or not. In this proposed work we are trying to present the answer for the same question. In this work we are presenting a set of software matrix that will check the interconnection between the software component and the application. How strong this relation defines the software quality after using this software component. The overall metrics will return the final result in terms of the boundless of the component with application.

  17. Research and implementation of software automatic test

    Science.gov (United States)

    Li-hong, LIAN

    2017-06-01

    With the fast development in IT technology nowadays, software is increasingly complex and large. Hundreds of people in the development team, thousands of modules and interfaces, across geographies and systems user are no longer a fresh thing. All of these put forward higher requirements for software testing. Due to the low cost of implementation and the advantage of effective inheritance and accumulation of test assets, software automation testing has gradually become one of the important means to ensure the quality of software for IT enterprises. This paper analyzes the advantages of automatic test, common misconceptions; puts forward unsuitable application scenarios and the best time to intervene; focus on the analysis of the feasibility of judging the interface automation test; and puts forward the function and elements of interface automatic test tools to have; provides a reference for large-scale project interface automated testing tool selection or custom development.

  18. Quality Attribute Driven Software Architecture Reconstruction

    Science.gov (United States)

    2016-06-21

    3890 Quality Attribute Driven Software Architecture Reconstruction SATURN Workshop April 7, 2005 Liam O’Brien © 2005 by Carnegie Mellon University...Version 1.0 QADSAR – SATURN 2005 - page 2 Motivation - 1 • Software architectures are critical to implement an organization’s business goals and critical...Attribute Driven Software Architecture Reconstruction 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER

  19. Building quality into medical product software design.

    Science.gov (United States)

    Mallory, S R

    1993-01-01

    The software engineering and quality assurance disciplines are a requisite to the design of safe and effective software-based medical devices. It is in the areas of software methodology and process that the most beneficial application of these disciplines to software development can be made. Software is a product of complex operations and methodologies and is not amenable to the traditional electromechanical quality assurance processes. Software quality must be built in by the developers, with the software verification and validation engineers acting as the independent instruments for ensuring compliance with performance objectives and with development and maintenance standards. The implementation of a software quality assurance program is a complex process involving management support, organizational changes, and new skill sets, but the benefits are profound. Its rewards provide safe, reliable, cost-effective, maintainable, and manageable software, which may significantly speed the regulatory review process and therefore potentially shorten the overall time to market. The use of a trial project can greatly facilitate the learning process associated with the first-time application of a software quality assurance program.

  20. Software Quality Assurance Audits Guidebooks

    Science.gov (United States)

    1990-01-01

    The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes that are used in software development. The Software Assurance Guidebook, NASA-GB-A201, issued in September, 1989, provides an overall picture of the NASA concepts and practices in software assurance. Second level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the second level Software Quality Assurance Audits Guidebook that describes software quality assurance audits in a way that is compatible with practices at NASA Centers.

  1. Software Quality Assurance for Nuclear Safety Systems

    Energy Technology Data Exchange (ETDEWEB)

    Sparkman, D R; Lagdon, R

    2004-05-16

    The US Department of Energy has undertaken an initiative to improve the quality of software used to design and operate their nuclear facilities across the United States. One aspect of this initiative is to revise or create new directives and guides associated with quality practices for the safety software in its nuclear facilities. Safety software includes the safety structures, systems, and components software and firmware, support software and design and analysis software used to ensure the safety of the facility. DOE nuclear facilities are unique when compared to commercial nuclear or other industrial activities in terms of the types and quantities of hazards that must be controlled to protect workers, public and the environment. Because of these differences, DOE must develop an approach to software quality assurance that ensures appropriate risk mitigation by developing a framework of requirements that accomplishes the following goals: {sm_bullet} Ensures the software processes developed to address nuclear safety in design, operation, construction and maintenance of its facilities are safe {sm_bullet} Considers the larger system that uses the software and its impacts {sm_bullet} Ensures that the software failures do not create unsafe conditions Software designers for nuclear systems and processes must reduce risks in software applications by incorporating processes that recognize, detect, and mitigate software failure in safety related systems. It must also ensure that fail safe modes and component testing are incorporated into software design. For nuclear facilities, the consideration of risk is not necessarily sufficient to ensure safety. Systematic evaluation, independent verification and system safety analysis must be considered for software design, implementation, and operation. The software industry primarily uses risk analysis to determine the appropriate level of rigor applied to software practices. This risk-based approach distinguishes safety

  2. Software Quality Metrics

    Science.gov (United States)

    1991-07-01

    March 1979, pp. 121-128. Gorla, Narasimhaiah, Alan C. Benander, and Barbara A. Benander, "Debugging Effort Estimation Using Software Metrics", IEEE...Society, IEEE Guide for the Use of IEEE Standard Dictionary of Measures to Produce Reliable Software, IEEE Std 982.2-1988, June 1989. Jones, Capers

  3. First experiences with the implementation of the European standard EN 62304 on medical device software for the quality assurance of a radiotherapy unit.

    Science.gov (United States)

    Höss, Angelika; Lampe, Christian; Panse, Ralf; Ackermann, Benjamin; Naumann, Jakob; Jäkel, Oliver

    2014-03-21

    According to the latest amendment of the Medical Device Directive standalone software qualifies as a medical device when intended by the manufacturer to be used for medical purposes. In this context, the EN 62304 standard is applicable which defines the life-cycle requirements for the development and maintenance of medical device software. A pilot project was launched to acquire skills in implementing this standard in a hospital-based environment (in-house manufacture). The EN 62304 standard outlines minimum requirements for each stage of the software life-cycle, defines the activities and tasks to be performed and scales documentation and testing according to its criticality. The required processes were established for the pre-existent decision-support software FlashDumpComparator (FDC) used during the quality assurance of treatment-relevant beam parameters. As the EN 62304 standard implicates compliance with the EN ISO 14971 standard on the application of risk management to medical devices, a risk analysis was carried out to identify potential hazards and reduce the associated risks to acceptable levels. The EN 62304 standard is difficult to implement without proper tools, thus open-source software was selected and integrated into a dedicated development platform. The control measures yielded by the risk analysis were independently implemented and verified, and a script-based test automation was retrofitted to reduce the associated test effort. After all documents facilitating the traceability of the specified requirements to the corresponding tests and of the control measures to the proof of execution were generated, the FDC was released as an accessory to the HIT facility. The implementation of the EN 62304 standard was time-consuming, and a learning curve had to be overcome during the first iterations of the associated processes, but many process descriptions and all software tools can be re-utilized in follow-up projects. It has been demonstrated that a

  4. Software cost/resource modeling: Software quality tradeoff measurement

    Science.gov (United States)

    Lawler, R. W.

    1980-01-01

    A conceptual framework for treating software quality from a total system perspective is developed. Examples are given to show how system quality objectives may be allocated to hardware and software; to illustrate trades among quality factors, both hardware and software, to achieve system performance objectives; and to illustrate the impact of certain design choices on software functionality.

  5. DESQA a Software Quality Assurance Framework

    Directory of Open Access Journals (Sweden)

    Dr.KhaledKh. S. Kh. Allanqawi

    2017-04-01

    Full Text Available In current software development lifecycles of heterogeneous environments, the pitfalls businesses have to face are that software defect tracking, measurements and quality assurance do not start early enough in the development process. In fact the cost of fixing a defect in a production environment is much higher than in the initial phases of the Software Development Life Cycle (SDLC which is particularly true for Service Oriented Architecture (SOA. Thus the aim of this study is to develop a new framework for defect tracking and detection and quality estimation for early stages particularly for the design stage of the SDLC. Part of the objectives of this work is to conceptualize, borrow and customize from known frameworks, such as object-oriented programming to build a solid framework using automated rule based intelligent mechanisms to detect and classify defects in software design of SOA. The implementation part demonstrated how the framework can predict the quality level of the designed software. The results showed a good level of quality estimation can be achieved based on the number of design attributes, the number of quality attributes and the number of SOA Design Defects. Assessment shows that metrics provide guidelines to indicate the progress that a software system has made and the quality of design. Using these guidelines, we can develop more usable and maintainable software systems to fulfill the demand of efficient systems for software applications. Another valuable result coming from this study is that developers are trying to keep backwards compatibility when they introduce new functionality. Sometimes, in the same newly-introduced elements developers perform necessary breaking changes in future versions. In that way they give time to their clients to adapt their systems. This is a very valuable practice for the developers because they have more time to assess the quality of their software before releasing it. Other improvements in

  6. Software quality engineering a practitioner's approach

    CERN Document Server

    Suryn, Witold

    2014-01-01

    Software quality stems from two distinctive, but associated, topics in software engineering: software functional quality and software structural quality. Software Quality Engineering studies the tenets of both of these notions, which focus on the efficiency and value of a design, respectively. The text addresses engineering quality on both the application and system levels with attention to Information Systems and Embedded Systems as well as recent developments. Targeted at graduate engineering students and software quality specialists, the book analyzes the relationship between functionality

  7. Continuous improvement of software quality

    Energy Technology Data Exchange (ETDEWEB)

    Sivertsen, Terje

    1999-04-15

    The present report is the first Halden Work Report delivered from the OECD Halden Reactor Project's research activity on formal methods and software quality. Of particular concern in this activity is to reach a consensus between regulators, licensees and the nuclear industry on questions related to the effective, industrial use of formal methods. The report gives considerable attention to the importance of continuous improvement as a characteristic of a living software quality system, and to the need of providing a basis for software process/product quality integration. In particular, the report discusses these aspects from the perspectives of defect prevention, formal methods, Total Quality Management (TQM), and Bayesian Belief Nets. Another concern is to promote controlled experiments on the use of new methods, techniques, and tools. This is achieved partly by reviewing suggestions on the collection and experimental use of data, and by surveying a number of metrics believed to have some potential for comparison studies (author) (ml)

  8. Set up and programming of an ALICE Time-Of-Flight trigger facility and software implementation for its Quality Assurance (QA) during LHC Run 2

    CERN Document Server

    Toschi, Francesco

    2016-01-01

    The Cosmic and Topology Trigger Module (CTTM) is the main component of a trigger based on the ALICE TOF detector. Taking advantage of the TOF fast response, this VME board implements the trigger logic and delivers several L0 trigger outputs, used since Run 1, to provide cosmic triggers and rare triggers in pp, p+Pb and Pb+Pb data taking. Due to TOF DCS architectural change of the PCs controlling the CTTM (from 32 bits to 64 bits) it is mandatory to upgrade the software related to the CTTM including the code programming the FPGA firmware. A dedicated CTTM board will be installed in a CERN lab (Meyrin site), with the aim of recreating the electronics chain of the TOF trigger, to get a comfortable porting of the code to the 64 bit environment. The project proposed to the summer student is the setting up of the CTTM and the porting of the software. Moreover, in order to monitor the CTTM Trigger board during the real data taking, the implementation of a new Quality Assurance (QA) code is also crucial, together wit...

  9. The Effects of Development Team Skill on Software Product Quality

    Science.gov (United States)

    Beaver, Justin M.; Schiavone, Guy A.

    2006-01-01

    This paper provides an analysis of the effect of the skill/experience of the software development team on the quality of the final software product. A method for the assessment of software development team skill and experience is proposed, and was derived from a workforce management tool currently in use by the National Aeronautics and Space Administration. Using data from 26 smallscale software development projects, the team skill measures are correlated to 5 software product quality metrics from the ISO/IEC 9126 Software Engineering Product Quality standard. in the analysis of the results, development team skill is found to be a significant factor in the adequacy of the design and implementation. In addition, the results imply that inexperienced software developers are tasked with responsibilities ill-suited to their skill level, and thus have a significant adverse effect on the quality of the software product. Keywords: software quality, development skill, software metrics

  10. Factors to Consider When Implementing Automated Software Testing

    Science.gov (United States)

    2016-11-10

    Factors to Consider When Implementing Automated Software Testing By Larry Yang, MBA, SSCP, Security+, Oracle DBA OCA, ASTQB CTFL, ITIL V3 ITM...Testing is a major component of the Software Development Lifecycle (SDLC), constituting a prominent cost driver for both government and industry entities...Therefore, many businesses are automating their software testing in order to save money and improve quality. When considering whether automation

  11. Software Quality Certification: identifying the real obstacles

    Directory of Open Access Journals (Sweden)

    Megan Baker

    1996-05-01

    Full Text Available A case study of software certification reveals the real difficulty of certifying quality beyond superficial assessment - readers are invited to form their own conclusions. AS 3563 Software Quality Management System is the Australian version of ISO 9001, developed specifically for the software industry. For many Australian software houses, gaining certification with AS 3563 is a priority since certification has become a prerequisite to doing business with government departments and major corporations. However, the process of achieving registration with this standard is a lengthy and resource intensive process, and may have little impact on actual software quality. This case study recounts the experience of the consulting arm of one of Australia's accounting firms in its quest for certification. By using a number of specific management strategies this company was able to successfully implement AS 3563 in less than half the time usually taken to achieve certification - a feat for which its management should be congratulated. However, because the focus of the project was on gaining certification, few internal benefits have been realised despite the successful implementation of the standard.

  12. Implementing software safety in the NASA environment

    Science.gov (United States)

    Wetherholt, Martha S.; Radley, Charles F.

    1994-05-01

    Until recently, NASA did not consider allowing computers total control of flight systems. Human operators, via hardware, have constituted the ultimate safety control. In an attempt to reduce costs, NASA has come to rely more and more heavily on computers and software to control space missions. (For example. software is now planned to control most of the operational functions of the International Space Station.) Thus the need for systematic software safety programs has become crucial for mission success. Concurrent engineering principles dictate that safety should be designed into software up front, not tested into the software after the fact. 'Cost of Quality' studies have statistics and metrics to prove the value of building quality and safety into the development cycle. Unfortunately, most software engineers are not familiar with designing for safety, and most safety engineers are not software experts. Software written to specifications which have not been safety analyzed is a major source of computer related accidents. Safer software is achieved step by step throughout the system and software life cycle. It is a process that includes requirements definition, hazard analyses, formal software inspections, safety analyses, testing, and maintenance. The greatest emphasis is placed on clearly and completely defining system and software requirements, including safety and reliability requirements. Unfortunately, development and review of requirements are the weakest link in the process. While some of the more academic methods, e.g. mathematical models, may help bring about safer software, this paper proposes the use of currently approved software methodologies, and sound software and assurance practices to show how, to a large degree, safety can be designed into software from the start. NASA's approach today is to first conduct a preliminary system hazard analysis (PHA) during the concept and planning phase of a project. This determines the overall hazard potential of

  13. Implementing an open-access CASA software for the assessment of stallion sperm motility: Relationship with other sperm quality parameters.

    Science.gov (United States)

    Giaretta, Elisa; Munerato, Mauro; Yeste, Marc; Galeati, Giovanna; Spinaci, Marcella; Tamanini, Carlo; Mari, Gaetano; Bucci, Diego

    2017-01-01

    Setting an open-access computer assisted sperm analysis (CASA) may benefit the evaluation of motility in mammalian sperm, especially when economic constraints do not allow the use of a commercial system. There have been successful attempts to develop such a device in Zebra fish sperm and the system has been used in very few studies on mammalian spermatozoa. Against this background, the present study aimed at developing an open-access CASA system for mammalian sperm using the horse as a model and based upon the Image J software previously established for Zebra fish sperm. Along with determining the sperm progressive motility and other kinetic parameters (such as amplitude of lateral head displacement), the "results" window was adjusted to simplify subsequent statistical analyses. The path window was enriched with colored sperm trajectories on the basis of the subpopulation they belong to and a number that allowed the sperm track to be associated to the sperm motility data shown in the "results" window. Data obtained from the novel plugin (named as CASA_bgm) were compared with those of the commercial CASA Hamilton-Thorn IVOS Vers.12, through Bland Altman's plots. While the percentage of total and progressive motile sperm, VCL, VAP, VSL, LIN and STR and ALH were in agreement with those obtained with the commercial system, BCF significantly differed between the two systems probably due to their settings. Interestingly, a positive and significant correlation between the percentages of total motile sperm evaluated through CASA_bgm and those showing high mitochondrial membrane potential evaluated by JC-1 staining was found. In conclusion, CASA_bgm ImageJ plugin could be useful and reliable for stallion sperm motility analysis and it is our aim to apply this system to other mammalian species. Copyright © 2016 Elsevier B.V. All rights reserved.

  14. A SELF PROCESS IMPROVEMENT FOR ACHIEVING HIGH SOFTWARE QUALITY

    Directory of Open Access Journals (Sweden)

    DR. SHASHANK.D.JOSHI

    2011-05-01

    Full Text Available Quality has been one of the most important factors in judging any product. Quality means “a degree or grade of excellence or worth”. Quality is a term that is usually described using adjectives. Quality has several attributes to it, some of which can be quantified using metrics. These attributes such as usability, portability, security, performance, reliability etc have different importance in different projects. Different software quality assurance methods & practices have been used in different software projects to attain the true value. Quality is an attribute which is a distinct feature and it differs with people’s perception. Achieving high software quality involves measurement of software metrics and optimization based on estimated values. As the software systems grow larger, complexity ofdesign and implementation increases, and this in turn is more prone to defects and hence directly affect the quality of the systems. However, in any software project, high quality is always desirable, and many projects have specific quality requirements. Achieving high software quality involves measurement of software metrics and optimization based on estimated values. Developing high quality software is governed by factors such as people, process, technology and cost. This paper attempts to present a novel approach towards achieving high software quality in various kinds of projects under given constraints.

  15. Software metrics: Software quality metrics for distributed systems. [reliability engineering

    Science.gov (United States)

    Post, J. V.

    1981-01-01

    Software quality metrics was extended to cover distributed computer systems. Emphasis is placed on studying embedded computer systems and on viewing them within a system life cycle. The hierarchy of quality factors, criteria, and metrics was maintained. New software quality factors were added, including survivability, expandability, and evolvability.

  16. Results of agile project management implementation in software engineering companies

    Directory of Open Access Journals (Sweden)

    Suetin Sergei

    2016-01-01

    Full Text Available Agile project management methodologies, tools and techniques have been becoming more and more popular among Russian and foreign software companies. Though the agile software project management methodologies appear to be more flexible and sound alternative to traditional project management approaches, the practice of agile project management implementation needs more research. The article presents the research on the practical results of agile project management implementation in Russian software engineering companies. The survey-based study covers 8 companies and 35 their projects managed with the help of agile methodologies. In contrast to some optimistic researchers of agile practices, the research findings show that in the investigated projects agile project management led to the deterioration of cost and schedule performance. However, the quality, both perceived by clients and assessed by internal technical analysts, improved after the implementation of agile software project management practices.

  17. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan : ASC software quality engineering practices Version 3.0.

    Energy Technology Data Exchange (ETDEWEB)

    Turgeon, Jennifer L.; Minana, Molly A.; Hackney, Patricia; Pilch, Martin M.

    2009-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in the US Department of Energy/National Nuclear Security Agency (DOE/NNSA) Quality Criteria, Revision 10 (QC-1) as 'conformance to customer requirements and expectations'. This quality plan defines the SNL ASC Program software quality engineering (SQE) practices and provides a mapping of these practices to the SNL Corporate Process Requirement (CPR) 001.3.6; 'Corporate Software Engineering Excellence'. This plan also identifies ASC management's and the software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals. This SNL ASC Software Quality Plan establishes the signatories commitments to improving software products by applying cost-effective SQE practices. This plan enumerates the SQE practices that comprise the development of SNL ASC's software products and explains the project teams opportunities for tailoring and implementing the practices.

  18. Software Quality Improvement in the OMC Team

    CERN Document Server

    Maier, Viktor

    Physicists use self-written software as a tool to fulfill their tasks and often the developed software is used for several years or even decades. If a software product lives for a long time, it has to be changed and adapted to external influences. This implies that the source code has to be read, understood and modified. The same applies to the software of the Optics Measurements and Corrections (OMC) team at CERN. Their task is to track, analyze and correct the beams in the LHC and other accelerators. To solve this task, they revert to a self-written software base with more than 150,000 physical lines of code. The base is subject to continuous changes as well. Their software does its job and is effective, but runs regrettably not efficient because some parts of the source code are in a bad shape and has a low quality. The implementation could be faster and more memory efficient. In addition it is difficult to read and understand the code. Source code files and functions are too big and identifiers do not rev...

  19. SWiFT Software Quality Assurance Plan.

    Energy Technology Data Exchange (ETDEWEB)

    Berg, Jonathan Charles [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2016-01-01

    This document describes the software development practice areas and processes which contribute to the ability of SWiFT software developers to provide quality software. These processes are designed to satisfy the requirements set forth by the Sandia Software Quality Assurance Program (SSQAP). APPROVALS SWiFT Software Quality Assurance Plan (SAND2016-0765) approved by: Department Manager SWiFT Site Lead Dave Minster (6121) Date Jonathan White (6121) Date SWiFT Controls Engineer Jonathan Berg (6121) Date CHANGE HISTORY Issue Date Originator(s) Description A 2016/01/27 Jon Berg (06121) Initial release of the SWiFT Software Quality Assurance Plan

  20. Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan. Part 2, Mappings for the ASC software quality engineering practices. Version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Ellis, Molly A.; Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, 'ASCI Software Quality Engineering: Goals, Principles, and Guidelines'. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  1. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1: ASC software quality engineering practices, Version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  2. Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan. Part 2, Mappings for the ASC software quality engineering practices. Version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Ellis, Molly A.; Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR 1.3.2 and 1.3.6 and to a Department of Energy document, 'ASCI Software Quality Engineering: Goals, Principles, and Guidelines'. This document also identifies ASC management and software project teams responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  3. Swarming Robot Design, Construction and Software Implementation

    Science.gov (United States)

    Stolleis, Karl A.

    2014-01-01

    In this paper is presented an overview of the hardware design, construction overview, software design and software implementation for a small, low-cost robot to be used for swarming robot development. In addition to the work done on the robot, a full simulation of the robotic system was developed using Robot Operating System (ROS) and its associated simulation. The eventual use of the robots will be exploration of evolving behaviors via genetic algorithms and builds on the work done at the University of New Mexico Biological Computation Lab.

  4. An Approach towards Software Quality Assessment

    Science.gov (United States)

    Srivastava, Praveen Ranjan; Kumar, Krishan

    Software engineer needs to determine the real purpose of the software, which is a prime point to keep in mind: The customer’s needs come first, and they include particular levels of quality, not just functionality. Thus, the software engineer has a responsibility to elicit quality requirements that may not even be explicit at the outset and to discuss their importance and the difficulty of attaining them. All processes associated with software quality (e.g. building, checking, improving quality) will be designed with these in mind and carry costs based on the design. Therefore, it is important to have in mind some of the possible attributes of quality. We start by identifying the metrics and measurement approaches that can be used to assess the quality of software product. Most of them can be measured subjectively because there is no solid statistics regarding them. Here, in this paper we propose an approach to measure the software quality statistically.

  5. Software Defect Prediction Models for Quality Improvement: A Literature Study

    Directory of Open Access Journals (Sweden)

    Mrinal Singh Rawat

    2012-09-01

    Full Text Available In spite of meticulous planning, well documentation and proper process control during software development, occurrences of certain defects are inevitable. These software defects may lead to degradation of the quality which might be the underlying cause of failure. In todays cutting edge competition its necessary to make conscious efforts to control and minimize defects in software engineering. However, these efforts cost money, time and resources. This paper identifies causative factors which in turn suggest the remedies to improve software quality and productivity. The paper also showcases on how the various defect prediction models are implemented resulting in reduced magnitude of defects.

  6. Implementation of a free software for quality control of IMRT; Puesta en marcha de un soltware de libre distribucion para el control de calidad IMRT

    Energy Technology Data Exchange (ETDEWEB)

    Chinillace, N.; Alonso, S.; Cortina, T.; Reinado, D.; Ricos, B.; Diaz, S.; Campayo, J. M.

    2011-07-01

    In this paper we focus on implementation and launch of software that allows us to compare quantitatively the two-dimensional dose distributions calculated and measured experimentally in IMRT treatment. The tool we are using to make this comparison is the free software DoseLab. This is a program written in MatLab and open source, thereby allowing in some cases adapt the program to the needs of each user. This program will be able to calculate the gamma function of these distributions, a parameter that simultaneously evaluates the difference in dose between two pixels of the image and the distance between them, giving us an objective and quantitative, allowing us to decide if both distributions are compatible or not.

  7. NIF Projects Controls and Information Systems Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Fishler, B

    2011-03-18

    Quality achievement for the National Ignition Facility (NIF) and the National Ignition Campaign (NIC) is the responsibility of the NIF Projects line organization as described in the NIF and Photon Science Directorate Quality Assurance Plan (NIF QA Plan). This Software Quality Assurance Plan (SQAP) is subordinate to the NIF QA Plan and establishes quality assurance (QA) activities for the software subsystems within Controls and Information Systems (CIS). This SQAP implements an activity level software quality assurance plan for NIF Projects as required by the LLNL Institutional Software Quality Assurance Program (ISQAP). Planned QA activities help achieve, assess, and maintain appropriate quality of software developed and/or acquired for control systems, shot data systems, laser performance modeling systems, business applications, industrial control and safety systems, and information technology systems. The objective of this SQAP is to ensure that appropriate controls are developed and implemented for management planning, work execution, and quality assessment of the CIS organization's software activities. The CIS line organization places special QA emphasis on rigorous configuration control, change management, testing, and issue tracking to help achieve its quality goals.

  8. Software quality assurance plan for GCS

    Science.gov (United States)

    Duncan, Stephen E.; Bailey, Elizabeth K.

    1990-01-01

    The software quality assurance (SQA) function for the Guidance and Control Software (GCS) project which is part of a software error studies research program is described. The SQA plan outlines all of the procedures, controls, and audits to be carried out by the SQA organization to ensure adherence to the policies, procedures, and standards for the GCS project.

  9. Software Quality in the Objectory Software Development Process

    NARCIS (Netherlands)

    van den Berg, Klaas; Demeyer, S.; Bosch, J.

    In this paper we discuss how software quality assurance is realized in Rational Objectory. Although much support is given through guidelines and checkpoints, the tool fails to provide clear goals and metrics for quality assessments and it only partially supports the phases in a measurement program.

  10. Software Quality in the Objectory Software Development Process

    OpenAIRE

    van den Berg, Klaas; Demeyer, S.; Bosch, J.

    1998-01-01

    In this paper we discuss how software quality assurance is realized in Rational Objectory. Although much support is given through guidelines and checkpoints, the tool fails to provide clear goals and metrics for quality assessments and it only partially supports the phases in a measurement program.

  11. The 7 Qualities of Highly Secure Software

    CERN Document Server

    Paul, Mano

    2012-01-01

    The 7 Qualities of Highly Secure Software provides a framework for designing, developing, and deploying hacker-resilient software. It uses engaging anecdotes and analogies-ranging from Aesop's fables, athletics, architecture, biology, nursery rhymes, and video games-to illustrate the qualities that are essential for the development of highly secure software. Each chapter details one of the seven qualities that can make your software highly secure and less susceptible to hacker threats. Leveraging real-world experiences and examples, the book: Explains complex security concepts in language that

  12. Requirements engineering: foundation for software quality

    NARCIS (Netherlands)

    Daneva, Maia; Pastor, Oscar

    2016-01-01

    Welcome to the proceedings of the 22nd edition of REFSQ: the International Working Conference on Requirements Engineering – Foundation for Software Quality! Requirements engineering (RE) has been recognized as a critical factor that impacts the quality of software, systems, and services. Since the

  13. Assessing Software Quality Through Visualised Cohesion Metrics

    Directory of Open Access Journals (Sweden)

    Timothy Shih

    2001-05-01

    Full Text Available Cohesion is one of the most important factors for software quality as well as maintainability, reliability and reusability. Module cohesion is defined as a quality attribute that seeks for measuring the singleness of the purpose of a module. The module of poor quality can be a serious obstacle to the system quality. In order to design a good software quality, software managers and engineers need to introduce cohesion metrics to measure and produce desirable software. A highly cohesion software is thought to be a desirable constructing. In this paper, we propose a function-oriented cohesion metrics based on the analysis of live variables, live span and the visualization of processing element dependency graph. We give six typical cohesion examples to be measured as our experiments and justification. Therefore, a well-defined, well-normalized, well-visualized and well-experimented cohesion metrics is proposed to indicate and thus enhance software cohesion strength. Furthermore, this cohesion metrics can be easily incorporated with software CASE tool to help software engineers to improve software quality.

  14. Proceedings of the Fifth Triennial Software Quality Forum 2000, Software for the Next Millennium, Software Quality Forum

    Energy Technology Data Exchange (ETDEWEB)

    Scientific Software Engineering Group, CIC-12

    2000-04-01

    The Software Quality Forum is a triennial conference held by the Software Quality Assurance Subcommittee for the Department of Energy's Quality Managers. The forum centers on key issues, information, and technology important in software development for the Nuclear Weapons Complex. This year it will be opened up to include local information technology companies and software vendors presenting their solutions, ideas, and lessons learned. The Software Quality Forum 2000 will take on a more hands-on, instructional tone than those previously held. There will be an emphasis on providing information, tools, and resources to assist developers in their goal of producing next generation software.

  15. Development and application of new quality model for software projects.

    Science.gov (United States)

    Karnavel, K; Dillibabu, R

    2014-01-01

    The IT industry tries to employ a number of models to identify the defects in the construction of software projects. In this paper, we present COQUALMO and its limitations and aim to increase the quality without increasing the cost and time. The computation time, cost, and effort to predict the residual defects are very high; this was overcome by developing an appropriate new quality model named the software testing defect corrective model (STDCM). The STDCM was used to estimate the number of remaining residual defects in the software product; a few assumptions and the detailed steps of the STDCM are highlighted. The application of the STDCM is explored in software projects. The implementation of the model is validated using statistical inference, which shows there is a significant improvement in the quality of the software projects.

  16. On the Role of Software Quality Management in Software Process Improvement

    DEFF Research Database (Denmark)

    Wiedemann Jacobsen, Jan; Kuhrmann, Marco; Münch, Jürgen

    2016-01-01

    Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. SPI addresses many aspects ranging from individual developer skills to entire organizations. It comprises, for instance, the optimization of specific activities...... in the software lifecycle as well as the creation of organizational awareness and project culture. In the course of conducting a systematic mapping study on the state-of-the-art in SPI from a general perspective, we observed Software Quality Management (SQM) being of certain relevance in SPI programs...... and to develop an initial picture of how these topics are addressed in SPI. Our findings show a fairly pragmatic contribution set in which different solutions are proposed, discussed, and evaluated. Among others, our findings indicate a certain reluctance towards standard quality or (test) maturity models...

  17. On the Role of Software Quality Management in Software Process Improvement

    DEFF Research Database (Denmark)

    Wiedemann Jacobson, Jan; Kuhrmann, Marco; Münch, Jürgen

    Software Process Improvement (SPI) programs have been implemented, inter alia, to improve quality and speed of software development. SPI addresses many aspects ranging from individual developer skills to entire organizations. It comprises, for instance, the optimization of specific activities...... in the software lifecycle as well as the creation of organizational awareness and project culture. In the course of conducting a systematic mapping study on the state-of-the-art in SPI from a general perspective, we observed Software Quality Management (SQM) being of certain relevance in SPI programs...... and to develop an initial picture of how these topics are addressed in SPI. Our findings show a fairly pragmatic contribution set in which different solutions are proposed, discussed, and evaluated. Among others, our findings indicate a certain reluctance towards standard quality or (test) maturity models...

  18. The Problems of project management software implementation in construction corporations

    Directory of Open Access Journals (Sweden)

    Kagan Pavel

    2016-01-01

    Full Text Available The article is devoted to the main problems facing many companies, implementing software project management activities have been working construction corporation. The company transferred the task performing the activities for the implementation, as well as its purpose. Revealed rules that must be followed to achieve the best results in the implementation of software project management.

  19. Lessons and challenges from software quality assessment: The ...

    African Journals Online (AJOL)

    Software development activities have continued to be plagued by a number of problems ... namely quality of design (life cycle stages) and quality of conformance. ... and challenges from software quality management for space system software.

  20. Implementing Large Projects in Software Engineering Courses

    Science.gov (United States)

    Coppit, David

    2006-01-01

    In software engineering education, large projects are widely recognized as a useful way of exposing students to the real-world difficulties of team software development. But large projects are difficult to put into practice. First, educators rarely have additional time to manage software projects. Second, classrooms have inherent limitations that…

  1. Implementing Large Projects in Software Engineering Courses

    Science.gov (United States)

    Coppit, David

    2006-01-01

    In software engineering education, large projects are widely recognized as a useful way of exposing students to the real-world difficulties of team software development. But large projects are difficult to put into practice. First, educators rarely have additional time to manage software projects. Second, classrooms have inherent limitations that…

  2. Spectrum analysis on quality requirements consideration in software design documents.

    Science.gov (United States)

    Kaiya, Haruhiko; Umemura, Masahiro; Ogata, Shinpei; Kaijiri, Kenji

    2013-12-01

    Software quality requirements defined in the requirements analysis stage should be implemented in the final products, such as source codes and system deployment. To guarantee this meta-requirement, quality requirements should be considered in the intermediate stages, such as the design stage or the architectural definition stage. We propose a novel method for checking whether quality requirements are considered in the design stage. In this method, a technique called "spectrum analysis for quality requirements" is applied not only to requirements specifications but also to design documents. The technique enables us to derive the spectrum of a document, and quality requirements considerations in the document are numerically represented in the spectrum. We can thus objectively identify whether the considerations of quality requirements in a requirements document are adapted to its design document. To validate the method, we applied it to commercial software systems with the help of a supporting tool, and we confirmed that the method worked well.

  3. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan part 2 mappings for the ASC software quality engineering practices, version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR001.3.2 and CPR001.3.6 and to a Department of Energy document, ''ASCI Software Quality Engineering: Goals, Principles, and Guidelines''. This document also identifies ASC management and software project teams' responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  4. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan part 2 mappings for the ASC software quality engineering practices, version 2.0.

    Energy Technology Data Exchange (ETDEWEB)

    Heaphy, Robert; Sturtevant, Judith E.; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Minana, Molly A.; Hackney, Patricia; Forsythe, Christi A.; Schofield, Joseph Richard, Jr. (,; .); Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2006-09-01

    The purpose of the Sandia National Laboratories Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. The plan defines the ASC program software quality practices and provides mappings of these practices to Sandia Corporate Requirements CPR001.3.2 and CPR001.3.6 and to a Department of Energy document, ''ASCI Software Quality Engineering: Goals, Principles, and Guidelines''. This document also identifies ASC management and software project teams' responsibilities in implementing the software quality practices and in assessing progress towards achieving their software quality goals.

  5. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  6. Software quality assurance: in large scale and complex software-intensive systems

    NARCIS (Netherlands)

    Mistrik, I.; Soley, R.; Ali, N.; Grundy, J.; Tekinerdogan, B.

    2015-01-01

    Software Quality Assurance in Large Scale and Complex Software-intensive Systems presents novel and high-quality research related approaches that relate the quality of software architecture to system requirements, system architecture and enterprise-architecture, or software testing. Modern software

  7. Software Engineering Management for Productivity and Quality

    Energy Technology Data Exchange (ETDEWEB)

    Karen White

    1999-10-01

    Since the advent of electronic computers, people have struggled to develop effective software engineering processes. While these processes are similar to those used by hardware engineers, the software industry has earned a reputation for late delivery of inadequate products. Most software managers are looking for ways to deliver quality products faster, or with fewer resources. The development time and product outcome of any software project can be influenced by four variables: the product characteristics, the people involved, the processes they use, and the underlying technology. In order to have an impact on the productivity of a software development effort, the manager must focus on and balance these areas. This paper will discuss effective ways to improve productivity by using this approach.

  8. A Study on the Software Quality Assurance Plan

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Hyun Tae [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    2006-07-01

    On 25 August 2006, the CMMI V1.2 (Capability Maturity Model Integration Version 1.2) was released with the new title CMMI-DEV (CMMI for Development) which supersedes the CMMI-SE/SW (CMMI for systems engineering and software engineering) V1.1. This study discusses the application of IEEE Std 730-2002, IEEE Standard for Software Quality Assurance Plans, for the implementation of the Process and Product Quality Assurance (PPQA) process area (PA) of the CMMI-DEV.

  9. An approach to ensuring quality in environmental software

    Energy Technology Data Exchange (ETDEWEB)

    Gelston, G.M.; Lundgren, R.E.; McDonald, J.P.; Hoopes, B.L.

    1998-05-01

    Environmental software is often used to determine impacts to the public, workers, and the environment from environmental contamination. It is vital, therefore, that the modeling results, and the software that provides them, be scientifically defensible and capable of withstanding the most rigorous of technical reviews. The control and assurance of quality is a critical factor for the project team that develops environmental software at the Pacific Northwest National Laboratory. This document describes the philosophy, process, and activities that ensure a quality product throughout the life cycle of requirements analysis, design, programming, modification, testing, and implementation of environmental software. Environmental software developed by the project team is designed using an object-oriented approach. This software offers increased benefits, such as ease of maintenance and retention of the development and testing legacy of individual components, over traditional hard wired software. These benefits allow the design and testing of the models and future additions to be faster and less costly. This software is developed using a modular framework concept that allows a variety of models to work within a single construct. This software has two parts: an overall system framework and a set of modules. Each module has up to three components: a user interface, a scientific model, and pre/post-processors. Each of these pieces has a different set of quality criteria associated with it. However, whatever form this software might take for a particular client, standard processes apply to protect the information from inappropriate use. The information contained within this document can be applied to most environmental software to analyze risk in multiple environmental media.

  10. Aspect-Oriented Software Quality Model: The AOSQ Model

    Directory of Open Access Journals (Sweden)

    Pankaj Kumar

    2012-04-01

    Full Text Available Nowadays, software development has become more complex and dynamic; they are expected more flexible, scalable and reusable. Under the umbrella of aspect, Aspect-Oriented Software Development (AOSD is relatively a modern programming paradigm to improve modularity in software development. Using Aspect-Oriented Programming (AOP language to implements crosscutting concerns through the introduction of a new construct Aspect like Class is defined as a modular unit of crosscutting behavior that affect multiple classes into reusable modules. Several quality models to measure the quality of software are available in literature. However, keep on developing software, and acceptance of new environment (i.e. AOP under conditions that give rise to an issue of evolvability. After the evolution of system, we have to find out how the new system needs to be extensible? What is the configurable status? Is designed pattern stable for new environment and technology? How the new system is sustainable? The objective of this paper is to propose a new quality model for AOSD to integrating some new qualityattributes in AOSQUAMO Model based which is based on ISO/IEC 9126 Quality Model, is called AspectOriented Quality (AOSQ Model. Analytic Hierarchy Process (AHP is used to evaluate an improved hierarchical quality model for AOSD.

  11. Designing, developing, and implementing software ecosystems

    DEFF Research Database (Denmark)

    Manikas, Konstantinos; Hämäläinen, Mervi; Tyrväinen, Pasi

    2017-01-01

    The notion of software ecosystems has been popular both in research and industry for more than a decade, but how software ecosystems are created still remains unclear. This becomes more of a challenge if one examines the ``creation'' of ecosystems that have high probability in surviving...... in the future, i.e. with respect to ecosystem health. In this paper, we focus on the creation of software ecosystems and propose a process for designing, developing, and establishing software ecosystems based on three basic steps and a set of activities for each step. We note that software ecosystem research...... identifies that ecosystems typically emerge from either a company deciding to allow development on their product platform or from a successful open source project. In our study we add to this knowledge by demonstrating, through two case studies, that ecosystems can emerge from more than a technological...

  12. Designing, developing, and implementing software ecosystems

    DEFF Research Database (Denmark)

    Manikas, Konstantinos; Hämäläinen, Mervi; Tyrväinen, Pasi

    2017-01-01

    The notion of software ecosystems has been popular both in research and industry for more than a decade, but how software ecosystems are created still remains unclear. This becomes more of a challenge if one examines the "creation'' of ecosystems that have high probability in surviving...... in the future, i.e. with respect to ecosystem health. In this paper, we focus on the creation of software ecosystems and propose a process for designing, developing, and establishing software ecosystems based on three basic steps and a set of activities for each step. We note that software ecosystem research...... identifies that ecosystems typically emerge from either a company deciding to allow development on their product platform or from a successful open source project. In our study we add to this knowledge by demonstrating, through two case studies, that ecosystems can emerge from more than a technological...

  13. Evaluation Framework for Quality Management Software

    Directory of Open Access Journals (Sweden)

    Nadica Hrgarek

    2008-06-01

    Full Text Available Identifying and specifying user requirements is an integral part of information systems design and is critical for the project success. More than 50% of the reasons for the project failure presented in the CHAOS report [36] and study of a US Air Force project by Sheldon et al. [33] are related to requirements. The goal of this paper is to assess the relevant user and software requirements which are the basis for an electronic quality management system selection in medical device companies. This paper describes the structured evaluation and selection process of different quality management software tools that shall support business processes. The purpose of this paper is to help the small to medium size medical device companies to choose the right quality management software which meets the company's business needs.

  14. Implementing large projects in software engineering courses

    Science.gov (United States)

    Coppit, David

    2006-03-01

    In software engineering education, large projects are widely recognized as a useful way of exposing students to the real-world difficulties of team software development. But large projects are difficult to put into practice. First, educators rarely have additional time to manage software projects. Second, classrooms have inherent limitations that threaten the realism of large projects. Third, quantitative evaluation of individuals who work in groups is notoriously difficult. As a result, many software engineering courses compromise the project experience by reducing the team sizes, project scope, and risk. In this paper, we present an approach to teaching a one-semester software engineering course in which 20 to 30 students work together to construct a moderately sized (15KLOC) software system. The approach combines carefully coordinated lectures and homeworks, a hierarchical project management structure, modern communication technologies, and a web-based project tracking and individual assessment system. Our approach provides a more realistic project experience for the students, without incurring significant additional overhead for the instructor. We present our experiences using the approach the last 2 years for the software engineering course at The College of William and Mary. Although the approach has some weaknesses, we believe that they are strongly outweighed by the pedagogical benefits.

  15. System Quality Management in Software Testing Laboratory that Chooses Accreditation

    Directory of Open Access Journals (Sweden)

    Yanet Brito R.

    2013-12-01

    Full Text Available The evaluation of software products will reach full maturity when executed by the scheme and provides third party certification. For the validity of the certification, the independent laboratory must be accredited for that function, using internationally recognized standards. This brings with it a challenge for the Industrial Laboratory Testing Software (LIPS, responsible for testing the products developed in Cuban Software Industry, define strategies that will permit it to offer services with a high level of quality. Therefore it is necessary to establish a system of quality management according to NC-ISO/IEC 17025: 2006 to continuously improve the operational capacity and technical competence of the laboratory, with a view to future accreditation of tests performed. This article discusses the process defined in the LIPS for the implementation of a Management System of Quality, from the current standards and trends, as a necessary step to opt for the accreditation of the tests performed.

  16. How To Improve Software Quality Assurance In Developing Countries

    Directory of Open Access Journals (Sweden)

    Ali Javed

    2012-04-01

    Full Text Available Quality is an important factor in software industry. Software quality depends upon the customer satisfaction which can be achieved through applying standards. In this era achieving quality software is very important because of the high customer demands. Developed countries are excelling in software industry and improving day by day. Meanwhile developing countries like Pakistan are struggling with software quality and cannot maintain reputation in International Market. Software Quality lacks due tomany reasons. This paper will address the problems for lacking interest in improving the software quality by higher authorities and software assurance team. We have provided solution to the addressed problems also.

  17. Sandia National Laboratories Advanced Simulation and Computing (ASC) software quality plan. Part 1 : ASC software quality engineering practices version 1.0.

    Energy Technology Data Exchange (ETDEWEB)

    Minana, Molly A.; Sturtevant, Judith E.; Heaphy, Robert; Hodges, Ann Louise; Boucheron, Edward A.; Drake, Richard Roy; Forsythe, Christi A.; Schofield, Joseph Richard, Jr.; Pavlakos, Constantine James; Williamson, Charles Michael; Edwards, Harold Carter

    2005-01-01

    The purpose of the Sandia National Laboratories (SNL) Advanced Simulation and Computing (ASC) Software Quality Plan is to clearly identify the practices that are the basis for continually improving the quality of ASC software products. Quality is defined in DOE/AL Quality Criteria (QC-1) as conformance to customer requirements and expectations. This quality plan defines the ASC program software quality practices and provides mappings of these practices to the SNL Corporate Process Requirements (CPR 1.3.2 and CPR 1.3.6) and the Department of Energy (DOE) document, ASCI Software Quality Engineering: Goals, Principles, and Guidelines (GP&G). This quality plan identifies ASC management and software project teams' responsibilities for cost-effective software engineering quality practices. The SNL ASC Software Quality Plan establishes the signatories commitment to improving software products by applying cost-effective software engineering quality practices. This document explains the project teams opportunities for tailoring and implementing the practices; enumerates the practices that compose the development of SNL ASC's software products; and includes a sample assessment checklist that was developed based upon the practices in this document.

  18. Software archeology: a case study in software quality assurance and design

    Energy Technology Data Exchange (ETDEWEB)

    Macdonald, John M [Los Alamos National Laboratory; Lloyd, Jane A [Los Alamos National Laboratory; Turner, Cameron J [COLORADO SCHOOL OF MINES

    2009-01-01

    Ideally, quality is designed into software, just as quality is designed into hardware. However, when dealing with legacy systems, demonstrating that the software meets required quality standards may be difficult to achieve. As the need to demonstrate the quality of existing software was recognized at Los Alamos National Laboratory (LANL), an effort was initiated to uncover and demonstrate that legacy software met the required quality standards. This effort led to the development of a reverse engineering approach referred to as software archaeology. This paper documents the software archaeology approaches used at LANL to document legacy software systems. A case study for the Robotic Integrated Packaging System (RIPS) software is included.

  19. GSM Trace Quality Analyzer (TQA) software

    OpenAIRE

    Blanchart Forne, Marc

    2016-01-01

    Connectivity is now the must-have service for enhancing passenger experience. To proof and also to show to the customers the quality of the connectivity system an user friendly mock-up has to be designed. A packet analyzer software designed to validate an existing SATCOM simulator and to improve future airline architecture networks.

  20. MCNP{trademark} Software Quality Assurance plan

    Energy Technology Data Exchange (ETDEWEB)

    Abhold, H.M.; Hendricks, J.S.

    1996-04-01

    MCNP is a computer code that models the interaction of radiation with matter. MCNP is developed and maintained by the Transport Methods Group (XTM) of the Los Alamos National Laboratory (LANL). This plan describes the Software Quality Assurance (SQA) program applied to the code. The SQA program is consistent with the requirements of IEEE-730.1 and the guiding principles of ISO 900.

  1. Software metrics: The key to quality software on the NCC project

    Science.gov (United States)

    Burns, Patricia J.

    1993-01-01

    Network Control Center (NCC) Project metrics are captured during the implementation and testing phases of the NCCDS software development lifecycle. The metrics data collection and reporting function has interfaces with all elements of the NCC project. Close collaboration with all project elements has resulted in the development of a defined and repeatable set of metrics processes. The resulting data are used to plan and monitor release activities on a weekly basis. The use of graphical outputs facilitates the interpretation of progress and status. The successful application of metrics throughout the NCC project has been instrumental in the delivery of quality software. The use of metrics on the NCC Project supports the needs of the technical and managerial staff. This paper describes the project, the functions supported by metrics, the data that are collected and reported, how the data are used, and the improvements in the quality of deliverable software since the metrics processes and products have been in use.

  2. Software Quality Perceptions of Stakeholders Involved in the Software Development Process

    Science.gov (United States)

    Padmanabhan, Priya

    2013-01-01

    Software quality is one of the primary determinants of project management success. Stakeholders involved in software development widely agree that quality is important (Barney and Wohlin 2009). However, they may differ on what constitutes software quality, and which of its attributes are more important than others. Although, software quality…

  3. Software Quality Perceptions of Stakeholders Involved in the Software Development Process

    Science.gov (United States)

    Padmanabhan, Priya

    2013-01-01

    Software quality is one of the primary determinants of project management success. Stakeholders involved in software development widely agree that quality is important (Barney and Wohlin 2009). However, they may differ on what constitutes software quality, and which of its attributes are more important than others. Although, software quality…

  4. IMPROVING (SOFTWARE) PATENT QUALITY THROUGH THE ADMINISTRATIVE PROCESS.

    Science.gov (United States)

    Rai, Arti K

    2013-11-24

    The available evidence indicates that patent quality, particularly in the area of software, needs improvement. This Article argues that even an agency as institutionally constrained as the U.S. Patent and Trademark Office ("PTO") could implement a portfolio of pragmatic, cost-effective quality improvement strategies. The argument in favor of these strategies draws upon not only legal theory and doctrine but also new data from a PTO software examination unit with relatively strict practices. Strategies that resolve around Section 112 of the patent statute could usefully be deployed at the initial examination stage. Other strategies could be deployed within the new post-issuance procedures available to the agency under the America Invents Act. Notably, although the strategies the Article discusses have the virtue of being neutral as to technology, they are likely to have a very significant practical impact in the area of software.

  5. QualitySpy: a framework for monitoring software development processes

    Directory of Open Access Journals (Sweden)

    Marian Jureczko

    2012-03-01

    Full Text Available The growing popularity of highly iterative, agile processes creates increasing need for automated monitoring of the quality of software artifacts, which would be focused on short terms (in the case of eXtreme Programming process iteration can be limited to one week. This paper presents a framework that calculates software metrics and cooperates with development tools (e.g. source version control system and issue tracking system to describe current state of a software project with regard to its quality. The framework is designed to support high level of automation of data collection and to be useful for researchers as well as for industry. The framework is currently being developed hence the paper reports already implemented features as well as future plans. The first release is scheduled for July.

  6. 高等级公路绿化质量评价系统软件的设计与实现%The design and implementation of quality evaluation software system of highway green landscape

    Institute of Scientific and Technical Information of China (English)

    袁荷伟; 谌昌强

    2012-01-01

    公路绿化质量评价系统是公路养护质量评价系统的一个子系统,其软件的设计对公路养护实现计算机管理、制定养护计划、提高养护水平起着重要的作用,本文运用了Visual FoxPro6.0编程语言,并以模块划分的方式设计出一套功能齐全的公路绿化质量评价系统,此系统包括3个子系统:高速公路、一级公路、二级公路绿化质量评价子系统,该软件能有效地完成评价标准和评价计分的计算,进而对公路绿化质量作出客观公正的评价。%Highway green landscape quality evaluation software system is a subsystem of highway maintenance quality e- valuation system, this system is important to implement computer management of highway maintenance, to make mainte- nance plan and to improve maintenance level. An evaluation software is introduced in this paper, which is developed with Visual Fox.Pro 6.0 language and is based on modularization. This system includes three subsystems : expressway, first - grade highway and second - grade highway evaluation system. This software can complete the calculation of evaluation standard and scoring efficiently and reach an objective and fair evaluation.

  7. Developing high-quality educational software.

    Science.gov (United States)

    Johnson, Lynn A; Schleyer, Titus K L

    2003-11-01

    The development of effective educational software requires a systematic process executed by a skilled development team. This article describes the core skills required of the development team members for the six phases of successful educational software development. During analysis, the foundation of product development is laid including defining the audience and program goals, determining hardware and software constraints, identifying content resources, and developing management tools. The design phase creates the specifications that describe the user interface, the sequence of events, and the details of the content to be displayed. During development, the pieces of the educational program are assembled. Graphics and other media are created, video and audio scripts written and recorded, the program code created, and support documentation produced. Extensive testing by the development team (alpha testing) and with students (beta testing) is conducted. Carefully planned implementation is most likely to result in a flawless delivery of the educational software and maintenance ensures up-to-date content and software. Due to the importance of the sixth phase, evaluation, we have written a companion article on it that follows this one. The development of a CD-ROM product is described including the development team, a detailed description of the development phases, and the lessons learned from the project.

  8. Total Quality Management Implementation Plan.

    Science.gov (United States)

    1989-06-01

    E 14. SUBJECT TERMS 15. NUMBER OF PAGES TOM ( Total Quality Management ), Continuous Process Improvement,_________ Depot Operations, Supply Support 16... Quality Management Implementation Plan 6. AUTHOR(S) 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) 8. PERFORMING ORGANIZATION Defense General...Reduction PrOtect (O704.Ot8SL Wasilngton, OC 20503j. .EPORT DATE 3. REPORT TYPE AND DATES COVERED June 19891 4. TITLE AND SUBTITLE S. FUNDING NUMBERS Total

  9. Master Pump Shutdown MPS Software Quality Assurance Plan (SQAP)

    Energy Technology Data Exchange (ETDEWEB)

    BEVINS, R.R.

    2000-09-20

    The MPSS Software Quality Assurance (SQAP) describes the tools and strategy used in the development of the MPSS software. The document also describes the methodology for controlling and managing changes to the software.

  10. Implementation of Digital Watermarking Using MATLAB Software

    Directory of Open Access Journals (Sweden)

    Karnpriya Vyas

    2012-11-01

    Full Text Available Digital watermarking holds significant promise as one of the keys to protecting proprietary digital content in the coming years. It focuses on embedding information inside a digital object such that the embedded information is in separable bound to the object. The proposed scheme has been implemented on MATLAB, as it is a high level technical computing language and interactive environment for algorithm development, data visualization, data analysis, and numerical computation. We wanted to know about how one can embed information in an image such that he can later claim the ownership of that image by extracting back the embedded information. Hence, “copyright protection” of images was our main motivation in starting this project. This paper includes, Implementation of several watermarking algorithms and examines them in terms of how they meet the requirements of different applications and general requirements of watermarking.

  11. Implementation of Digital Watermarking Using MATLAB Software

    Directory of Open Access Journals (Sweden)

    Karnpriya Vyas

    2012-10-01

    Full Text Available Digital watermarking holds significant promise as one of the keys to protecting proprietary digital content in the coming years. It focuses on embedding information inside a digital object such that the embedded information is in separable bound to the object. The proposed scheme has been implemented on MATLAB, as it is a high level technical computing language and interactive environment for algorithm development, data visualization, data analysis, and numerical computation. We wanted to know about how one can embed information in an image such that he can later claim the ownership of that image by extracting back the embedded information. Hence, “copyright protection” of images was our main motivation in starting this project. This paper includes, Implementation of several watermarking algorithms and examines them in terms of how they meet the requirements of different applications and general requirements of watermarking.

  12. N286.7-99, A Canadian standard specifying software quality management system requirements for analytical, scientific, and design computer programs and its implementation at AECL

    Energy Technology Data Exchange (ETDEWEB)

    Abel, R. [R and M Abel Consultants Inc. (Canada)

    2000-07-01

    Analytical, scientific, and design computer programs (referred to in this paper as 'scientific computer programs') are developed for use in a large number of ways by the user-engineer to support and prove engineering calculations and assumptions. These computer programs are subject to frequent modifications inherent in their application and are often used for critical calculations and analysis relative to safety and functionality of equipment and systems. N286.7-99(4) was developed to establish appropriate quality management system requirements to deal with the development, modification, and application of scientific computer programs. N286.7-99 provides particular guidance regarding the treatment of legacy codes.

  13. Operational excellence (six sigma) philosophy: Application to software quality assurance

    Energy Technology Data Exchange (ETDEWEB)

    Lackner, M.

    1997-11-01

    This report contains viewgraphs on operational excellence philosophy of six sigma applied to software quality assurance. This report outlines the following: goal of six sigma; six sigma tools; manufacturing vs administrative processes; Software quality assurance document inspections; map software quality assurance requirements document; failure mode effects analysis for requirements document; measuring the right response variables; and questions.

  14. FSO and quality of service software prediction

    Science.gov (United States)

    Bouchet, O.; Marquis, T.; Chabane, M.; Alnaboulsi, M.; Sizun, H.

    2005-08-01

    Free-space optical (FSO) communication links constitute an alternative option to radio relay links and to optical cables facing growth needs in high-speed telecommunications (abundance of unregulated bandwidth, rapid installation, availability of low-cost optical components offering a high data rate, etc). Their operationalisation requires a good knowledge of the atmospheric effects which can negatively affect role propagation and the availability of the link, and thus to the quality of service (QoS). Better control of these phenomena will allow for the evaluation of system performance and thus assist with improving reliability. The aim of this paper is to compare the behavior of a FSO link located in south of France (Toulouse: with the following parameters: around 270 meters (0.2 mile) long, 34 Mbps data rate, 850 nm wavelength and PDH frame) with airport meteorological data. The second aim of the paper is to assess in-house FSO quality of service prediction software, through comparing simulations with the optical link data and the weather data. The analysis uses in-house software FSO quality of service prediction software ("FSO Prediction") developed by France Telecom Research & Development, which integrates news fog fading equations (compare to Kim & al.) and includes multiple effects (geometrical attenuation, atmospheric fading, rain, snow, scintillation and refraction attenuation due to atmospheric turbulence, optical mispointing attenuation). The FSO link field trial, intended to enable the demonstration and evaluation of these different effects, is described; and preliminary results of the field trial, from December 2004 to May 2005, are then presented.

  15. Software Quality Measurement Demonstration Project II.

    Science.gov (United States)

    1987-10-01

    Report ............................... 1-7 1.4-4 Time Log .................................................. 1-8 2.0-1 ECOAEA Scores for Worksheet 1...34 Conduct independent software quality evaluation and validation to include goal specification, data collection, worksheet scoring, assessment of time logs ...8217n ’~’t"~ i m ’u ~ ~ ej 𔃺 vall-datc. s;Cores )..3SieL dd Spcf ’,t~ii,, it -~ L Th~ 1r 3QM prc:"d’:ae 1c ricvn r )o aL used i n measuring the cpaiity

  16. Rules of thumb to increase the software quality through testing

    Science.gov (United States)

    Buttu, M.; Bartolini, M.; Migoni, C.; Orlati, A.; Poppi, S.; Righini, S.

    2016-07-01

    The software maintenance typically requires 40-80% of the overall project costs, and this considerable variability mostly depends on the software internal quality: the more the software is designed and implemented to constantly welcome new changes, the lower will be the maintenance costs. The internal quality is typically enforced through testing, which in turn also affects the development and maintenance costs. This is the reason why testing methodologies have become a major concern for any company that builds - or is involved in building - software. Although there is no testing approach that suits all contexts, we infer some general guidelines learned during the Development of the Italian Single-dish COntrol System (DISCOS), which is a project aimed at producing the control software for the three INAF radio telescopes (the Medicina and Noto dishes, and the newly-built SRT). These guidelines concern both the development and the maintenance phases, and their ultimate goal is to maximize the DISCOS software quality through a Behavior-Driven Development (BDD) workflow beside a continuous delivery pipeline. We consider different topics and patterns; they involve the proper apportion of the tests (from end-to-end to low-level tests), the choice between hardware simulators and mockers, why and how to apply TDD and the dependency injection to increase the test coverage, the emerging technologies available for test isolation, bug fixing, how to protect the system from the external resources changes (firmware updating, hardware substitution, etc.) and, eventually, how to accomplish BDD starting from functional tests and going through integration and unit tests. We discuss pros and cons of each solution and point out the motivations of our choices either as a general rule or narrowed in the context of the DISCOS project.

  17. Unisys' experience in software quality and productivity management of an existing system

    Science.gov (United States)

    Munson, John B.

    1988-01-01

    A summary of Quality Improvement techniques, implementation, and results in the maintenance, management, and modification of large software systems for the Space Shuttle Program's ground-based systems is provided.

  18. A Framework for Analyzing Software Quality using Hierarchical Clustering

    Directory of Open Access Journals (Sweden)

    Arashdeep Kaur

    2011-02-01

    Full Text Available Fault proneness data available in the early software life cycle from previous releases or similar kind of projects will aid in improving software quality estimations. Various techniques have been proposed in the literature which includes statistical method, machine learning methods, neural network techniques and clustering techniques for the prediction of faulty and non faulty modules in the project. In this study, Hierarchical clustering algorithm is being trained and tested with lifecycle data collected from NASA projects namely, CM1, PC1 and JM1 as predictive models. These predictive models contain requirement metrics and static code metrics. We have combined requirement metric model with static code metric model to get fusion metric model. Further we have investigated that which of the three prediction models is found to be the best prediction model on the basis of fault detection. The basic hypothesis of software quality estimation is that automatic quality prediction models enable verificationexperts to concentrate their attention and resources at problem areas of the system under development. The proposed approach has been implemented in MATLAB 7.4. The results show that when all the prediction techniques are evaluated, the best prediction model is found to be the fusion metric model. This proposed model is also compared with other quality models available in the literature and is found to be efficient for predicting faulty modules.

  19. Design and implementation of modular software for programming mobile robots

    Directory of Open Access Journals (Sweden)

    Luca Iocchi

    2008-11-01

    Full Text Available This article describes a software development toolkit for programming mobile robots, that has been used on different platforms and for different robotic applications. We address design choices, implementation issues and results in the realization of our robot programming environment, that has been devised and built from many people since 1998. We believe that the proposed framework is extremely useful not only for experienced robotic software developers, but also for students approaching robotic research projects.

  20. Software Communication Architecture Implementation and Its Waveform Application

    Institute of Scientific and Technical Information of China (English)

    SUN Pei-gang; ZHAO Hai; WANG Ting-chang; FAN Jian-hua

    2006-01-01

    This paper attempts to do a research on the development of software defined radio(SDR) based on software communication architecture(SCA). Firstly, SCA is studied and a whole reference model of SCA3.0 core framework (CF)is realized; Secondly, an application-specific FM3TR waveform is implemented on the platform of common software based on the reference model; Thirdly, from the point of view of real-time performance and software reuse, tests and validations are made on the above realized CF reference model and FM3TR waveform. As a result, the SCA-compliant SDR has favorable interoperability and software portability and can satisfy the real-time performance requirements which are not too rigorous.

  1. Simplifying Complex Software Assembly: The Component Retrieval Language and Implementation

    CERN Document Server

    Seidel, Eric L; Brandt, Steven; Löffler, Frank; Schnetter, Erik; 10.1145/1838574.1838592

    2010-01-01

    Assembling simulation software along with the associated tools and utilities is a challenging endeavor, particularly when the components are distributed across multiple source code versioning systems. It is problematic for researchers compiling and running the software across many different supercomputers, as well as for novices in a field who are often presented with a bewildering list of software to collect and install. In this paper, we describe a language (CRL) for specifying software components with the details needed to obtain them from source code repositories. The language supports public and private access. We describe a tool called GetComponents which implements CRL and can be used to assemble software. We demonstrate the tool for application scenarios with the Cactus Framework on the NSF TeraGrid resources. The tool itself is distributed with an open source license and freely available from our web page.

  2. Software Implemented Fault-Tolerant (SIFT) user's guide

    Science.gov (United States)

    Green, D. F., Jr.; Palumbo, D. L.; Baltrus, D. W.

    1984-01-01

    Program development for a Software Implemented Fault Tolerant (SIFT) computer system is accomplished in the NASA LaRC AIRLAB facility using a DEC VAX-11 to interface with eight Bendix BDX 930 flight control processors. The interface software which provides this SIFT program development capability was developed by AIRLAB personnel. This technical memorandum describes the application and design of this software in detail, and is intended to assist both the user in performance of SIFT research and the systems programmer responsible for maintaining and/or upgrading the SIFT programming environment.

  3. Effective organizational solutions for implementation of DBMS software packages

    Science.gov (United States)

    Jones, D.

    1984-01-01

    The space telescope management information system development effort is a guideline for discussing effective organizational solutions used in implementing DBMS software. Focus is on the importance of strategic planning. The value of constructing an information system architecture to conform to the organization's managerial needs, the need for a senior decision maker, dealing with shifting user requirements, and the establishment of a reliable working relationship with the DBMS vendor are examined. Requirements for a schedule to demonstrate progress against a defined timeline and the importance of continued monitoring for production software control, production data control, and software enhancements are also discussed.

  4. A cooperative approach to ensure software product quality

    Directory of Open Access Journals (Sweden)

    M Sangeetha

    2011-07-01

    Full Text Available Quality is the process of ensuring that software developed satisfies customer requirements. Among the many software quality attributes like functionality, usability, capability, maintainability, etc., reliability is a major factor to assure quality of the software. Reliability ensures that software is failure free. In this work, we propose to ensure quality through a behavioral model that evaluates business requirements and gives priority for quality attributes. The model consists of behavioural and human perspectives in assessment. It is used for assessment of software developed.

  5. Lessons learned from development and quality assurance of software systems at the Halden Project

    Energy Technology Data Exchange (ETDEWEB)

    Bjorlo, T.J.; Berg, O.; Pehrsen, M.; Dahll, G.; Sivertsen, T. [OECD Halden Reactor Project (Norway)

    1996-03-01

    The OECD Halden Reactor Project has developed a number of software systems within the research programmes. These programmes have comprised a wide range of topics, like studies of software for safety-critical applications, development of different operator support systems, and software systems for building and implementing graphical user interfaces. The systems have ranged from simple prototypes to installations in process plants. In the development of these software systems, Halden has gained much experience in quality assurance of different types of software. This paper summarises the accumulated experience at the Halden Project in quality assurance of software systems. The different software systems being developed at the Halden Project may be grouped into three categories. These are plant-specific software systems (one-of-a-kind deliveries), generic software products, and safety-critical software systems. This classification has been found convenient as the categories have different requirements to the quality assurance process. In addition, the experience from use of software development tools and proprietary software systems at Halden, is addressed. The paper also focuses on the experience gained from the complete software life cycle, starting with the software planning phase and ending with software operation and maintenance.

  6. Maximizing ROI on software development

    CERN Document Server

    Sikka, Vijay

    2004-01-01

    A brief review of software development history. Software complexity crisis. Software development ROI. The case for global software development and testing. Software quality and test ROI. How do you implement global software development and testing. Case studies.

  7. The use of Software Quality Metrics in Software Maintenance

    OpenAIRE

    Kafura, Dennis G.; Reddy, Geereddy R.

    1985-01-01

    This paper reports on a modest study which relates seven different software complexity metrics to the experience of maintenance activities performed on a medium size sofhvare system. Three different versions of the system that evolved over aperiod of three years were analyzed in this study. A major revision of the system, while still in its design phase, was also analyzed. The results of this study indicate: (1) that the growth in system complexity as determined by the software...

  8. Resistance Factors in the Implementation of Software Process Improvement Project in Malaysia

    Directory of Open Access Journals (Sweden)

    Mohd H.N.M. Nasir

    2008-01-01

    Full Text Available For the past two decades, improving the quality of software has become an important agenda in the software industries as they have been assessed according to standards such as the CMM, CMM-I, ISO 9000 SIX-SIGMA and etc. As the result, software process improvement project implementations have been the main targets by most software companies. However, many initiatives are facing variety of problems and difficulties due to lack of guidance and experience. Hence, this research attempted to identify and analyze main resistance factors which influenced the implementation of the software process improvement project specifically companies operated in Malaysia including local and multi-national companies. The findings helped other software companies to manage future projects through the use of preventive actions or proper planning which intended to lessen anticipated problems during software process improvement projects implementation. This research used a survey instrument to gather data from 29 companies operated across Malaysia with the total of 174 business and software professionals responded. Average of 4 to 8 questionnaires were distributed to each company with the objective of getting wider views on each SPI project. The questionnaires were mainly distributed to professionals who are directly involved in SPI projects. The results showed that the most critical resistance factor is lack of adhesion and participation of the entire individual involved in SPI projects. This result is similar with the result gained by Brietzke and Rabello which they have conducted it in Brazil and corroborated the research findings experience in SPI project.

  9. PARTICULARITIES OF QUALITY EVALUATION IN A SOFTWARE COMPANY

    Directory of Open Access Journals (Sweden)

    Cătălin AFRĂSINEI

    2010-01-01

    Full Text Available Quality management is a management domain very discussed and disputed nowadays and this is the first sign it is a very modern, needed and present concept in theory and practice. Some are seeing it as a solution to prepare things in the way they are needed, and the instrument which might guarantee a proper environment of keeping them in a specified and constant form. The application of quality management is a quality management system that has to be designed, developed and implemented to achieve the aim of quality. The article has the purpose to briefly present what a quality management system should mean in a software company, why it should be periodically evaluated and how it might be done. In the second part it points out the characteristics of the audit as a general evaluation instrument and the main contribution consists on the author’s endeavor to mark out the particularities of an audit process carried out on a software company, considering the fact that particularization increases the changes to easier and earlier succeed with such an activity on a practical basis.

  10. Transitions in Classroom Technology: Instructor Implementation of Classroom Management Software

    Science.gov (United States)

    Ackerman, David; Chung, Christina; Sun, Jerry Chih-Yuan

    2014-01-01

    The authors look at how business instructor needs are fulfilled by classroom management software (CMS), such as Moodle, and why instructors are sometimes slow to implement it. Instructors at different universities provided both qualitative and quantitative responses regarding their use of CMS. The results indicate that the top needs fulfilled by…

  11. Design and Implementation of Software for Resistance Welding Process Simulations

    DEFF Research Database (Denmark)

    Zhang, Wenqi

    2003-01-01

    by engineers and technicians in industry, all of the important parameters in resistance welding are considered and automatically implemented into the software. With the specially designed graphic user interface for Windows, engineers (even without prior knowledge of FEM) can quickly learn and easily operate...... and utilize the software. All industrial users, including welding engineers from DaimlerChrysler, Volkswangen, PSA Peugeot Citroen, VOLVO, Siemens, ABB and so on, have started using the software just after taking a one-day training course. With the user-friendly facilities for flexible geometric design...... of work pieces and electrodes as well as process parameter settings similar to real machine settings, the software has been readily applied in industry for supporting product development and process optimization. After simulation, the dynamic process parameters are graphically displayed. The distributions...

  12. Design and Implementation of Software for Resistance Welding Process Simulations

    DEFF Research Database (Denmark)

    Zhang, Wenqi

    2003-01-01

    by engineers and technicians in industry, all of the important parameters in resistance welding are considered and automatically implemented into the software. With the specially designed graphic user interface for Windows, engineers (even without prior knowledge of FEM) can quickly learn and easily operate...... and utilize the software. All industrial users, including welding engineers from DaimlerChrysler, Volkswangen, PSA Peugeot Citroen, VOLVO, Siemens, ABB and so on, have started using the software just after taking a one-day training course. With the user-friendly facilities for flexible geometric design...... of work pieces and electrodes as well as process parameter settings similar to real machine settings, the software has been readily applied in industry for supporting product development and process optimization. After simulation, the dynamic process parameters are graphically displayed. The distributions...

  13. SRGM Analyzers Tool of SDLC for Software Improving Quality

    Directory of Open Access Journals (Sweden)

    Mr. Girish Nille

    2014-11-01

    Full Text Available Software Reliability Growth Models (SRGM have been developed to estimate software reliability measures such as software failure rate, number of remaining faults and software reliability. In this paper, the software analyzers tool proposed for deriving several software reliability growth models based on Enhanced Non-homogeneous Poisson Process (ENHPP in the presence of imperfect debugging and error generation. The proposed models are initially formulated for the case when there is no differentiation between failure observation and fault removal testing processes and then this extended for the case when there is a clear differentiation between failure observation and fault removal testing processes. Many Software Reliability Growth Models (SRGM have been developed to describe software failures as a random process and can be used to measure the development status during testing. With SRGM software consultants can easily measure (or evaluate the software reliability (or quality and plot software reliability growth charts.

  14. Experiences with Software Quality Metrics in the EMI middleware

    Science.gov (United States)

    Alandes, M.; Kenny, E. M.; Meneses, D.; Pucciani, G.

    2012-12-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering - Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project to extract “code metrics” on the status of the software products and “process metrics” related to the quality of the development and support process such as reaction time to critical bugs, requirements tracking and delays in product releases.

  15. Improving Performance of Software Implemented Floating Point Addition

    DEFF Research Database (Denmark)

    Hindborg, Andreas Erik; Karlsson, Sven

    2011-01-01

    We outline and evaluate hardware extensions to an integer processor pipeline which allow IEEE 754 oating point, FP, addition to be eciently implemented in software. With a very moderate increase in hardware resources, our perfor- mance evaluation shows that, for a benchmark that executes 12.5% FP...... addition instructions, our approach exhibits a rel- ative slowdown of 3.38 to 15.15 as compared to dedicated hardware. This is a signicant improvement of pure software emulation which leads to relative slowdowns up to 45.33....

  16. A new quality of service FSO software

    Science.gov (United States)

    Chabane, Mourad; Al Naboulsi, Maher C.; Sizun, Herve; Bouchet, M. O.

    2004-09-01

    Free space Optical links (FSO) in visible and infrared wavelengths constitute an interesting alternative to new transmission channels for cordless phone, data-processing networks and high definition television. One finds a choice of varied manufacturers and they propose products whose performances are characterized by a raised rate of transmission, from 2 Mbps to 10 Gbps. But the announced ranges are very important, from 100 to 10 000 meters, in spite of the fact that many manufacturers try to indicate the possible ranges according to time, these indications completely miss standardization and are hardly exploitable because, generally, it is very difficult to know the percentage of time during which a value is reached or exceeded. Availability and reliability of a FSO link depend on used systems but also on climatic and atmospheric parameters such as rain, snow or fog. It is the purpose of our study. Its finality is software which integrates: - Results of a bibliographical search (geometrical, aerosols, rain, snow, fog, scintillation, and environment solar light attenuation), - European integrated surface weather data, hour per hour, over several years (1995-1999). The result is the presentation of the Quality of Service Light version software which allows predicting; starting from the data of equipment (power, wavelength, receiver sensibility), geographical situation of a site in Europe (geographical coordinates, altitude, height/ground) and climatic and atmospheric parameter (relative humidity, ground rugosity, albedo, solar radiation, etc) the availability of a FSO link for the following period (year, the most unfavourable month, 8am to 8pm period and 8 pm to 8 am period. The interruption probabilities for each type of attenuation are also mentioned (aerosols, scintillation, ambient solar light, rain, snow, etc).

  17. On Quality and Measures in Software Engineering

    Science.gov (United States)

    Bucur, Ion I.

    2006-01-01

    Complexity measures are mainly used to estimate vital information about reliability and maintainability of software systems from regular analysis of the source code. Such measures also provide constant feedback during a software project to assist the control of the development procedure. There exist several models to classify a software product's…

  18. Pragmatic quality metrics for evolutionary software development models

    Science.gov (United States)

    Royce, Walker

    1990-01-01

    Due to the large number of product, project, and people parameters which impact large custom software development efforts, measurement of software product quality is a complex undertaking. Furthermore, the absolute perspective from which quality is measured (customer satisfaction) is intangible. While we probably can't say what the absolute quality of a software product is, we can determine the relative quality, the adequacy of this quality with respect to pragmatic considerations, and identify good and bad trends during development. While no two software engineers will ever agree on an optimum definition of software quality, they will agree that the most important perspective of software quality is its ease of change. We can call this flexibility, adaptability, or some other vague term, but the critical characteristic of software is that it is soft. The easier the product is to modify, the easier it is to achieve any other software quality perspective. This paper presents objective quality metrics derived from consistent lifecycle perspectives of rework which, when used in concert with an evolutionary development approach, can provide useful insight to produce better quality per unit cost/schedule or to achieve adequate quality more efficiently. The usefulness of these metrics is evaluated by applying them to a large, real world, Ada project.

  19. Quality Assurance in Software Development: An Exploratory Investigation in Software Project Failures and Business Performance

    Science.gov (United States)

    Ichu, Emmanuel A.

    2010-01-01

    Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…

  20. Quality Assurance in Software Development: An Exploratory Investigation in Software Project Failures and Business Performance

    Science.gov (United States)

    Ichu, Emmanuel A.

    2010-01-01

    Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…

  1. Requirements Prioritization: Challenges and Techniques for Quality Software Development

    Directory of Open Access Journals (Sweden)

    Muhammad Abdullah Awais

    2016-04-01

    Full Text Available Every organization is aware of the consequences and importance of requirements for the development of quality software product whether local or global. Requirement engineering phase of development with focus on the prioritization of requirements is going under huge research every day because in any development methodology, all requirements cannot be implemented at same time so requirements are prioritized to be implemented to give solution as early as possible in phases as scheduled in incremental fashion. Numerous frameworks and practices have been devised, in progress and some being discovered day by day. With such huge knowledge database and research available, it has always been confusing to decide which technique to follow to gain maximum results. Thus many projects fail because of the wrong choice in requirement prioritization because it’s really difficult to employ right technique and framework at right time. And problems do not end here rather due to strict deadlines, it’s often best to develop system in parts by different team members dispersed globally with diverse methodologies and differences and in this situation it becomes more difficult to prioritize requirements. Main focus would be on ETVX based prioritization [1] for in house development and requirement prioritization of software developed globally by diverse team members [2]. This paper will try to provide an overview of different prioritization techniques for software requirement, and a critical analysis of ETVX based model will be presented to highlight issues and challenges in this proposed model of requirement prioritization in [1] and improved version of this model will be presented while an analysis of requirement prioritization for software developed in global environment [2] also be presented.

  2. Design and Implementation of Convex Analysis of Mixtures Software Suite

    OpenAIRE

    Meng, Fan

    2012-01-01

    Various convex analysis of mixtures (CAM) based algorithms have been developed to address real world blind source separation (BSS) problems and proven to have good performances in previous papers. This thesis reported the implementation of a comprehensive software CAM-Java, which contains three different CAM based algorithms, CAM compartment modeling (CAM-CM), CAM non-negative independent component analysis (CAM-nICA), and CAM non-negative well-grounded component analysis (CAM-nWCA). The imp...

  3. IMPROVED SOFTWARE QUALITY ASSURANCE TECHNIQUES USING SAFE GROWTH MODEL

    Directory of Open Access Journals (Sweden)

    M.Sangeetha

    2010-09-01

    Full Text Available In our lives are governed by large, complex systems with increasingly complex software, and the safety, security, and reliability of these systems has become a major concern. As the software in today’ssystems grows larger, it has more defects, and these defects adversely affect the safety, security, and reliability of the systems. Software engineering is the application of a systematic, disciplined, quantifiable approach to the development, operation, andmaintenance of software. Software divides into two pieces: internal and external quality characteristics.External quality characteristics are those parts of a product that face its users, where internal quality characteristics are those that do not.Quality is conformance to product requirements and should be free. This research concerns the role of software Quality. Software reliability is an important facet of software quality. It is the probability of failure-freeoperation of a computer program in a specified environment for a specified time. In software reliability modeling, the parameters of the model are typically estimated from the test data of the corresponding component. However, the widely used point estimatorsare subject to random variations in the data, resulting in uncertainties in these estimated parameters. This research describes a new approach to the problem of software testing. The approach is based on Bayesian graphical models and presents formal mechanisms forthe logical structuring of the software testing problem, the probabilistic and statistical treatment of the uncertainties to be addressed, the test design and analysis process, and the incorporation and implication of test results. Once constructed, the models produced are dynamic representations of the software testingproblem. It explains need of the common test-and-fix software quality strategy is no longer adequate, and characterizes the properties of the quality strategy.

  4. The 7 C's for Creating Living Software: A Research Perspective for Quality-Oriented Software Engineering

    NARCIS (Netherlands)

    Aksit, Mehmet

    2004-01-01

    This article proposes the 7 C's for realizing quality-oriented software engineering practices. All the desired qualities of this approach are expressed in short by the term living software. The 7 C's are: Concern-oriented processes, Canonical models, Composable models, Certifiable models,

  5. The 7 C's for Creating Living Software: A Research Perspective for Quality Oriented Software Engineering.

    NARCIS (Netherlands)

    Aksit, Mehmet

    2004-01-01

    This article proposes the 7 C's for realizing quality-oriented software engineering practices. All the desired qualities of this approach are expressed in short by the term living software. The 7 C's are: Concern-oriented processes, Canonical models, Composable models, Certiable models, Constructibl

  6. Systems and software quality the next step for industrialisation

    CERN Document Server

    Wieczorek, Martin; Bons, Heinz

    2014-01-01

    Software and systems quality is playing an increasingly important role in the growth of almost all - profit and non-profit - organisations. Quality is vital to the success of enterprises in their markets. Most small trade and repair businesses use software systems in their administration and marketing processes. Every doctor's surgery is managing its patients using software. Banking is no longer conceivable without software. Aircraft, trucks and cars use more and more software to handle their increasingly complex technical systems. Innovation, competition and cost pressure are always present i

  7. Experiences with Software Quality Metrics in the EMI middlewate

    CERN Document Server

    Alandes, M; Meneses, D; Pucciani, G

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project to ...

  8. Experiences with Software Quality Metrics in the EMI Middleware

    CERN Document Server

    CERN. Geneva

    2012-01-01

    The EMI Quality Model has been created to define, and later review, the EMI (European Middleware Initiative) software product and process quality. A quality model is based on a set of software quality metrics and helps to set clear and measurable quality goals for software products and processes. The EMI Quality Model follows the ISO/IEC 9126 Software Engineering – Product Quality to identify a set of characteristics that need to be present in the EMI software. For each software characteristic, such as portability, maintainability, compliance, etc, a set of associated metrics and KPIs (Key Performance Indicators) are identified. This article presents how the EMI Quality Model and the EMI Metrics have been defined in the context of the software quality assurance activities carried out in EMI. It also describes the measurement plan and presents some of the metrics reports that have been produced for the EMI releases and updates. It also covers which tools and techniques can be used by any software project t...

  9. Total Quality Management Implementation Strategy: Directorate of Quality Assurance

    Science.gov (United States)

    1989-05-01

    Directorate of Quality Assurance Total Quality Management Implementation Strategy 6. AUTHOR(S) 7. PERFORMING ORGANIZATION NAME(S) AND ADORESS(ES) 8...that job. DTIC 7: t-I ECTE I SEP29 1989 14. SUBJECT TERMS 15. NUMBER OF PAGES TOM ( Total Quality Management ), Quality Assurance, Continuous Process...Prescribed by ANSI Std Z39-18 29d- 102 DEFENSE LOGISTICS AGENCY DIRECTORATE OF QUALITY ASSURANCE 0 TOTAL QUALITY MANAGEMENT IMPLEMENTATION STRATEGY 0 89 9

  10. Software Quality Metrics for Geant4: An Initial Assessment

    CERN Document Server

    Ronchieri, Elisabetta; Giacomini, Francesco

    2016-01-01

    In the context of critical applications, such as shielding and radiation protection, ensuring the quality of simulation software they depend on is of utmost importance. The assessment of simulation software quality is important not only to determine its adoption in experimental applications, but also to guarantee reproducibility of outcome over time. In this study, we present initial results from an ongoing analysis of Geant4 code based on established software metrics. The analysis evaluates the current status of the code to quantify its characteristics with respect to documented quality standards; further assessments concern evolutions over a series of release distributions. We describe the selected metrics that quantify software attributes ranging from code complexity to maintainability, and highlight what metrics are most effective at evaluating radiation transport software quality. The quantitative assessment of the software is initially focused on a set of Geant4 packages, which play a key role in a wide...

  11. A Software Quality Evaluation System: JT-SQE

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    JT-SQE system is a software quality and measurement system. Itsdesign w a s based on the Chinese national standards of software product evaluation and qua lity characteristics. The JT-SQE system consists of two parts. One is the mode l for software quality measurement, which is of hierarchical structure. The other is the process of requirements definition, measurement and rating. The system i s a feasible model for software quality evaluation and measurement, and it has t he advantage of a friendly user interface, simple operation, ease of revision an d maintenance, and expansible measurements.

  12. The CMS Data Quality Monitoring software experience and future improvements

    CERN Document Server

    De Guio, Federico

    2013-01-01

    The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release Validation, to constantly validate the functionality and the performance of the reconstruction software; in Monte Carlo productions. The central tool to deliver Data Quality information is a web site for browsing data quality histograms (DQM GUI). In this contribution the usage of the DQM Software in the different environments and its integration in the CMS Reconstruction Software Framework and in all production workflows are presented.

  13. A multiobjective module-order model for software quality enhancement

    NARCIS (Netherlands)

    Khoshgoftaar, TM; Liu, Y; Seliya, N

    2004-01-01

    The knowledge, prior to system operations, of which program modules are problematic is valuable to a software quality assurance team, especially when there is a constraint on software quality enhancement resources. A cost-effective approach for allocating such resources is to obtain a prediction in

  14. 2003 SNL ASCI applications software quality engineering assessment report.

    Energy Technology Data Exchange (ETDEWEB)

    Schofield, Joseph Richard, Jr.; Ellis, Molly A.; Williamson, Charles Michael; Bonano, Lora A.

    2004-02-01

    This document describes the 2003 SNL ASCI Software Quality Engineering (SQE) assessment of twenty ASCI application code teams and the results of that assessment. The purpose of this assessment was to determine code team compliance with the Sandia National Laboratories ASCI Applications Software Quality Engineering Practices, Version 2.0 as part of an overall program assessment.

  15. Faster issue resolution with higher technical quality of software

    NARCIS (Netherlands)

    Bijlsma, D.; Ferreira, M.A.; Luijten, B.; Visser, J.

    2011-01-01

    We performed an empirical study of the relation between technical quality of software products and the issue resolution performance of their maintainers. In particular, we tested the hypothesis that ratings for source code maintainability, as employed by the Software Improvement Group (SIG) quality

  16. A pattern framework for software quality assessment and tradeoff analysis

    NARCIS (Netherlands)

    Folmer, Eelke; Boscht, Jan

    2007-01-01

    The earliest design decisions often have a significant impact on software quality and are the most costly to revoke. One of the challenges in architecture design is to reduce the frequency of retrofit problems in software designs; not being able to improve the quality of a system cost effectively, a

  17. Software Quality - Introduction to the Special Theme

    NARCIS (Netherlands)

    Cleve, A.; Vinju, J.J.

    2014-01-01

    The introduction of fast and cheap computer and networking hardware enables the spread of software. Software, in a nutshell, represents an unprecedented ability to channel creativity and innovation. The joyful act of simply writing computer programs for existing ICT infrastructure can change the wor

  18. Software quality assurance plan for the National Ignition Facility integrated computer control system

    Energy Technology Data Exchange (ETDEWEB)

    Woodruff, J.

    1996-11-01

    Quality achievement is the responsibility of the line organizations of the National Ignition Facility (NIF) Project. This Software Quality Assurance Plan (SQAP) applies to the activities of the Integrated Computer Control System (ICCS) organization and its subcontractors. The Plan describes the activities implemented by the ICCS section to achieve quality in the NIF Project`s controls software and implements the NIF Quality Assurance Program Plan (QAPP, NIF-95-499, L-15958-2) and the Department of Energy`s (DOE`s) Order 5700.6C. This SQAP governs the quality affecting activities associated with developing and deploying all control system software during the life cycle of the NIF Project.

  19. Total Quality Management (TQM). Implementers Workshop

    Science.gov (United States)

    1990-05-15

    SHEE’T :s t’ii ,rrl DEPARTMENT OF DEFENSE May 15, 1990 Lfl CN I TOTAL QUALITY MANAGEMENT (TQM) Implementers Workshop © Copyright 1990 Booz.Allen...must be continually performed in order to achieve successful TQM implementation. 1-5 = TOTAL QUALITY MANAGEMENT Implementers Workshop Course Content...information, please refer to the student manual, Total Quality Management (TOM) Awareness Seminar, that was provided for the Awareness Course. You may

  20. Guidance and Control Software Project Data - Volume 4: Configuration Management and Quality Assurance Documents

    Science.gov (United States)

    Hayhurst, Kelly J. (Editor)

    2008-01-01

    The Guidance and Control Software (GCS) project was the last in a series of software reliability studies conducted at Langley Research Center between 1977 and 1994. The technical results of the GCS project were recorded after the experiment was completed. Some of the support documentation produced as part of the experiment, however, is serving an unexpected role far beyond its original project context. Some of the software used as part of the GCS project was developed to conform to the RTCA/DO-178B software standard, "Software Considerations in Airborne Systems and Equipment Certification," used in the civil aviation industry. That standard requires extensive documentation throughout the software development life cycle, including plans, software requirements, design and source code, verification cases and results, and configuration management and quality control data. The project documentation that includes this information is open for public scrutiny without the legal or safety implications associated with comparable data from an avionics manufacturer. This public availability has afforded an opportunity to use the GCS project documents for DO-178B training. This report provides a brief overview of the GCS project, describes the 4-volume set of documents and the role they are playing in training, and includes configuration management and quality assurance documents from the GCS project. Volume 4 contains six appendices: A. Software Accomplishment Summary for the Guidance and Control Software Project; B. Software Configuration Index for the Guidance and Control Software Project; C. Configuration Management Records for the Guidance and Control Software Project; D. Software Quality Assurance Records for the Guidance and Control Software Project; E. Problem Report for the Pluto Implementation of the Guidance and Control Software Project; and F. Support Documentation Change Reports for the Guidance and Control Software Project.

  1. An Approach to Early Prediction of Software Quality

    Institute of Scientific and Technical Information of China (English)

    YAO Lan; YANG Bo

    2007-01-01

    Due to the rapid development of computers and their applications, early software quality prediction in software industry becomes more and more crucial. Software quality prediction model is very helpful for decision-makings such as the allocation of resource in module verification and validation. Nevertheless, due to the complicated situations of software development process in the early stage, the applicability and accuracy of these models are still under research. In this paper, a software quality prediction model based on a fuzzy neural network is presented, which takes into account both the internal factors and external factors of software. With hybrid-learning algorithm, the proposed model can deal with multiple forms of data as well as incomplete information, which helps identify design errors early and avoid expensive rework.

  2. MCTSSA Software Reliability Handbook, Volume III: Integration of Software Metrics with Quality and Reliability

    OpenAIRE

    Schneidewind, Norman F.

    1997-01-01

    The purpose of this handbook is threefold. Specifically, it: (1) Serves as a reference guide for implementing standard software reliability practices at Marine Corps Tactical Systems Support Activity and aids in applying the software reliability model. (2) Serves as a tool for managing the software reliability program. (3) Serves as a training aid. U.S. Marine Corps Tactical Systems Support Activity, Camp Pendleton, CA. RLACH

  3. An Approach to Measuring Software Quality Perception

    Science.gov (United States)

    Hofman, Radoslaw

    Perception measuring and perception management is an emerging approach in the area of product management. Cognitive, psychological, behavioral and neurological theories, tools and methods are being employed for a better understanding of the mechanisms of a consumer's attitude and decision processes. Software is also being defined as a product, however this kind of product is significantly different from all other products. Software products are intangible and it is difficult to trace their characteristics which are strongly dependant on a dynamic context of use.

  4. Total quality management implementation guidelines

    Energy Technology Data Exchange (ETDEWEB)

    1993-12-01

    These Guidelines were designed by the Energy Quality Council to help managers and supervisors in the Department of Energy Complex bring Total Quality Management to their organizations. Because the Department is composed of a rich mixture of diverse organizations, each with its own distinctive culture and quality history, these Guidelines are intended to be adapted by users to meet the particular needs of their organizations. For example, for organizations that are well along on their quality journeys and may already have achieved quality results, these Guidelines will provide a consistent methodology and terminology reference to foster their alignment with the overall Energy quality initiative. For organizations that are just beginning their quality journeys, these Guidelines will serve as a startup manual on quality principles applied in the Energy context.

  5. Ada Implementation Guide. Software Engineering With Ada. Volume 2

    Science.gov (United States)

    1994-04-01

    international standard LADY-LOV 11/25/91 10240 Article on life of Ada Lovelace by Carol L. James and Duncan E. Morrill with note on the naming of the Ada ...A aa a- S I !/ A -- I Volume II I I! I Ii Ada Implementation GuideI Software Engineering With Ada I DTIC QUALn1T :1 1 ’. I April 1994 *ýS94-18856 JJ...A-13 A.1 Government Sources ...................................... A-1 A.1.1 Organizations .................................... A-2 Ada

  6. Implementing Ideas for Improving Software Citation and Credit

    CERN Document Server

    Teuben, Peter; Berriman, G Bruce; DuPrie, Kimberly; Mink, Jessica; Robitaille, Thomas; Shortridge, Keith; Taylor, Mark; Warmels, Rein

    2016-01-01

    Improving software citation and credit continues to be a topic of interest across and within many disciplines, with numerous efforts underway. In this Birds of a Feather (BoF) session, we started with a list of actionable ideas from last year's BoF and other similar efforts and worked alone or in small groups to begin implementing them. Work was captured in a common Google document; the session organizers will disseminate or otherwise put this information to use in or for the community in collaboration with those who contributed.

  7. A Systems Perspective on the Quality Description of Software Components

    Directory of Open Access Journals (Sweden)

    Otto Preiss

    2003-02-01

    Full Text Available In this paper we present our rational for proposing a conceptual model for the description of quality attributes of software artifacts, in particular suited to software components. The scientific foundations for our quality description model are derived from researching systems science for its value to software engineering. In this work we realized that software engineering is concerned with a number of interrelated conceptual as well as concrete systems. Each of them exhibits the basic system theoretic principles and is strongly related to certain types of qualities. Such qualities receive particular attention in the context of large software systems, where systems are a combination of in-house and third party products and are increasingly integrated by means of software component technology. Consequently, a quality data sheet is needed by component users to gain trust in, and to evaluate the possible employment of, a candidate component. Interestingly, the concept of a software component appears in most of the aforementioned different types of systems. Hence, it is an excellent means to carry quality related information that belonged to different spheres up to now. The qualities range from those related to the development economics to those related to the execution performance.

  8. A Systems Perspective on the Quality Description of Software Components

    Directory of Open Access Journals (Sweden)

    Otto Preiss

    2003-02-01

    Full Text Available In this paper we present our rational for proposing a conceptual model for the description of quality attributes of software artifacts, in particular suited to software components. The scientific foundations for our quality description model are derived from researching systems science for its value to software engineering. In this work we realized that software engineering is concerned with a number of interrelated conceptual as well as concrete systems. Each of them exhibits the basic system theoretic principles and is strongly related to certain types of qualities. Such qualities receive particular attention in the context of large software systems, where systems are a combination of in-house and third party products and are increasingly integrated by means of software component technology. Consequently, a quality data sheet is needed by component users to gain trust in, and to evaluate the possible employment of, a candidate component. Interestingly, the concept of a software component appears in most of the aforementioned different types of systems. Hence, it is an excellent means to carry quality related information that belonged to different spheres up to now. The qualities range from those related to the development economics to those related to the execution performance.

  9. Tool Use Within NASA Software Quality Assurance

    Science.gov (United States)

    Shigeta, Denise; Port, Dan; Nikora, Allen P.; Wilf, Joel

    2013-01-01

    As space mission software systems become larger and more complex, it is increasingly important for the software assurance effort to have the ability to effectively assess both the artifacts produced during software system development and the development process itself. Conceptually, assurance is a straightforward idea - it is the result of activities carried out by an organization independent of the software developers to better inform project management of potential technical and programmatic risks, and thus increase management's confidence in the decisions they ultimately make. In practice, effective assurance for large, complex systems often entails assessing large, complex software artifacts (e.g., requirements specifications, architectural descriptions) as well as substantial amounts of unstructured information (e.g., anomaly reports resulting from testing activities during development). In such an environment, assurance engineers can benefit greatly from appropriate tool support. In order to do so, an assurance organization will need accurate and timely information on the tool support available for various types of assurance activities. In this paper, we investigate the current use of tool support for assurance organizations within NASA, and describe on-going work at JPL for providing assurance organizations with the information about tools they need to use them effectively.

  10. Software Design Improvements. Part 2; Software Quality and the Design and Inspection Process

    Science.gov (United States)

    Lalli, Vincent R.; Packard, Michael H.; Ziemianski, Tom

    1997-01-01

    The application of assurance engineering techniques improves the duration of failure-free performance of software. The totality of features and characteristics of a software product are what determine its ability to satisfy customer needs. Software in safety-critical systems is very important to NASA. We follow the System Safety Working Groups definition for system safety software as: 'The optimization of system safety in the design, development, use and maintenance of software and its integration with safety-critical systems in an operational environment. 'If it is not safe, say so' has become our motto. This paper goes over methods that have been used by NASA to make software design improvements by focusing on software quality and the design and inspection process.

  11. Embedding Quality Function Deployment In Software Development ...

    African Journals Online (AJOL)

    2013-03-01

    Mar 1, 2013 ... product development: Marketing, Design. Engineering,. Quality ... The basic Quality Function Deployment methodology involves ... structure and planning approach, requires that more time be ..... process, and strategy. This is ...

  12. Continuous Software Quality analysis for the ATLAS experiment

    CERN Document Server

    Washbrook, Andrew; The ATLAS collaboration

    2017-01-01

    The regular application of software quality tools in large collaborative projects is required to reduce code defects to an acceptable level. If left unchecked the accumulation of defects invariably results in performance degradation at scale and problems with the long-term maintainability of the code. Although software quality tools are effective for identification there remains a non-trivial sociological challenge to resolve defects in a timely manner. This is a ongoing concern for the ATLAS software which has evolved over many years to meet the demands of Monte Carlo simulation, detector reconstruction and data analysis. At present over 3.8 million lines of C++ code (and close to 6 million total lines of code) are maintained by a community of hundreds of developers worldwide. It is therefore preferable to address code defects before they are introduced into a widely used software release. Recent wholesale changes to the ATLAS software infrastructure have provided an ideal opportunity to apply software quali...

  13. Conflict Resolution (CORE) for Software Quality Factors

    Science.gov (United States)

    1993-05-01

    theory. Evaluate recent algorithm and concept developments for possible use in CORE. IH-34 SECTION IV REFERENCES [1] Jeffrey A. Lasky and Alan R...February, 1985. [5] Ruben Prieto -Diaz and Guillermo Arango (eds.), Domain Analysis and Software Systems Modeling, IEEE Computer Society Press, 1991. [6

  14. Quality Management Systems Implementation Compared With Organizational Maturity in Hospital.

    Science.gov (United States)

    Moradi, Tayebeh; Jafari, Mehdi; Maleki, Mohammad Reza; Naghdi, Seyran; Ghiasvand, Hesam

    2015-07-27

    A quality management system can provide a framework for continuous improvement in order to increase the probability of customers and other stakeholders' satisfaction. The test maturity model helps organizations to assess the degree of maturity in implementing effective and sustained quality management systems; plan based on the current realities of the organization and prioritize their improvement programs. We aim to investigate and compare the level of organizational maturity in hospitals with the status of quality management systems implementation. This analytical cross sectional study was conducted among hospital administrators and quality experts working in hospitals with over 200 beds located in Tehran. In the first step, 32 hospitals were selected and then 96 employees working in the selected hospitals were studied. The data were gathered using the implementation checklist of quality management systems and the organization maturity questionnaire derived from ISO 10014. The content validity was calculated using Lawshe method and the reliability was estimated using test - retest method and calculation of Cronbach's alpha coefficient. The descriptive and inferential statistics were used to analyze the data using SPSS 18 software. According to the table, the mean score of organizational maturity among hospitals in the first stage of quality management systems implementation was equal to those in the third stage and hypothesis was rejected (p-value = 0.093). In general, there is no significant difference in the organizational maturity between the first and third level hospitals (in terms of implementation of quality management systems). Overall, the findings of the study show that there is no significant difference in the organizational maturity between the hospitals in different levels of the quality management systems implementation and in fact, the maturity of the organizations cannot be attributed to the implementation of such systems. As a result, hospitals

  15. Assessing quality in software development: An agile methodology approach

    OpenAIRE

    V. Rodríguez-Hernández; M.C. Espino-Gudiño; J.L. González-Pérez; J. Gudiño-Bazaldúa; Victor Castano

    2015-01-01

    A novel methodology, result of 10 years of in-field testing, which makes possible the convergence of different types of models and quality standards for Engineering and Computer Science Faculties, is presented. Since most software-developing companies are small and medium sized, the projects developed must focuson SCRUM and Extreme Programming (XP), opposed to a RUP, which is quite heavy, as well as on Personal Software Process (PSP) and Team Software Process (TSP), which provide students wit...

  16. Quality Market: Design and Field Study of Prediction Market for Software Quality Control

    Science.gov (United States)

    Krishnamurthy, Janaki

    2010-01-01

    Given the increasing competition in the software industry and the critical consequences of software errors, it has become important for companies to achieve high levels of software quality. While cost reduction and timeliness of projects continue to be important measures, software companies are placing increasing attention on identifying the user…

  17. Human Factors in Software Development Processes: Measuring System Quality

    DEFF Research Database (Denmark)

    Abrahão, Silvia; Baldassarre, Maria Teresa; Caivano, Danilo

    2016-01-01

    Software Engineering and Human-Computer Interaction look at the development process from different perspectives. They apparently use very different approaches, are inspired by different principles and address different needs. But, they definitively have the same goal: develop high quality software...

  18. Implementation of Software Configuration Management Process by Models: Practical Experiments and Learned Lessons

    Directory of Open Access Journals (Sweden)

    Bartusevics Arturs

    2014-12-01

    Full Text Available Nowadays software configuration management process is not only dilemma which system should be used for version control or how to merge changes from one source code branch to other. There are multiple tasks such as version control, build management, deploy management, status accounting, bug tracking and many others that should be solved to support full configuration management process according to most popular quality standards. The main scope of the mentioned process is to include only valid and tested software items to final version of product and prepare a new version as soon as possible. To implement different tasks of software configuration management process, a set of different tools, scripts and utilities should be used. The current paper provides a new model-based approach to implementation of configuration management. Using different models, a new approach helps to organize existing solutions and develop new ones by a parameterized way, thus increasing reuse of solutions. The study provides a general description of new model-based conception and definitions of all models needed to implement a new approach. The second part of the paper contains an overview of criteria, practical experiments and lessons learned from using new models in software configuration management. Finally, further works are defined based on results of practical experiments and lessons learned.

  19. Achieving Product Qualities Through Software Architecture Practices

    Science.gov (United States)

    2016-06-14

    maintenance Mass customization Mind share improved efficiency and productivity © 2004 by Carnegie Mellon University page 10 Substantial Quick Sustainable...problems articulated by the specification. • Boxes define the elements or “parts” of the system. • Lines define the interactions or between the parts...function were all that mattered , any monolithic software would do, ..but other things matter … • Modifiability • Interoperability • Availability

  20. Implementation of an OAIS Repository Using Free, Open Source Software

    Science.gov (United States)

    Flathers, E.; Gessler, P. E.; Seamon, E.

    2015-12-01

    The Northwest Knowledge Network (NKN) is a regional data repository located at the University of Idaho that focuses on the collection, curation, and distribution of research data. To support our home institution and others in the region, we offer services to researchers at all stages of the data lifecycle—from grant application and data management planning to data distribution and archive. In this role, we recognize the need to work closely with other data management efforts at partner institutions and agencies, as well as with larger aggregation efforts such as our state geospatial data clearinghouses, data.gov, DataONE, and others. In the past, one of our challenges with monolithic, prepackaged data management solutions is that customization can be difficult to implement and maintain, especially as new versions of the software are released that are incompatible with our local codebase. Our solution is to break the monolith up into its constituent parts, which offers us several advantages. First, any customizations that we make are likely to fall into areas that can be accessed through Application Program Interfaces (API) that are likely to remain stable over time, so our code stays compatible. Second, as components become obsolete or insufficient to meet new demands that arise, we can replace the individual components with minimal effect on the rest of the infrastructure, causing less disruption to operations. Other advantages include increased system reliability, staggered rollout of new features, enhanced compatibility with legacy systems, reduced dependence on a single software company as a point of failure, and the separation of development into manageable tasks. In this presentation, we describe our application of the Service Oriented Architecture (SOA) design paradigm to assemble a data repository that conforms to the Open Archival Information System (OAIS) Reference Model primarily using a collection of free and open-source software. We detail the design

  1. A survey of Canadian medical physicists: software quality assurance of in-house software.

    Science.gov (United States)

    Salomons, Greg J; Kelly, Diane

    2015-01-05

    This paper reports on a survey of medical physicists who write and use in-house written software as part of their professional work. The goal of the survey was to assess the extent of in-house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple-choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software-related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines.

  2. Predicting Software Assurance Using Quality and Reliability Measures

    Science.gov (United States)

    2014-12-01

    quality and security results ap- pear to be related. Section 5 describes other efforts underway to evaluate quality and security that are providing use...potential faults. Software security failure modes are exploita- ble conditions, such as unverified input data. A design review should confirm that the...deployed as the software is developed, whether it be on checklists for inspections or other approaches.” [Shull 2013] Staff members were trained to

  3. Quality Attributes for Mission Flight Software: A Reference for Architects

    Science.gov (United States)

    Wilmot, Jonathan; Fesq, Lorraine; Dvorak, Dan

    2016-01-01

    In the international standards for architecture descriptions in systems and software engineering (ISO/IEC/IEEE 42010), "concern" is a primary concept that often manifests itself in relation to the quality attributes or "ilities" that a system is expected to exhibit - qualities such as reliability, security and modifiability. One of the main uses of an architecture description is to serve as a basis for analyzing how well the architecture achieves its quality attributes, and that requires architects to be as precise as possible about what they mean in claiming, for example, that an architecture supports "modifiability." This paper describes a table, generated by NASA's Software Architecture Review Board, which lists fourteen key quality attributes, identifies different important aspects of each quality attribute and considers each aspect in terms of requirements, rationale, evidence, and tactics to achieve the aspect. This quality attribute table is intended to serve as a guide to software architects, software developers, and software architecture reviewers in the domain of mission-critical real-time embedded systems, such as space mission flight software.

  4. Designing and implementing a state quality award

    Science.gov (United States)

    Dobson, E. N.

    1993-02-01

    To remain competitive in today's global economy, businesses need to ensure customer satisfaction by offering high-quality products and services. Governors and state governments can play a critical role in ensuring the economic health of the business in their state by encouraging the adoption of quality practices and recognizing successful efforts by firms to improve quality and productivity. The manual is intended to help state government officials and other individuals implement a state quality award program.

  5. The effect of competition from open source software on the quality of proprietary software in the presence of network externalities

    Directory of Open Access Journals (Sweden)

    Mingqing Xing

    2015-05-01

    Full Text Available Purpose: A growing number of open source software emerges in many segments of the software market. In addition, software products usually exhibit network externalities. The purpose of this paper is to study the impact of open source software on the quality choices of proprietary software vendors when the market presents positive network externalities. Design/methodology: To analyze how open source software affects the optimal quality of proprietary software, this paper constructs two vertical differentiation models: the basic model considers proprietary software monopolizing the market, and its extended one considers proprietary software competing with open source substitute. Findings: This paper mainly finds that the presence of open source software does not necessarily lead to the quality of proprietary software decreases (or increases. The network externalities and compatibility between open source and proprietary software may change the impact of open source software on the quality of proprietary software and may affect the quality choices of proprietary software vendors. Originality/value: The main contribution of this paper is to examine the effect of open source software on the quality choices for proprietary software vendors in software markets exhibiting positive network externalities.

  6. A supernova feedback implementation for the astrophysical simulation software Arepo

    CERN Document Server

    Bubel, André-Patrick

    2016-01-01

    Supernova (SN) explosions play an important role in the development of galactic structures. The energy and momentum imparted on the interstellar medium (ISM) in so called "supernova feedback" drives turbulence, heats the gas, enriches it with heavy elements, can lead to the formation of new stars or even suppress star formation by disrupting stellar nurseries. In the numerical simulation at the sub-galactic level, not including the energy and momentum of supernovas in the physical description of the problem can also lead to several problems that might partially be resolved by including a description of supernovas. In this thesis such an implementation is attempted for the combined numerical hydrodynamics and N-body simulation software Arepo (Springel, 2010). In a stochastic process a large amount of thermal energy is imparted on a number of neighbouring cells, mimicking the effect of a supernova explosions. We test this approach by modelling the explosion of a single supernova in a uniform density medium and ...

  7. A direct implementation for influence lines in finite element software

    DEFF Research Database (Denmark)

    Jepsen, Michael S.; Damkilde, Lars

    2014-01-01

    The use of influence lines is a recognized method for determining the critical design load conditions and this paper shows a direct method for applying influence lines in any structural finite element software. The main idea is to equate displacement or angular discontinuities with nodal forces...... to consistent nodal forces, which makes it very suitable for implementation in finite element schemes and applicable for all element types, such as shell, plates, beams etc. This paper derives the consistent nodal forces for angular, lateral and axial displacement discontinuities for a Bernoulli-Euler beam......, and subsequently obtain the influence function only applying a single load case without changing the geometry or boundary conditions of the model. The new approach for determining Influence lines is based on the Müller-Breslau principle, but the discontinuous displacement fields are in the new approach equated...

  8. Implementation of Physical Layer Key Distribution using Software Defined Radios

    Directory of Open Access Journals (Sweden)

    S. Kambala

    2013-01-01

    Full Text Available It was well known from Shannon’s days that characteristics of the physical channel like attenuation, fading and noise can impair reliable communication. But it was more recently that the beneficial side effects of channel characteristics in ensuring secret communication started getting attention. Studies have been made to quantify the amount of secrecy that can be reaped by combining channel coding with security protocols. The Wiretap channel proposed by Wyner is arguably one of the oldest models of physical layer security protocols. In this paper, we present a brief tutorial introduction to the Wiretap channel, followed by an application of the physical layer model to a class of Key Distribution protocols. We present results from an implementation of key distribution protocols using Software Defined Radio tools along with physical RF hardware peripherals. We believe this approach is much more tangible and informative than computer based simulation studies.

  9. ISO9126 BASED SOFTWARE QUALITY EVALUATION USING CHOQUET INTEGRAL

    Directory of Open Access Journals (Sweden)

    Abdelkareem M. Alashqar

    2015-01-01

    Full Text Available Evaluating software quality is an important and essential issue in the development process because it helps to deliver a competitive software product. A decision of selecting the best software based on quality attributes is a type of multi-criteria decision-making (MCDM processes where interactions among criteria should be considered. This paper presents and develops quantitative evaluations by considering interactions among criteria in the MCDM problems. The aggregator methods such as Arithmetic Mean (AM and Weighted Arithmetic Mean (WAM are introduced, described and compared to Choquet Integral (CI approach which is a type of fuzzy measure used as a new method for MCDM. The comparisons are shown by evaluating and ranking software alternatives based on six main quality attributes as identified by the ISO 9126-1 standard. The evaluation experiments depend on real data collected from case studies.

  10. A Quality Based Method to Analyze Software Architectures

    Directory of Open Access Journals (Sweden)

    Farzaneh Hoseini Jabali

    2011-07-01

    Full Text Available In order to produce and develop a software system, it is necessary to have a method of choosing a suitable software architecture which satisfies the required quality attributes and maintains a trade-off between sometimes conflicting ones. Each software architecture includes a set of design decisions for each of which there are various alternatives, satisfying the quality attributes differently. At the same time various stakeholders with various quality goals participate in decision-making. In this paper a numerical method is proposed that based on the quality attributes selects the suitable software architecture for a certain software. In this method, for each design decision, different alternatives are compared in view of a certain quality attribute, and the other way around. Multi-criteria decision-making methods are used and, at the same time, time and cost constraints are considered in decision-making, too. The proposed method applies the stakeholders' opinions in decision-making according to the degree of their importance and helps the architect to select the best software architecture with more certainty.

  11. C++ Software Quality in the ATLAS experiment

    CERN Document Server

    Roe, Shaun; The ATLAS collaboration; Kluth, Stefan; Seuster, Rolf; Snyder, Scott; Obreshkov, Emil; Sherwood, Peter; Stewart, Graeme

    2016-01-01

    In this paper we explain how the C++ code quality is managed in ATLAS using a range of tools from compile-time through to run time testing and reflect on the substantial progress made in the last two years largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other available code analysis tools are also discussed, as is the role of unit testing with an example of how the googlemock framework can be applied to our codebase.

  12. Faster Defect Resolution with Higher Technical Quality of Software

    NARCIS (Netherlands)

    Luijten, B.; Visser, J.

    2010-01-01

    We performed an empirical study of the relation between technical quality of software products and the defect resolution performance of their maintainers. In particular, we tested the hypothesis that ratings for source code maintainability, as employed by the SIG quality model, are correlated with r

  13. Cluster implementation for parallel computation within MATLAB software environment

    Energy Technology Data Exchange (ETDEWEB)

    Santana, Antonio O. de; Dantas, Carlos C.; Charamba, Luiz G. da R.; Souza Neto, Wilson F. de; Melo, Silvio B. Melo; Lima, Emerson A. de O., E-mail: mailto.aos@ufpe.br, E-mail: ccd@ufpe.br, E-mail: sbm@cin.ufpe.br, E-mail: emathematics@gmail.com [Universidade Federal de Pernambuco (UFPE), Recife, PE (Brazil)

    2013-07-01

    A cluster for parallel computation with MATLAB software the COCGT - Cluster for Optimizing Computing in Gamma ray Transmission methods, is implemented. The implementation correspond to creation of a local net of computers, facilities and configurations of software, as well as the accomplishment of cluster tests for determine and optimizing of performance in the data processing. The COCGT implementation was required by data computation from gamma transmission measurements applied to fluid dynamic and tomography reconstruction in a FCC-Fluid Catalytic Cracking cold pilot unity, and simulation data as well. As an initial test the determination of SVD - Singular Values Decomposition - of random matrix with dimension (n , n), n=1000, using the Girco's law modified, revealed that COCGT was faster in comparison to the literature [1] cluster, which is similar and operates at the same conditions. Solution of a system of linear equations provided a new test for the COCGT performance by processing a square matrix with n=10000, computing time was 27 s and for square matrix with n=12000, computation time was 45 s. For determination of the cluster behavior in relation to 'parfor' (parallel for-loop) and 'spmd' (single program multiple data), two codes were used containing those two commands and the same problem: determination of SVD of a square matrix with n= 1000. The execution of codes by means of COCGT proved: 1) for the code with 'parfor', the performance improved with the labs number from 1 to 8 labs; 2) for the code 'spmd', just 1 lab (core) was enough to process and give results in less than 1 s. In similar situation, with the difference that now the SVD will be determined from square matrix with n1500, for code with 'parfor', and n=7000, for code with 'spmd'. That results take to conclusions: 1) for the code with 'parfor', the behavior was the same already described above; 2) for code with &apos

  14. The Discovery of the Implemented Software Engineering Process Using Process Mining Techniques

    Directory of Open Access Journals (Sweden)

    Mostafa Adel Zayed

    2016-02-01

    Full Text Available Process model guidance is an important feature by which the software process is orchestrated. Without complying with this guidance, the production lifecycle deviates from producing a reliable software with high-quality standards. Usually, teams break the process deliberately or impulsively. Application Lifecycle Management (ALM tools log what teams do even if they break the process. The log file could be a key to discover the behavior of the undertaken process against the targeted process model. Since the date of its introduction, Process Mining techniques have been used in business process domains with no focus on the software engineering processes. This research brings the Process Mining techniques to the software engineering domain. The research shows a conclusive effort that used a Scrum adapted process model as an example of Agile adoption. This research has applied Process Mining discovery techniques to capture the actually implemented process by the Scrum team. This application clarifies the gap between the standard process guidance and the actually implemented one. The research’s results showed that Process Mining techniques have the ability to discover and verify the deviation on both levels; the process itself as well as the work items state-machine workflows.

  15. 如何实施软件质量保证的过程和产品审计%How to Implement Software Process and Product Audit of Process and Product Quality Assurance

    Institute of Scientific and Technical Information of China (English)

    王威

    2011-01-01

    软件质量保证是CMMI模型中很重要的一个过程域,通过质量保证来客观评价过程和工作产品,使组织开发人员和管理者能客观了解过程和相关的工作产品,减少并纠正、预防实际的开发过程和结果与预期不符合。根据实践,对软件质量保证的过程和产品审计的实施经验进行了分析总结。%CMMI model for software quality assurance is a very important process area in the field,and the purpose of Process and Product Quality Assurance(PPQA) is to provide staff and management with objective insight into processes and associated work products.PP

  16. Modernized build and test infrastructure for control software at ESO: highly flexible building, testing, and automatic quality practices for telescope control software

    Science.gov (United States)

    Pellegrin, F.; Jeram, B.; Haucke, J.; Feyrin, S.

    2016-07-01

    The paper describes the introduction of a new automatized build and test infrastructure, based on the open-source software Jenkins1, into the ESO Very Large Telescope control software to replace the preexisting in-house solution. A brief introduction to software quality practices is given, a description of the previous solution, the limitations of it and new upcoming requirements. Modifications required to adapt the new system are described, how these were implemented to current software and the results obtained. An overview on how the new system may be used in future projects is also presented.

  17. A Framework For Software Quality Assurance Using Agile Methodology

    Directory of Open Access Journals (Sweden)

    Maria Sagheer

    2015-02-01

    Full Text Available Abstract Agile methodology that uses iterative and incremental approach is widely used in the industry projects as they can satisfy to the change of requirements. Efficient product delivery is achieved by using short iterations. But assuring the quality in any system is essential and imperative part which is very difficult task which raises a question that Is quality is assured in the software which is developed using agile methodology The research paper proposed a solution and summarizes that how the quality is achieved or assure in agile software development using different factors. The major focus of this research paper is to analyse the quality parameter which assure quality in the agile methodology. The research paper includes the brief overview of the parameters which assure quality in agile.

  18. Color Image Quality in Presentation Software

    Directory of Open Access Journals (Sweden)

    María S. Millán

    2008-11-01

    Full Text Available The color image quality of presentation programs is evaluated and measured using S-CIELAB and CIEDE2000 color difference formulae. A color digital image in its original format is compared with the same image already imported by the program and introduced as a part of a slide. Two widely used presentation programs—Microsoft PowerPoint 2004 for Mac and Apple's Keynote 3.0.2—are evaluated in this work.

  19. ARCHITECTURE SOFTWARE SOLUTION TO SUPPORT AND DOCUMENT MANAGEMENT QUALITY SYSTEM

    Directory of Open Access Journals (Sweden)

    Milan Eric

    2010-12-01

    Full Text Available One of the basis of a series of standards JUS ISO 9000 is quality system documentation. An architecture of the quality system documentation depends on the complexity of business system. An establishment of an efficient management documentation of system of quality is of a great importance for the business system, as well as in the phase of introducing the quality system and in further stages of its improvement. The study describes the architecture and capability of software solutions to support and manage the quality system documentation in accordance with the requirements of standards ISO 9001:2001, ISO 14001:2005 HACCP etc.

  20. SIFT - Multiprocessor architecture for Software Implemented Fault Tolerance flight control and avionics computers

    Science.gov (United States)

    Forman, P.; Moses, K.

    1979-01-01

    A brief description of a SIFT (Software Implemented Fault Tolerance) Flight Control Computer with emphasis on implementation is presented. A multiprocessor system that relies on software-implemented fault detection and reconfiguration algorithms is described. A high level reliability and fault tolerance is achieved by the replication of computing tasks among processing units.

  1. Software Industry-oriented Education with Embedded Quality Assurance Mechanisms

    Institute of Scientific and Technical Information of China (English)

    HUSSEY Matt; WU Bing

    2012-01-01

    This paper presents a broad range of suggestions on the concept of quality-assured industry-oriented higher education in software engineering, a central theme of the annual CEISIE (CEISEE this year) workshops since the first one held in Harbin, China, in 2005. It draws on the lessons of collaborative experiences involving academics and industrialists from Europe and China. These experiences make the case for a strong role for software industry- oriented higher education in the production of the software architects, developers and engineers required for the future.

  2. Software Defined GPS API: Development and Implementation of GPS Correlator Architectures Using MATLAB with Focus on SDR Implementations

    Science.gov (United States)

    2014-05-18

    NO. 0704-0188 3. DATES COVERED (From - To) - UU UU UU UU Approved for public release; distribution is unlimited. Software Defined GPS API : Development...and Implementation of GPS Correlator Architectures Using MATLAB with Focus on SDR Implementations The Software Defined GPS API was created with the...ABOVE ADDRESS. Polytechnic University of Puerto Rico 377 Ponce de Leon Ave. Hato Rey, PR 00918 -0000 ABSTRACT Software Defined GPS API : Development and

  3. Software Acquisition: Evolution, Total Quality Management, and Applications to the Army Tactical Missile System.

    Science.gov (United States)

    presents the concept of software Total Quality Management (TQM) which focuses on the entire process of software acquisition, as a partial solution to...software TQM can be applied to software acquisition. Software Development, Software Acquisition, Total Quality management (TQM), Army Tactical Missile

  4. Metric-based Evaluation of Implemented Software Architectures

    NARCIS (Netherlands)

    Bouwers, E.M.

    2013-01-01

    Software systems make up an important part of our daily lives. Just like all man- made objects, the possibilities of a software system are constrained by the choices made during its creation. The complete set of these choices can be referred to as the software architecture of a system. Since the so

  5. Implementation of Physical Layer Key Distribution using Software Defined Radios

    Directory of Open Access Journals (Sweden)

    S. Kambala

    2013-01-01

    Full Text Available It was well known from Shannon’s days that characteristics of the physical channel like attenuation, fadingand noise can impair reliable communication. But it was more recently that the beneficial side effects of channelcharacteristics in ensuring secret communication started getting attention. Studies have been made to quantify theamount of secrecy that can be reaped by combining channel coding with security protocols. The Wiretap channelproposed by Wyner is arguably one of the oldest models of physical layer security protocols. In this paper, wepresent a brief tutorial introduction to the Wiretap channel, followed by an application of the physical layer modelto a class of Key Distribution protocols. We present results from an implementation of key distribution protocolsusing Software Defined Radio tools along with physical RF hardware peripherals. We believe this approach is muchmore tangible and informative than computer based simulation studies.Defence Science Journal, 2013, 63(1, pp.6-14, DOI:http://dx.doi.org/10.14429/dsj.63.3758

  6. Scrum Method Implementation in a Software Development Project Management

    Directory of Open Access Journals (Sweden)

    Putu Adi Guna Permana

    2015-09-01

    Full Text Available To maximize the performance, companies conduct a variety of ways to increase the business profit. The work management between one company and the other company is different, so the differences in the management may cause the software to have a different business process. Software development can be defined as creating a new software or fixing the existing one. Technology developments led to increasing demand for software, Industrial Technology (IT Companies should be able to project well maintenance. The methodology in software development is used in accordance with the company's needs based on the SDLC (Software Development Life Cycle. Scrum method is a part of the Agile method that is expected to increase the speed and flexibility in software development project management.

  7. Requirements for guidelines systems: implementation challenges and lessons from existing software-engineering efforts.

    Science.gov (United States)

    Shah, Hemant; Allard, Raymond D; Enberg, Robert; Krishnan, Ganesh; Williams, Patricia; Nadkarni, Prakash M

    2012-03-09

    A large body of work in the clinical guidelines field has identified requirements for guideline systems, but there are formidable challenges in translating such requirements into production-quality systems that can be used in routine patient care. Detailed analysis of requirements from an implementation perspective can be useful in helping define sub-requirements to the point where they are implementable. Further, additional requirements emerge as a result of such analysis. During such an analysis, study of examples of existing, software-engineering efforts in non-biomedical fields can provide useful signposts to the implementer of a clinical guideline system. In addition to requirements described by guideline-system authors, comparative reviews of such systems, and publications discussing information needs for guideline systems and clinical decision support systems in general, we have incorporated additional requirements related to production-system robustness and functionality from publications in the business workflow domain, in addition to drawing on our own experience in the development of the Proteus guideline system (http://proteme.org). The sub-requirements are discussed by conveniently grouping them into the categories used by the review of Isern and Moreno 2008. We cite previous work under each category and then provide sub-requirements under each category, and provide example of similar work in software-engineering efforts that have addressed a similar problem in a non-biomedical context. When analyzing requirements from the implementation viewpoint, knowledge of successes and failures in related software-engineering efforts can guide implementers in the choice of effective design and development strategies.

  8. Cost Analysis of Poor Quality Using a Software Simulation

    Directory of Open Access Journals (Sweden)

    Jana Fabianová

    2017-02-01

    Full Text Available The issues of quality, cost of poor quality and factors affecting quality are crucial to maintaining a competitiveness regarding to business activities. Use of software applications and computer simulation enables more effective quality management. Simulation tools offer incorporating the variability of more variables in experiments and evaluating their common impact on the final output. The article presents a case study focused on the possibility of using computer simulation Monte Carlo in the field of quality management. Two approaches for determining the cost of poor quality are introduced here. One from retrospective scope of view, where the cost of poor quality and production process are calculated based on historical data. The second approach uses the probabilistic characteristics of the input variables by means of simulation, and reflects as a perspective view of the costs of poor quality. Simulation output in the form of a tornado and sensitivity charts complement the risk analysis.

  9. Assuring Quality and Reliability in Complex Avionics Systems hardware & Software

    Directory of Open Access Journals (Sweden)

    V. Haridas

    1997-01-01

    Full Text Available It is conventional wisdom in defence systems that electronic brains are where much of the present and future weapons system capability is developed. Electronic hardware advances, particularly in microprocessor, allow highly complex and sophisticated software to provide high degree of system autonomy and customisation to mission at hand. Since modern military systems are so much dependent on the proper functioning of electronics, the quality and reliability of electronic hardware and software have a profound impact on defensive capability and readiness. At the hardware level, due to the advances in microelectronics, functional capabilities of today's systems have increased. The advances in the hardware field have an impact on software also. Now a days, it is possible to incorporate more and more system functions through software, rather than going for a pure hardware solution. On the other hand complexities the systems are increasing, working energy levels of the systems are decreasing and the areas of reliability and quality assurance are becoming more and more wide. This paper covers major failure modes in microelectronic devices. The various techniques used to improve component and system reliability are described. The recent trends in expanding the scope of traditional quality assurance techniques are also discussed, considering both hardware and software.

  10. Evaluating the Effect of Software Quality Characteristics on Health Care Quality Indicators

    Directory of Open Access Journals (Sweden)

    Sakineh Aghazadeh

    2015-07-01

    Full Text Available Introduction: Various types of software are used in health care organizations to manage information and care processes. The quality of software has been an important concern for both health authorities and designers of Health Information Technology. Thus, assessing the effect of software quality on the performance quality of healthcare institutions is essential. Method: The most important health care quality indicators in relation to software quality characteristics are provided via an already performed literature review. ISO 9126 standard model is used for definition and integration of various characteristics of software quality. The effects of software quality characteristics and sub-characteristics on the healthcare indicators are evaluated through expert opinion analyses. A questionnaire comprising of 126 questions of 10-point Likert scale was used to gather opinions of experts in the field of Medical/Health Informatics. The data was analyzed using Structural Equation Modeling. Results: Our findings showed that software Maintainability was rated as the most effective factor on user satisfaction (R2 =0.89 and Functionality as the most important and independent variable affecting patient care quality (R2 =0.98. Efficiency was considered as the most effective factor on workflow (R2 =0.97, and Maintainability as the most important factor that affects healthcare communication (R2 =0.95. Usability and Efficiency were rated as the most effectual factor affecting patient satisfaction (R2 =0.80, 0.81. Reliability, Maintainability, and Efficiency were considered as the main factors affecting care costs (R2 =0.87, 0.74, 0.87. Conclusion: We presented a new model based on ISO standards. The model demonstrates and weighs the relations between software quality characteristics and healthcare quality indicators. The clear relationships between variables and the type of the metrics and measurement methods used in the model make it a reliable method to assess

  11. Software Quality Attribute Analysis by Architecture Reconstruction (SQUA3RE)

    NARCIS (Netherlands)

    Stormer, C.

    2007-01-01

    Software Quality Attribute Analysis by Architecture Reconstruction (SQUA3RE) is a method that fosters a goal-driven process to evaluate the impact of what-if scenarios on existing systems. The method is partitioned into SQA2 and ARE. The SQA2 part provides the analysis models that can be used for q

  12. Expert System for Software Quality Assurance. User’s Manual,

    Science.gov (United States)

    1986-11-01

    This user’s manual describes the execution of an expert system for Software Quality Assurance (SQA). The objective of the expert system is to capture...which are all included in the batch file called SQA.BAT. The primary component is the expert system , which was developed using the EXSYS development

  13. Software Quality Attribute Analysis by Architecture Reconstruction (SQUA3RE)

    NARCIS (Netherlands)

    Stormer, C.

    2007-01-01

    Software Quality Attribute Analysis by Architecture Reconstruction (SQUA3RE) is a method that fosters a goal-driven process to evaluate the impact of what-if scenarios on existing systems. The method is partitioned into SQA2 and ARE. The SQA2 part provides the analysis models that can be used for q

  14. The software product assurance metrics study: JPL's software systems quality and productivity

    Science.gov (United States)

    Bush, Marilyn W.

    1989-01-01

    The findings are reported of the Jet Propulsion Laboratory (JPL)/Software Product Assurance (SPA) Metrics Study, conducted as part of a larger JPL effort to improve software quality and productivity. Until recently, no comprehensive data had been assembled on how JPL manages and develops software-intensive systems. The first objective was to collect data on software development from as many projects and for as many years as possible. Results from five projects are discussed. These results reflect 15 years of JPL software development, representing over 100 data points (systems and subsystems), over a third of a billion dollars, over four million lines of code and 28,000 person months. Analysis of this data provides a benchmark for gauging the effectiveness of past, present and future software development work. In addition, the study is meant to encourage projects to record existing metrics data and to gather future data. The SPA long term goal is to integrate the collection of historical data and ongoing project data with future project estimations.

  15. Accomplished Teachers Implementation of Quality Teaching Practices

    Science.gov (United States)

    Chen, Weiyun; Hammond-Bennett, Austin; Upton, Ashely; Mason, Steve

    2014-01-01

    The purpose of this study was to describe how accomplished teachers implement the quality of teaching practices in their daily lessons. The participants were four elementary physical education teachers (one male, three female). The data sources consisted of videotape of the teachers teaching 12 lessons, transcription of the taped lessons,…

  16. Bridging the gap between PAT concepts and implementation: An integrated software platform for fermentation.

    Science.gov (United States)

    Chopda, Viki R; Gomes, James; Rathore, Anurag S

    2016-01-01

    Bioreactor control significantly impacts both the amount and quality of the product being manufactured. The complexity of the control strategy that is implemented increases with reactor size, which may vary from thousands to tens of thousands of litres in commercial manufacturing. The Process Analytical Technology (PAT) initiative has highlighted the need for having robust monitoring tools and effective control schemes that are capable of taking real time information about the critical quality attributes (CQA) and the critical process parameters (CPP) and executing immediate response as soon as a deviation occurs. However, the limited flexibility that present commercial software packages offer creates a hurdle. Visual programming environments have gradually emerged as potential alternatives to the available text based languages. This paper showcases development of an integrated programme using a visual programming environment for a Sartorius BIOSTAT® B Plus 5L bioreactor through which various peripheral devices are interfaced. The proposed programme facilitates real-time access to data and allows for execution of control actions to follow the desired trajectory. Major benefits of such integrated software system include: (i) improved real time monitoring and control; (ii) reduced variability; (iii) improved performance; (iv) reduced operator-training time; (v) enhanced knowledge management; and (vi) easier PAT implementation.

  17. Develop Quality Characteristics Based Quality Evaluation Process for Ready to Use Software Products

    OpenAIRE

    Daiju Kato; Hiroshi Ishikawa

    2016-01-01

    The users who use ready to use software product had better get the products’ quality information with classified by some kind of global standard metrics or technique for their evaluation. But many of those software products’ co mpanies don’t provide the quality information because of the products are developed b y their own development and evaluation process. But those users want to get quality i...

  18. Continuous integration and quality control for scientific software

    Science.gov (United States)

    Neidhardt, A.; Ettl, M.; Brisken, W.; Dassing, R.

    2013-08-01

    Modern software has to be stable, portable, fast and reliable. This is going to be also more and more important for scientific software. But this requires a sophisticated way to inspect, check and evaluate the quality of source code with a suitable, automated infrastructure. A centralized server with a software repository and a version control system is one essential part, to manage the code basis and to control the different development versions. While each project can be compiled separately, the whole code basis can also be compiled with one central “Makefile”. This is used to create automated, nightly builds. Additionally all sources are inspected automatically with static code analysis and inspection tools, which check well-none error situations, memory and resource leaks, performance issues, or style issues. In combination with an automatic documentation generator it is possible to create the developer documentation directly from the code and the inline comments. All reports and generated information are presented as HTML page on a Web server. Because this environment increased the stability and quality of the software of the Geodetic Observatory Wettzell tremendously, it is now also available for scientific communities. One regular customer is already the developer group of the DiFX software correlator project.

  19. Arquitectura orientada a servicios para software de apoyo para el proceso personal de software A service oriented architecture for the implementation of the personal software process

    Directory of Open Access Journals (Sweden)

    Erick Salinas

    2011-06-01

    Full Text Available El presente trabajo describe una arquitectura orientada a servicios para un software que tiene como objetivo facilitar la implementación de un Proceso Personal de Software en un equipo de desarrollo u organización. Entre las características que posee este software y que son relevantes de mencionar están las de entregar extensibilidad e independencia, esto se ve reflejado en la facilidad para agregar nuevas herramientas al proceso de desarrollo de software integradas al Proceso Personal de Software con un máximo de independencia de sistemas operativos y lenguajes de programación. El software implementado realiza la recolección de los datos necesarios para el Proceso Personal de Software casi completamente automática, considerando que el administrador solamente clasifica los errores que pueden ocurrir cuando se utiliza algún lenguaje de programación en particular, entre otras pequeñas tareas. Esta facilidad de uso hace que la implementación del Proceso Personal de Software se realice exitosamente con un bajo esfuerzo requerido por los integrantes del equipo de desarrollo.This work describes a service oriented architecture of a software application that facilitates the implementation of the Personal Software Process by a development team or an organization. Some of the characteristics of this software and which are important to mention are extensibility and technical environment independence. These characteristics facilitate the process of adding new tools to the software development process integrating them to the Personal Software Process independently of the operating systems and programming languages being used. The implemented software undertakes the data collection necessary to the Personal Software Process almost automatically, since the administrator must only classify the errors that may occur when a particular programming language is used, among other small tasks. This ease of use approach helps to make the implementation of

  20. Software development tool for PicoBlaze multi-processor implementation

    OpenAIRE

    Claudiu Lung; Buchman Attila

    2012-01-01

    This paper presents a useful software tool for projects with multi PicoBlaze microprocessors implemented in FPGA circuits. Application presented in this paper which use for software development PicoBlaze SDK tool is an Automatic Packet Report System (APRS), with three PicoBlaze microprocessors implemented in FPGA circuit.

  1. Design study of Software-Implemented Fault-Tolerance (SIFT) computer

    Science.gov (United States)

    Wensley, J. H.; Goldberg, J.; Green, M. W.; Kutz, W. H.; Levitt, K. N.; Mills, M. E.; Shostak, R. E.; Whiting-Okeefe, P. M.; Zeidler, H. M.

    1982-01-01

    Software-implemented fault tolerant (SIFT) computer design for commercial aviation is reported. A SIFT design concept is addressed. Alternate strategies for physical implementation are considered. Hardware and software design correctness is addressed. System modeling and effectiveness evaluation are considered from a fault-tolerant point of view.

  2. Machine Learning and Software Quality Prediction: As an Expert System

    Directory of Open Access Journals (Sweden)

    Ekbal A. Rashid

    2014-04-01

    Full Text Available To improve the software quality the number of errors from the software must be removed. The research paper presents a study towards machine learning and software quality prediction as an expert system. The purpose of this paper is to apply the machine learning approaches, such as case-based reasoning, to predict software quality. The main objective of this research is to minimize software costs. Predict the error in software module correctly and use the results in future estimation. The novel idea behind this system is that Knowledge base (KBS building is an important task in CBR and the knowledge base can be built based on world new problems along with world new solutions. Second, reducing the maintenance cost by removing the duplicate record set from the KBS. Third, error prediction with the help of similarity functions. In this research four similarity functions have been used and these are Euclidean, Manhattan, Canberra, and Exponential. We feel that case-based models are particularly useful when it is difficult to define actual rules about a problem domain. For this purpose we have developed a case-based reasoning model and have validated it upon student data. It was observed that, Euclidean and Exponential both are good for error calculation in comparison to Manhattan and Canberra after performing five experiments. In order to obtain a result we have used indigenous tool. For finding the mean and standard deviation, SPSS version 16 and for generating graphs MATLAB 7.10.0 version have been used as an analyzing tool.

  3. QoSLight: a new quality of service FSO software

    Science.gov (United States)

    Chabane, Mourad; Alnaboulsi, Maher; Sizun, Herve; Bouchet, Olivier

    2003-08-01

    The atmospheric optical links (FSO) in visible and infrared wavelengths constitute an interesting alternative to creation of new transmission channels for the cordless phone, data-processing networks and high definition television. One finds a choice of varied manufacturers and they propose products whose performances are characterized by a raised rate of transmission, from 2 Mbps to 10 Gbps. But the announced ranges are very important, from 100 to 10,000 meters, in spite of the fact that many manufacturers try to indicate the possible ranges according to time, these indications completely miss standardization and are hardly exploitable because, generally, it is very difficult to know the percentage of time during which a value is reached or exceeded. Availability and reliability of a FSO link depend on used systems but also on climatic and atmospheric parameters such as rain, snow or fog. Library search underlined the lack of reliable data to be able to lay down, in a precise way, the statistical availability of such links, like one usually does for the radio transmission. Before to implement an effective FSO links, we need to know their availability and their reliability. It is the purpose of our study. Its finality is a software which integrate (1) Results of a library search (geometrical attenuation, aerosols, scintillation, environment light, etc), (2) English and French integrated surface weather data, hour per hour, over several years (1995-1999). The result is the presentation of this software, "QoSLight" (Quality of Service Light), making it possible to predict; starting from the data of equipment (power, wavelength, receiver sensibility), geographical situation of a site in France or England (geographical coordinates, altitude, height/ground) and climatic and atmospheric parameter (relative humidity, ground rugosity, albedo, solar radiation, etc) the availability of a FSO link for the following period (year, the most unfavourable month, 8am to 8pm period

  4. Image quality dependence on image processing software in computed radiography

    Directory of Open Access Journals (Sweden)

    Lourens Jochemus Strauss

    2012-06-01

    Full Text Available Background. Image post-processing gives computed radiography (CR a considerable advantage over film-screen systems. After digitisation of information from CR plates, data are routinely processed using manufacturer-specific software. Agfa CR readers use MUSICA software, and an upgrade with significantly different image appearance was recently released: MUSICA2. Aim. This study quantitatively compares the image quality of images acquired without post-processing (flatfield with images processed using these two software packages. Methods. Four aspects of image quality were evaluated. An aluminium step-wedge was imaged using constant mA at tube voltages varying from 40 to 117kV. Signal-to-noise ratios (SNRs and contrast-to-noise Ratios (CNRs were calculated from all steps. Contrast variation with object size was evaluated with visual assessment of images of a Perspex contrast-detail phantom, and an image quality figure (IQF was calculated. Resolution was assessed using modulation transfer functions (MTFs. Results. SNRs for MUSICA2 were generally higher than the other two methods. The CNRs were comparable between the two software versions, although MUSICA2 had slightly higher values at lower kV. The flatfield CNR values were better than those for the processed images. All images showed a decrease in CNRs with tube voltage. The contrast-detail measurements showed that both MUSICA programmes improved the contrast of smaller objects. MUSICA2 was found to give the lowest (best IQF; MTF measurements confirmed this, with values at 3.5 lp/mm of 10% for MUSICA2, 8% for MUSICA and 5% for flatfield. Conclusion. Both MUSICA software packages produced images with better contrast resolution than unprocessed images. MUSICA2 has slightly improved image quality than MUSICA.

  5. 自动气象站质量控制系统中覆盖软件的设计与实现%Design and implementation of covering software in automatic meteorological station quality control system

    Institute of Scientific and Technical Information of China (English)

    张志强; 孙超

    2013-01-01

    区域气象自动站收集降水数据的质量,直接决定着气象预警的漏警与虚假概率,但是仅依靠自动站间采集数据的时空相关性难以准确验证数据的有效性.通过天气雷达对降水滴谱的分析可以对自动站数据进行校验.但是,雷达站的系统异构性和地理环境的复杂性使其难以判定所覆盖的气象站,易产生错误校验.结合天气雷达的层体扫模式和雷达波束的遮挡模型,构建了雷达站与自动气象站点的覆盖匹配算法,基于该算法实现的自动站与雷达站匹配软件可有效的实现雷达站与自动站配对,从而为对自动站降水数据进行校验提供基础保障.%Precipitation data quality of regional automatic weather stations determines the missed alarm and false probability of weather warning, but it is difficult to verify the validity of the data by only relying on space and time correlation of automatic data. Precipitation of weather radar can verify automatic weather station data. However, it is difficult to determine the coverage of weather stations by radar station due to heterogeneous and complex geographical environment of radar station, which results in constant error-checking. This paper built a coverage matching algorithm of radar stations and automatic weather stations by using the layer volume scan mode and the radar beam occlusion model of weather radar. Automatic stations and radar stations matching software based on the algorithm can effectively support Automatic Weather Station ( AWS) precipitation data inspection.

  6. Implementation and Simulation Results using Autonomous Aerobraking Development Software

    Science.gov (United States)

    Maddock, Robert W.; DwyerCianciolo, Alicia M.; Bowes, Angela; Prince, Jill L. H.; Powell, Richard W.

    2011-01-01

    An Autonomous Aerobraking software system is currently under development with support from the NASA Engineering and Safety Center (NESC) that would move typically ground-based operations functions to onboard an aerobraking spacecraft, reducing mission risk and mission cost. The suite of software that will enable autonomous aerobraking is the Autonomous Aerobraking Development Software (AADS) and consists of an ephemeris model, onboard atmosphere estimator, temperature and loads prediction, and a maneuver calculation. The software calculates the maneuver time, magnitude and direction commands to maintain the spacecraft periapsis parameters within design structural load and/or thermal constraints. The AADS is currently tested in simulations at Mars, with plans to also evaluate feasibility and performance at Venus and Titan.

  7. Real-time Kernel Implementation Practice Program for Embedded Software Engineers' Education and its Evaluation

    Science.gov (United States)

    Yoshida, Toshio; Matsumoto, Masahide; Seo, Katsuhiko; Chino, Shinichiro; Sugino, Eiji; Sawamoto, Jun; Koizumi, Hisao

    A real-time kernel (henceforth RTK) is in the center place of embedded software technology, and the understanding of RTK is indispensable for the embedded system design. To implement RTK, it is necessary to understand languages that describe RTK software program code, system programming manners, software development tools, CPU on that RTK runs and the interface between software and hardware, etc. in addition to understanding of RTK itself. This means RTK implementation process largely covers embedded software implementation process. Therefore, it is thought that RTK implementation practice program is very effective as a means of the acquisition of common embedded software skill in addition to deeper acquisition of RTK itself. In this paper, we propose to apply RTK implementing practice program to embedded software engineers educational program. We newly developed very small and step-up type RTK named μK for educational use, and held a seminar that used μK as a teaching material for the students of information science and engineers of the software house. As a result, we confirmed that RTK implementation practice program is very effective for the acquisition of embedded software common skill.

  8. Software Implementation of Elliptic Curve Encryption over Binary Field

    Institute of Scientific and Technical Information of China (English)

    ZHANG Xianfeng; QIN Zhiguang; ZHOU Shijie; LIU Jinde

    2003-01-01

    The mathematical theory for elliptic curve encryption based on optimal normal basis (ONB) over Fm2 is introduced. Then an elliptic curve cryptography(ECC) based encryption scheme is analyzed and designed. The mechanism for key exchange based on Diffie-Hellman is described in details for further applications. Based on these theoretic foundations, the software based on ECC is developed and an application is provided. The software is characterized by excellent security as well as high efficiency.

  9. ALMA software releases versus quality management: and the winner is...

    Science.gov (United States)

    Allaert, Erik; Pasquato, Moreno; Soto, Rubén.

    2016-08-01

    After its inauguration and the formal completion of the construction phase, the software development effort at the Atacama Large Millimeter/submillimeter Array (ALMA) continues at roughly the same level as during construction - gradually adding capabilities as required by and offered to the scientific community. In the run-up to a new yearly Observing Cycle several software releases have to be prepared, incorporating this new functionality. However, the ALMA observatory is used on a daily basis to produce scientific data for the approved projects within the current Observing Cycle, and also by engineering teams to extend existing capabilities or to diagnose and fix problems - so the preparation of new software releases up to their deployment competes for resources with all other activities. Testing a new release and ensuring its quality is of course fundamental, but can on the other hand not monopolize the observatory's resources or jeopardize its commitments to the scientific community.

  10. Software Tools for Electrical Quality Assurance in the LHC

    CERN Document Server

    Bednarek, Mateusz

    2011-01-01

    There are over 1600 superconducting magnet circuits in the LHC machine. Many of them consist of a large number of components electrically connected in series. This enhances the sensitivity of the whole circuits to electrical faults of individual components. Furthermore, circuits are equipped with a large number of instrumentation wires, which are exposed to accidental damage or swapping. In order to ensure safe operation, an Electrical Quality Assurance (ELQA) campaign is needed after each thermal cycle. Due to the complexity of the circuits, as well as their distant geographical distribution (tunnel of 27km circumference divided in 8 sectors), suitable software and hardware platforms had to be developed. The software combines an Oracle database, LabView data acquisition applications and PHP-based web follow-up tools. This paper describes the software used for the ELQA of the LHC.

  11. Softcopy quality ruler method: implementation and validation

    Science.gov (United States)

    Jin, Elaine W.; Keelan, Brian W.; Chen, Junqing; Phillips, Jonathan B.; Chen, Ying

    2009-01-01

    A softcopy quality ruler method was implemented for the International Imaging Industry Association (I3A) Camera Phone Image Quality (CPIQ) Initiative. This work extends ISO 20462 Part 3 by virtue of creating reference digital images of known subjective image quality, complimenting the hardcopy Standard Reference Stimuli (SRS). The softcopy ruler method was developed using images from a Canon EOS 1Ds Mark II D-SLR digital still camera (DSC) and a Kodak P880 point-and-shoot DSC. Images were viewed on an Apple 30in Cinema Display at a viewing distance of 34 inches. Ruler images were made for 16 scenes. Thirty ruler images were generated for each scene, representing ISO 20462 Standard Quality Scale (SQS) values of approximately 2 to 31 at an increment of one just noticeable difference (JND) by adjusting the system modulation transfer function (MTF). A Matlab GUI was developed to display the ruler and test images side-by-side with a user-adjustable ruler level controlled by a slider. A validation study was performed at Kodak, Vista Point Technology, and Aptina Imaging in which all three companies set up a similar viewing lab to run the softcopy ruler method. The results show that the three sets of data are in reasonable agreement with each other, with the differences within the range expected from observer variability. Compared to previous implementations of the quality ruler, the slider-based user interface allows approximately 2x faster assessments with 21.6% better precision.

  12. Automated Theorem Proving in High-Quality Software Design

    Science.gov (United States)

    Schumann, Johann; Swanson, Keith (Technical Monitor)

    2001-01-01

    The amount and complexity of software developed during the last few years has increased tremendously. In particular, programs are being used more and more in embedded systems (from car-brakes to plant-control). Many of these applications are safety-relevant, i.e. a malfunction of hardware or software can cause severe damage or loss. Tremendous risks are typically present in the area of aviation, (nuclear) power plants or (chemical) plant control. Here, even small problems can lead to thousands of casualties and huge financial losses. Large financial risks also exist when computer systems are used in the area of telecommunication (telephone, electronic commerce) or space exploration. Computer applications in this area are not only subject to safety considerations, but also security issues are important. All these systems must be designed and developed to guarantee high quality with respect to safety and security. Even in an industrial setting which is (or at least should be) aware of the high requirements in Software Engineering, many incidents occur. For example, the Warshaw Airbus crash, was caused by an incomplete requirements specification. Uncontrolled reuse of an Ariane 4 software module was the reason for the Ariane 5 disaster. Some recent incidents in the telecommunication area, like illegal "cloning" of smart-cards of D2GSM handies, or the extraction of (secret) passwords from German T-online users show that also in this area serious flaws can happen. Due to the inherent complexity of computer systems, most authors claim that only a rigorous application of formal methods in all stages of the software life cycle can ensure high quality of the software and lead to real safe and secure systems. In this paper, we will have a look, in how far automated theorem proving can contribute to a more widespread application of formal methods and their tools, and what automated theorem provers (ATPs) must provide in order to be useful.

  13. Applying total quality management techniques to improve software development.

    Science.gov (United States)

    Mezher, T; Assem Abdul Malak, M; el-Medawar, H

    1998-01-01

    Total Quality Management (TQM) is a new management philosophy and a set of guiding principles that represent the basis of a continuously improving organization. This paper sheds light on the application of TQM concepts for software development. A fieldwork study was conducted on a Lebanese software development firm and its customers to determine the major problems affecting the organization's operation and to assess the level of adoption of TQM concepts. Detailed questionnaires were prepared and handed out to the firm's managers, programmers, and customers. The results of the study indicate many deficiencies in applying TQM concepts, especially in the areas of planning, defining customer requirements, teamwork, relationship with suppliers, and adopting standards and performance measures. One of the major consequences of these deficiencies is considerably increased programming errors and delays in delivery. Recommendations on achieving quality are discussed.

  14. The design strategy of scientific data quality control software for Euclid mission

    CERN Document Server

    Brescia, Massimo; Fredvik, Terje; Haugan, Stein Vidar Hagfors; Gozaliasl, Ghassem; Kirkpatrick, Charles; Kurki-Suonio, Hannu; Longo, Giuseppe; Nilsson, Kari; Wiesmann, Martin

    2016-01-01

    The most valuable asset of a space mission like Euclid are the data. Due to their huge volume, the automatic quality control becomes a crucial aspect over the entire lifetime of the experiment. Here we focus on the design strategy for the Science Ground Segment (SGS) Data Quality Common Tools (DQCT), which has the main role to provide software solutions to gather, evaluate, and record quality information about the raw and derived data products from a primarily scientific perspective. The SGS DQCT will provide a quantitative basis for evaluating the application of reduction and calibration reference data, as well as diagnostic tools for quality parameters, flags, trend analysis diagrams and any other metadata parameter produced by the pipeline. In a large programme like Euclid, it is prohibitively expensive to process large amount of data at the pixel level just for the purpose of quality evaluation. Thus, all measures of quality at the pixel level are implemented in the individual pipeline stages, and passed ...

  15. EPIQR software. [Energy Performance, Indoor Environmental Quality, Retrofit

    Energy Technology Data Exchange (ETDEWEB)

    Flourentzos, F. (Federal Institute of Technology, Lausanne (Switzerland)); Droutsa, K. (National Observatory of Athens, Athens (Greece)); Wittchen, K.B. (Danish Building Research Institute, Hoersholm (Denmark))

    1999-01-01

    The support of the EPIQR method is a multimedia computer program. Several modules help the users of the method to treat the data collected during a diagnosis survey, to set up refurbishment scenario and calculate their cost or energy performance, and finally to visualize the results in a comprehensive way and to prepare quality reports. This article presents the structure and the main features of the software. (au)

  16. Modeling TCP/IP software implementation performance and its application for software routers

    OpenAIRE

    Lepe Aldama, Oscar Iván

    2002-01-01

    Existen numerosos trabajos que estudian o tratan la realización software de los protocolos de comunicaciones para el acceso a la Internet-TCP/IP. Sin embargo, no conocemos ninguno que modele de forma general y precisa la ejecución de este software.La presente tesis aporta una caracterización detallada de la ejecución de la realización software de los mencionados protocolos sobre un computador personal y bajo un sistema operativo UNIX. Esta caracterización muestra cómo varía el rendimiento del...

  17. Implementation of a new picking procedure in the Antelope software

    Science.gov (United States)

    Tiberi, Lara; Costa, Giovanni; Spallarossa, Daniele

    2014-05-01

    quite similar, even if Antelope locates only 38 of the 44 events studied, despite of the 43 of the AutoPicker. So AutoPicker finds more and preciser phases than Antelope both P- (16% more) and mainly S-phases(90% more); but Orbassoc process in Antelope, is able to correctly associate the detections and to find the right location. So in this work we illustrate the implementation of the AutoPicker picking procedure in Antelope software. In the final configuration the AutoPicker is used to rilevate the P- and S- phases, and then these P- and S- onset are associated through the use of the Antelope associator and localizator.

  18. QUALITY SERVICES EVALUATION MODEL BASED ON DEDICATED SOFTWARE TOOL

    Directory of Open Access Journals (Sweden)

    ANDREEA CRISTINA IONICĂ

    2012-10-01

    Full Text Available In this paper we introduced a new model, called Service Quality (SQ, which combines QFD and SERVQUAL methods. This model takes from the SERVQUAL method the five dimensions of requirements and three of characteristics and from the QFD method the application methodology. The originality of the SQ model consists in computing a global index that reflects the customers’ requirements accomplishment level by the quality characteristics. In order to prove the viability of the SQ model, there was developed a software tool that was applied for the evaluation of a health care services provider.

  19. 78 FR 34177 - Implementation of the 2008 National Ambient Air Quality Standards for Ozone: State Implementation...

    Science.gov (United States)

    2013-06-06

    ... the 2008 National Ambient Air Quality Standards for Ozone: State Implementation Plan Requirements... Ambient Air Quality Standards for Ozone: State Implementation Plan Requirements AGENCY: Environmental... 2008 ozone national ambient air quality standards (NAAQS) (the ``2008 ozone NAAQS'') that...

  20. A Software Implementation of a Satellite Interface Message Processor.

    Science.gov (United States)

    Eastwood, Margaret A.; Eastwood, Lester F., Jr.

    A design for network control software for a computer network is described in which some nodes are linked by a communications satellite channel. It is assumed that the network has an ARPANET-like configuration; that is, that specialized processors at each node are responsible for message switching and network control. The purpose of the control…

  1. Fuzzy Based Evaluation of Software Quality Using Quality Models and Goal Models

    Directory of Open Access Journals (Sweden)

    Arfan Mansoor

    2015-09-01

    Full Text Available Software quality requirements are essential part for the success of software development. Defined and guaranteed quality in software development requires identifying, refining, and predicting quality properties by appropriate means. Goal models of goal oriented requirements engineering (GORE and quality models are useful for modelling of functional goals as well as for quality goals. Once the goal models are obtained representing the functional requirements and integrated quality goals, there is need to evaluate each functional requirement arising from functional goals and quality requirement arising from quality goals. The process consist of two main parts. In first part, the goal models are used to evaluate functional goals. The leaf level goals are used to establish the evaluation criteria. Stakeholders are also involved to contribute their opinions about the importance of each goal (functional and/or quality goal. Stakeholder opinions are then converted into quantifiable numbers using triangle fuzzy numbers (TFN. After applying the defuzzification process on TFN, the scores (weights are obtained for each goal. In second part specific quality goals are identified, refined/tailored based on existing quality models and their evaluation is performed similarly using TFN and by applying defuzzification process. The two step process helps to evaluate each goal based on stakeholder opinions and to evaluate the impact of quality requirements. It also helps to evaluate the relationships among functional goals and quality goals. The process is described and applied on ’cyclecomputer’ case study.

  2. Software quality for 1997 - what works and what doesn`t?

    Energy Technology Data Exchange (ETDEWEB)

    Jones, C. [Software Productivity Research, Burlington, MA (United States)

    1997-11-01

    This presentation provides a view of software quality for 1997 - what works and what doesn`t. For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Development (WFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels.

  3. Implementing Extreme Programming in Distributed Software Project Teams: Strategies and Challenges

    Science.gov (United States)

    Maruping, Likoebe M.

    Agile software development methods and distributed forms of organizing teamwork are two team process innovations that are gaining prominence in today's demanding software development environment. Individually, each of these innovations has yielded gains in the practice of software development. Agile methods have enabled software project teams to meet the challenges of an ever turbulent business environment through enhanced flexibility and responsiveness to emergent customer needs. Distributed software project teams have enabled organizations to access highly specialized expertise across geographic locations. Although much progress has been made in understanding how to more effectively manage agile development teams and how to manage distributed software development teams, managers have little guidance on how to leverage these two potent innovations in combination. In this chapter, I outline some of the strategies and challenges associated with implementing agile methods in distributed software project teams. These are discussed in the context of a study of a large-scale software project in the United States that lasted four months.

  4. Delivering successful projects with TSP(SM) and six sigma a practical guide to implementing team software process(SM)

    CERN Document Server

    Jain, Mukesh

    2008-01-01

    Delivering successful projects means the ability to produce high quality software within budget and on time-consistently, but when one mentions quality to software engineers or project managers, they talk about how impossible it is to eliminate defects from software. This assumption is passed on and on until it becomes accepted wisdom, with the power of a self-fulfilling prophecy. And when a project fails to arrive on time or up to standards, team members will turn on each other. The project got delayed because the engineers did a poor job in development or too much was promised upfront for this short of a timeline.In Delivering Successful Projects with TSPSM and Six Sigma: A Practical Guide to Implementing Team Software ProcessSM, you will learn how to effectively manage the development of a software project and deliver it in line with customer expectations. This refreshing volume -Offers real-world case studies about the author's experience at Microsoft successfully implementing TSP to achieve higher qualit...

  5. Semantic Entity-Component State Management Techniques to Enhance Software Quality for Multimodal VR-Systems.

    Science.gov (United States)

    Fischbach, Martin; Wiebusch, Dennis; Latoschik, Marc Erich

    2017-04-01

    Modularity, modifiability, reusability, and API usability are important software qualities that determine the maintainability of software architectures. Virtual, Augmented, and Mixed Reality (VR, AR, MR) systems, modern computer games, as well as interactive human-robot systems often include various dedicated input-, output-, and processing subsystems. These subsystems collectively maintain a real-time simulation of a coherent application state. The resulting interdependencies between individual state representations, mutual state access, overall synchronization, and flow of control implies a conceptual close coupling whereas software quality asks for a decoupling to develop maintainable solutions. This article presents five semantics-based software techniques that address this contradiction: Semantic grounding, code from semantics, grounded actions, semantic queries, and decoupling by semantics. These techniques are applied to extend the well-established entity-component-system (ECS) pattern to overcome some of this pattern's deficits with respect to the implied state access. A walk-through of central implementation aspects of a multimodal (speech and gesture) VR-interface is used to highlight the techniques' benefits. This use-case is chosen as a prototypical example of complex architectures with multiple interacting subsystems found in many VR, AR and MR architectures. Finally, implementation hints are given, lessons learned regarding maintainability pointed-out, and performance implications discussed.

  6. EARLY IDENTIFICATION OF SOFTWARE DEFECTS USING OCL PREDICATES TO IMPROVE SOFTWARE QUALITY

    Directory of Open Access Journals (Sweden)

    A. JALILA

    2015-03-01

    Full Text Available Formal specification-based testing has been used widely to assess potential faults or prove their absence in a given system at the earliest. This research work has proposed an automated fault-based testing framework to test the specification of the system using Object Constraint Language (OCL. Accordingly, the possible faults in OCL specification of the system has been anticipated by mutating its method based on OCL predicate-based fault classes. Then, test cases are generated using Genetic Algorithm with simulated annealing technique. In this paper, a novel OCL-predicate based fitness function is defined to evaluate the generated test data. Finally, this paper presents the experimental results, which indicate that the proposed methodology provides more test coverage with the reduced test suite and test run. This results in cost-effective software development so as to improve software quality.

  7. Implementation of wireless communication based on Software Defined Radio

    OpenAIRE

    Zhang, Lei

    2013-01-01

    In current communication systems, there are many new challenges like various competitive standards, the scarcity of frequency resource, etc., especially the development of personal wireless communication systems result the new system update faster than ever before, the conventional hardware-based wireless communication system is difficult to adapt to this situation. The emergence of SDR enabled the third revolution of wireless communication which from hardware to software and build a flexible...

  8. Implementation of Pulse Radar Waveform Based on Software Radio Platform

    OpenAIRE

    Wang Dong; Dong Jian; Xiao Shunping

    2015-01-01

    Based on the frequency and phase modulated signal, the authors design some commonly-used pulse radar baseband waveform, such as linear frequency modulated waveform, nonlinear frequency modulated waveform, Costas waveform, Barker coding waveform and multi-phase coded waveform, and the authors compare their performance, such as the peak side lobe ratio, the Rayleigh resolution in time and distance resolution. Then, based on the software radio platform NI PXIe-5644R, the authors design the timin...

  9. The aerospace energy systems laboratory: Hardware and software implementation

    Science.gov (United States)

    Glover, Richard D.; Oneil-Rood, Nora

    1989-01-01

    For many years NASA Ames Research Center, Dryden Flight Research Facility has employed automation in the servicing of flight critical aircraft batteries. Recently a major upgrade to Dryden's computerized Battery Systems Laboratory was initiated to incorporate distributed processing and a centralized database. The new facility, called the Aerospace Energy Systems Laboratory (AESL), is being mechanized with iAPX86 and iAPX286 hardware running iRMX86. The hardware configuration and software structure for the AESL are described.

  10. NARAC SOFTWARE QUALITY ASSURANCE: ADAPTING FORMALISM TO MEET VARYING NEEDS

    Energy Technology Data Exchange (ETDEWEB)

    Walker, H; Nasstrom, J S; Homann, S G

    2007-11-20

    The National Atmospheric Release Advisory Center (NARAC) provides tools and services that predict and map the spread of hazardous material accidentally or intentionally released into the atmosphere. NARAC is a full function system that can meet a wide range of needs with a particular focus on emergency response. The NARAC system relies on computer software in the form of models of the atmosphere and related physical processes supported by a framework for data acquisition and management, user interface, visualization, communications and security. All aspects of the program's operations and research efforts are predicated to varying degrees on the reliable and correct performance of this software. Consequently, software quality assurance (SQA) is an essential component of the NARAC program. The NARAC models and system span different levels of sophistication, fidelity and complexity. These different levels require related but different approaches to SQA. To illustrate this, two different levels of software complexity are considered in this paper. As a relatively simple example, the SQA procedures that are being used for HotSpot, a straight-line Gaussian model focused on radiological releases, are described. At the other extreme, the SQA issues that must be considered and balanced for the more complex NARAC system are reviewed.

  11. Design, Implementation, and Performance of CREAM Data Acquisition Software

    CERN Document Server

    Zinn, S Y; Bagliesi, M G; Beatty, J J; Childers, J T; Coutu, S; Duvernois, M A; Ganel, O; Kim, H J; Lee, M H; Lutz, L; Malinine, A; Maestro, P; Marrocchesi, P S; Park, I H; Seo, E S; Song, C; Swordy, S; Wu, J

    2005-01-01

    Cosmic Ray Energetics and Mass (CREAM) is a balloon-borne experiment scheduled for launching from Antarctica in late 2004. Its aim is to measure the energy spectrum and composition of cosmic rays from proton to iron nuclei at ultra high energies from 1 to 1,000 TeV. Ultra long duration balloons are expected to fly about 100 days. One special feature of the CREAM data acquisition software (CDAQ) is the telemetric operation of the instrument using satellites. During a flight the science event and housekeeping data are sent from the instrument to a ground facility. Likewise, commands for controlling both the hardware and the software are uploaded from the ground facility. This requires a robust, reliable, and fast software system. CDAQ has been developed and tested during three beam tests at CERN in July, September, and November 2003. Recently the interfaces to the transition radiation detector (TRD) and to the timing-based charge detector (TCD) have been added. These new additions to CDAQ will be checked at a t...

  12. Quality-driven multi-objective optimization of software architecture design : method, tool, and application

    NARCIS (Netherlands)

    Etemadi Idgahi (Etemaadi), Ramin

    2014-01-01

    Software architecting is a non-trivial and demanding task for software engineers to perform. The architecture is a key enabler for software systems. Besides being crucial for user functionality, the software architecture has deep impact on software qualities such as performance, safety, and cost.

  13. Quality-driven multi-objective optimization of software architecture design : method, tool, and application

    NARCIS (Netherlands)

    Etemadi Idgahi (Etemaadi), Ramin

    2014-01-01

    Software architecting is a non-trivial and demanding task for software engineers to perform. The architecture is a key enabler for software systems. Besides being crucial for user functionality, the software architecture has deep impact on software qualities such as performance, safety, and cost.

  14. E-learning Materials Development: Applying and Implementing Software Reuse Principles and Granularity Levels in the Small

    OpenAIRE

    Nabil Arman,

    2010-01-01

    E-learning materials development is typically acknowledged as an expensive, complicated, and lengthy process, often producing materials that are of low quality and difficult to adaptand maintain. It has always been a challenge to identify proper e-learning materials that can be reused at a reasonable cost and effort. In this paper, software engineering reuse principlesare applied to e-learning materials development process. These principles are then applied and implemented in a prototype that...

  15. Global Software Engineering: A Software Process Approach

    Science.gov (United States)

    Richardson, Ita; Casey, Valentine; Burton, John; McCaffery, Fergal

    Our research has shown that many companies are struggling with the successful implementation of global software engineering, due to temporal, cultural and geographical distance, which causes a range of factors to come into play. For example, cultural, project managementproject management and communication difficulties continually cause problems for software engineers and project managers. While the implementation of efficient software processes can be used to improve the quality of the software product, published software process models do not cater explicitly for the recent growth in global software engineering. Our thesis is that global software engineering factors should be included in software process models to ensure their continued usefulness in global organisations. Based on extensive global software engineering research, we have developed a software process, Global Teaming, which includes specific practices and sub-practices. The purpose is to ensure that requirements for successful global software engineering are stipulated so that organisations can ensure successful implementation of global software engineering.

  16. Software for improving the quality of project management, a case study: international manufacture of electrical equipment

    Science.gov (United States)

    Preradović, D. M.; Mićić, Lj S.; Barz, C.

    2017-05-01

    Production conditions in today’s world require software support at every stage of production and development of new products, for quality assurance and compliance with ISO standards. In addition to ISO standards such as usual metrics of quality, companies today are focused on other optional standards, such as CMMI (Capability Maturity Model Integrated) or prescribing they own standards. However, while there is intensive progress being made in the PM (project management), there is still a significant number of projects, at the global level, that are failures. These have failed to achieve their goals, within budget or timeframe. This paper focuses on checking the role of software tools through the rate of success in projects implemented in the case of internationally manufactured electrical equipment. The results of this research show the level of contribution of the project management software used to manage and develop new products to improve PM processes and PM functions, and how selection of the software tools affects the quality of PM processes and successfully completed projects.

  17. Ada Implementation Guide. Software Engineering With Ada. Volume 1

    Science.gov (United States)

    1994-04-01

    teaching, the student is less likely to readily adopt new, more powerful ways of accomplishing old tasks 122 Depatn of the NaY I ! Trablng and Educaion and...Maturity Model3 (CMU/SEI-92-TR-25, ESC-TR-/92-0M5). Pittsburgh, PA : Carnegie-Mellon University, 1992. SBoehm. B.W. Software Engineering Economics...Pittsburgh, PA : Carnegie-Mellon University, 19-21 March 1991. £ Contrast: Ada 9X and C++, Schonberg, E. New York University, 1992 (Distributed by Ada IC on

  18. Maximize Your Investment 10 Key Strategies for Effective Packaged Software Implementations

    CERN Document Server

    Beaubouef, Grady Brett

    2009-01-01

    This is a handbook covering ten principles for packaged software implementations that project managers, business owners, and IT developers should pay attention to. The book also has practical real-world coverage including a sample agenda for conducting business solution modeling, customer case studies, and a road map to implement guiding principles. This book is aimed at enterprise architects, development leads, project managers, business systems analysts, business systems owners, and anyone who wants to implement packaged software effectively. If you are a customer looking to implement COTS s

  19. Re-constructible CMM software system modeling and its implementation

    Science.gov (United States)

    Bai, Y. W.; Wei, S. Y.; Yang, X. H.; Liu, S. P.

    2008-12-01

    This paper presents a novel way for the re-constructible CMM software system modeling by taking advantage of a tiered modeling strategy. It consists of four tiers: (1) the bottom layer is the CAD model manager which encapsulates geometric engine and 3D object displaying engine as a COM; (2) the middle is the kernel components which is designed to manage the objects of geometric entity, coordinate system, probe and the system environment parameters etc; (3) the third layer is function modules layer that is used to manage and handle the messages and events of the windows/dialog, menus and toolbars; (4) the top layer is GUI module that is designed to initialize the application with the resource of GUI with the manner of dynamic loading. A set of commercial CMM software, Direct DMIS, has applied the method in a R&D of China National Institute of Measuring and Test Technology (NIMTT). It proves that the developed system can effectively integrate the modules distributed in different layers developed with C++ or C# and the proposed method is feasible.

  20. Development and case study of a science-based software platform to support policy making on air quality

    Institute of Scientific and Technical Information of China (English)

    Yun Zhu; Yahweh Lao; Carey Jang; Chen-Jen Lin; Jia Xing; Shuxiao Wang; Joshua S.Fu

    2015-01-01

    This article describes the development and implementations of a novel software platform that supports real-time,science-based policy making on air quality through a user-friendly interface.The software,RSM-VAT,uses a response surface modeling (RSM) methodology and serves as a visualization and analysis tool (VAT) for three-dimensional air quality data obtained by atmospheric models.The software features a number of powerful and intuitive data visualization functions for illustrating the complex nonlinear relationship between emission reductions and air quality benefits.The case study of contiguous U.S.demonstrates that the enhanced RSM-VAT is capable of reproducing the air quality model results with Normalized Mean Bias <2% and assisting in air quality policy making in near real time.

  1. Implementing an composition architecture for an online game software

    Directory of Open Access Journals (Sweden)

    Jaime Alberto Guzmán Luna

    2011-01-01

    Full Text Available Este artículo presenta una arquitectura para la composición de servicios Web, en la que un plan de composición se construye sobre la base de un agente de planificación, que puede ser ejecutado en forma concurrente durante su composición a fin de estimar la siguiente acción a ejecutar en lugar de preparar un completo plan que con frecuencia será invalidado. Esta característica es muy valiosa cuando se trata de abordar los problemas en tiempo real. Específicamente se propone como un dominio de prueba un software de juegos en línea, llamado ENVIRO.

  2. Considerations on the quality of medical software and information services.

    Science.gov (United States)

    Forsström, J J; Rigby, M

    1999-12-01

    Fast developments in information and communication technology have made it possible to develop new services for citizens. One of the most interesting areas is health care. Medical knowledge is usually valid all over the world that makes the market global. Information services and decision support software are becoming important tools for medical professionals but also ordinary citizens are interested in health related information. It has been estimated that by the year 2010 the turnover of health care telematics industry may be close to that of the drug industry today. The nature of this global information industry is very different from any industry in history. Since there are no frontiers, no clear products and no shops in the information market, it is difficult to develop any effective legislation. However, the history of medicine has shown that health care sector cannot be free from regulation without risking citizens' health. The huge commercial potential of the Internet has already been used to promote products and services that have no proven effect on health and that may sometimes be even dangerous. In this paper we discuss the needs and possibilities to assess the quality of medical decision support software and information services. For brevity the terms medical software and medical knowledge are used, but the issues also relate to informatics systems used by any health professional, and to computerised systems used to schedule care or to organise record systems.

  3. Improving the Quality of Published Chemical Names with Nomenclature Software

    Directory of Open Access Journals (Sweden)

    Gernot A. Eller

    2006-11-01

    Full Text Available This work deals with the use of organic systematic nomenclature in scientific literature, its quality, and computerized methods for its improvement. Criteria for classification of systematic names in terms of quality/correctness are discussed and applied to a sample set of several hundred names extracted from the literature. The same structures are named with three popular state-of-the-art nomenclature programs – AutoNom 2000, ChemDraw 10.0, and ACD/Name 9.0. When comparing the results, all nomenclature tools show a significantly better performance than 'average chemists'. One program allows the generation not only of IUPAC names but also of CAS-like index names that are compared with the officially registered names. The scope and limitations of nomenclature software are discussed and a comparison of the programs' actual capabilities is given.

  4. Improving Software Quality Prediction by Noise Filtering Techniques

    Institute of Scientific and Technical Information of China (English)

    Taghi M. Khoshgoftaar; Pierre Rebours

    2007-01-01

    Accuracy of machine learners is affected by quality of the data the learners are induced on. In this paper,quality of the training dataset is improved by removing instances detected as noisy by the Partitioning Filter. The fit datasetis first split into subsets, and different base learners are induced on each of these splits. The predictions are combined insuch a way that an instance is identified as noisy if it is misclassified by a certain number of base learners. Two versionsof the Partitioning Filter are used: Multiple-Partitioning Filter and Iterative-Partitioning Filter. The number of instancesremoved by the filters is tuned by the voting scheme of the filter and the number of iterations. The primary aim of thisstudy is to compare the predictive performances of the final models built on the filtered and the un-filtered training datasets.A case study of software measurement data of a high assurance software project is performed. It is shown that predictiveperformances of models built on the filtered fit datasets and evaluated on a noisy test dataset are generally better than thosebuilt on the noisy (un-filtered) fit dataset. However, predictive performance based on certain aggressive filters is affected bypresence of noise in the evaluation dataset.

  5. Analysis of quality raw data of second generation sequencers with Quality Assessment Software

    Directory of Open Access Journals (Sweden)

    Schneider Maria PC

    2011-04-01

    Full Text Available Abstract Background Second generation technologies have advantages over Sanger; however, they have resulted in new challenges for the genome construction process, especially because of the small size of the reads, despite the high degree of coverage. Independent of the program chosen for the construction process, DNA sequences are superimposed, based on identity, to extend the reads, generating contigs; mismatches indicate a lack of homology and are not included. This process improves our confidence in the sequences that are generated. Findings We developed Quality Assessment Software, with which one can review graphs showing the distribution of quality values from the sequencing reads. This software allow us to adopt more stringent quality standards for sequence data, based on quality-graph analysis and estimated coverage after applying the quality filter, providing acceptable sequence coverage for genome construction from short reads. Conclusions Quality filtering is a fundamental step in the process of constructing genomes, as it reduces the frequency of incorrect alignments that are caused by measuring errors, which can occur during the construction process due to the size of the reads, provoking misassemblies. Application of quality filters to sequence data, using the software Quality Assessment, along with graphing analyses, provided greater precision in the definition of cutoff parameters, which increased the accuracy of genome construction.

  6. A genetic algorithm for structure-activity relationships: software implementation

    CERN Document Server

    Jantschi, Lorentz

    2009-01-01

    The design and the implementation of a genetic algorithm are described. The applicability domain is on structure-activity relationships expressed as multiple linear regressions and predictor variables are from families of structure-based molecular descriptors. An experiment to compare different selection and survival strategies was designed and realized. The genetic algorithm was run using the designed experiment on a set of 206 polychlorinated biphenyls searching on structure-activity relationships having known the measured octanol-water partition coefficients and a family of molecular descriptors. The experiment shows that different selection and survival strategies create different partitions on the entire population of all possible genotypes.

  7. Development and implementation of a 'Mental Health Finder' software tool within an electronic medical record system.

    Science.gov (United States)

    Swan, D; Hannigan, A; Higgins, S; McDonnell, R; Meagher, D; Cullen, W

    2017-02-01

    In Ireland, as in many other healthcare systems, mental health service provision is being reconfigured with a move toward more care in the community, and particularly primary care. Recording and surveillance systems for mental health information and activities in primary care are needed for service planning and quality improvement. We describe the development and initial implementation of a software tool ('mental health finder') within a widely used primary care electronic medical record system (EMR) in Ireland to enable large-scale data collection on the epidemiology and management of mental health and substance use problems among patients attending general practice. In collaboration with the Irish Primary Care Research Network (IPCRN), we developed the 'Mental Health Finder' as a software plug-in to a commonly used primary care EMR system to facilitate data collection on mental health diagnoses and pharmacological treatments among patients. The finder searches for and identifies patients based on diagnostic coding and/or prescribed medicines. It was initially implemented among a convenience sample of six GP practices. Prevalence of mental health and substance use problems across the six practices, as identified by the finder, was 9.4% (range 6.9-12.7%). 61.9% of identified patients were female; 25.8% were private patients. One-third (33.4%) of identified patients were prescribed more than one class of psychotropic medication. Of the patients identified by the finder, 89.9% were identifiable via prescribing data, 23.7% via diagnostic coding. The finder is a feasible and promising methodology for large-scale data collection on mental health problems in primary care.

  8. 78 FR 34964 - Implementation of the 2008 National Ambient Air Quality Standards for Ozone: State Implementation...

    Science.gov (United States)

    2013-06-11

    ... AGENCY 40 CFR Parts 50, 51, 70 and 71 Implementation of the 2008 National Ambient Air Quality Standards... proposed rule ``Implementation of the 2008 National Ambient Air Quality Standards for Ozone: State... proposed rulemaking proposes to implement the 2008 ozone national ambient air quality standards...

  9. 77 FR 73320 - Approval of Air Quality Implementation Plans; California; South Coast Air Quality Management...

    Science.gov (United States)

    2012-12-10

    ... Quality Management District (SCAQMD or District) portion of the California State Implementation Plan (SIP... Air Quality Management District regarding specific implementation of parts of the Prevention of Significant Deterioration program. (i) Incorporation by reference. (A) South Coast Air Quality...

  10. Contribuição dos modelos de qualidade e maturidade na melhoria dos processos de software Contribution of quality and maturity models to software process improvement

    Directory of Open Access Journals (Sweden)

    Antonio Carlos Tonini

    2008-01-01

    Full Text Available Grande parte das empresas desenvolvedoras de software criou seu próprio processo de trabalho. Devido à rápida expansão do mercado de software, a concorrência ocorre muito mais em custo do que em diferenciação. Para obter vantagem competitiva, as empresas devem atualizar-se continuamente na tecnologia, buscar a maturidade nos processos e eliminar a ineficiência operacional. Isso requer um envolvimento das pessoas, dos processos e da organização como um todo. O artigo discute a implementação de melhorias nos processos de software segundo os principais modelos de qualidade e de maturidade. Com base em um Estudo de Casos Múltiplos, verifica-se que a melhoria dos processos de software requer que a melhoria ocorra primeiramente entre cada um dos desenvolvedores e, a seguir, envolva os grupos de desenvolvimento e por fim, a organização como um todo. A pesquisa conclui que os modelos de qualidade e maturidade servem como orientadores do processo de melhoria.Many software development companies have developed their own work method. Due to the fast software market growth, the competition focuses more on cost than on differentiation. To achieve competitive advantage, software developer organizations must continually update their technology, reach high level process maturity and eliminate all the operational inefficiency. These procedures involve people, processes and the whole organization. The aim of the paper is to discuss software process improvement implementation according to the most important quality and maturity models. Based on a Multiple Case Study, it is verified that the software process improvement needs firstly individual improvement and, later, it involves the developer teams and the whole organization. The research concludes that the quality and maturity models must be used as improvement process drivers.

  11. Development and analysis of the Software Implemented Fault-Tolerance (SIFT) computer

    Science.gov (United States)

    Goldberg, J.; Kautz, W. H.; Melliar-Smith, P. M.; Green, M. W.; Levitt, K. N.; Schwartz, R. L.; Weinstock, C. B.

    1984-01-01

    SIFT (Software Implemented Fault Tolerance) is an experimental, fault-tolerant computer system designed to meet the extreme reliability requirements for safety-critical functions in advanced aircraft. Errors are masked by performing a majority voting operation over the results of identical computations, and faulty processors are removed from service by reassigning computations to the nonfaulty processors. This scheme has been implemented in a special architecture using a set of standard Bendix BDX930 processors, augmented by a special asynchronous-broadcast communication interface that provides direct, processor to processor communication among all processors. Fault isolation is accomplished in hardware; all other fault-tolerance functions, together with scheduling and synchronization are implemented exclusively by executive system software. The system reliability is predicted by a Markov model. Mathematical consistency of the system software with respect to the reliability model has been partially verified, using recently developed tools for machine-aided proof of program correctness.

  12. Software Quality Evaluation Models Applicable in Health Information and Communications Technologies. A Review of the Literature.

    Science.gov (United States)

    Villamor Ordozgoiti, Alberto; Delgado Hito, Pilar; Guix Comellas, Eva María; Fernandez Sanchez, Carlos Manuel; Garcia Hernandez, Milagros; Lluch Canut, Teresa

    2016-01-01

    Information and Communications Technologies in healthcare has increased the need to consider quality criteria through standardised processes. The aim of this study was to analyse the software quality evaluation models applicable to healthcare from the perspective of ICT-purchasers. Through a systematic literature review with the keywords software, product, quality, evaluation and health, we selected and analysed 20 original research papers published from 2005-2016 in health science and technology databases. The results showed four main topics: non-ISO models, software quality evaluation models based on ISO/IEC standards, studies analysing software quality evaluation models, and studies analysing ISO standards for software quality evaluation. The models provide cost-efficiency criteria for specific software, and improve use outcomes. The ISO/IEC25000 standard is shown as the most suitable for evaluating the quality of ICTs for healthcare use from the perspective of institutional acquisition.

  13. Quality assessment with the AGIR software results and experience; Qualitaetsmanagement mittels AGIR: Ergebnisse und Erfahrungen

    Energy Technology Data Exchange (ETDEWEB)

    Rauch, D.; Kotter, E.; Kurtz, C.; Schaefer, O.; Ehritt-Braun, C.; Burger, D.; Schaper, J.; Uhrmeister, P. [Freiburg Univ. (Germany). Radiologische Universitaetsklinik, Abt. fuer Roentgendiagnostik

    2001-02-01

    Purpose: To evaluate whether a new software from the working group for interventional radiology (AGIR) is an appropriate tool for quality assurance in interventional radiology, and presentation of results acquired within the quality improvement process in 1999. Patients and methods: AGIR-defined parameters such as patient data, risk profile, given interventions as well as complications were registered by a recently developed software. Based on monthly data analyses, possible complications were identified and discussed in morbidity and mortality conferences. Results: 1014 interventions were performed in our institution in 1999. According to criteria established by AGIR, the complication rate was 2.7%. In addition and according to SCVIR criteria, complications were distinguished quantitatively in five classes and semiquantitatively in minor and major groups. The result was a minor complication rate of 1.8%, and a major rate of 0.9%. There were no cases of death associated with the intervention. Further strategies were developed in order to reduce the complication rate. Conclusion: Extensive quality assurance methods can be integrated in daily routine work. These methods lead to an intensive transparency of treatment results, and allow the implementation of continuous quality improvements. The development of the software is a first step in establishing a nation-wide quality assurance system. Nevertheless, modification and additional definition of the AGIR predefined parameters are required, for example, to avoid unnecessary procedures. (orig.) [German] Ziel: Evaluierung einer neuen Software der arbeitsgemeinschaft interventionelle radiologie (AGIR) hinsichtlich der Eignung als Instrument zur Qualitaetssicherung in der interventionellen Radiologie sowie Praesentation der Ergebnisse innerhalb des Qualitaetsprozesses aus dem Jahre 1999. Patienten und Methoden: Mit Hilfe einer neuen Software wurden AGIR definierte Parameter wie Patientendaten, Risikoprofil, die jeweilige

  14. SOFTWARE TRAINING AIDS DEVELOPMENT AND IMPLEMENTATION IN PROFESSIONAL PREPARATION PRACTICE OF TECHNOLOGICAL EDUCATION TEACHER

    Directory of Open Access Journals (Sweden)

    Anatoliy G. Gritchenko

    2013-03-01

    Full Text Available The article outlines the theoretical and practical aspects of software training aids development and implementation in professional preparation practice of technological education teacher. The myriad opportunities of new information technologies are described; the characteristic features of modern software training tool (STT are revealed; the main algorithmic structure circuits of training programs construction (linear, cyclic, with hyperlinks, to the labels, which enable the development of STT variety and functionality are given; the methodology of STT creating is described based on the analysis of the technology teacher preparation in HEE content, MITE didactic functions and selection criteria of educational software for this area of specialist’s preparation.

  15. Total Quality Management Implementation Plan Defense Depot Memphis

    Science.gov (United States)

    1989-07-01

    Total Quality Management Implementation...improvement goals, implementation strategy and milestones. 6’ SEP 291989 /; ELECTE i= E 14. SUBJECT TERMS 15. NUMBER OF PAGES TQM ( Total Quality Management ), Depot...changing work environment where change is the norm. We are talking about changes in attitudes and habits. Total Quality Management is not a

  16. Aspects regarding implementation of quality management in health system

    OpenAIRE

    Mãdãlina ALBU

    2013-01-01

    Services offered by Romanian health system must be managed both quantitatively and qualitatively. This can be initiated and maintained by implementing quality management systems to ensure the provision of quality health services. Nationally, a number of hospitals are certified under ISO 9001. This paper presents how it was implemented quality management system in Emergency Hospital of Ploieºti.

  17. Software implementation of floating-Point arithmetic on a reduced-Instruction-set processor

    Energy Technology Data Exchange (ETDEWEB)

    Gross, T.

    1985-11-01

    Current single chip implementations of reduced-instruction-set processors do not support hardware floating-point operations. Instead, floating-point operations have to be provided either by a coprocessor or by software. This paper discusses issues arising from a software implementation of floating-point arithmetic for the MIPS processor, an experimental VLSI architecture. Measurements indicate that an acceptable level of performance is achieved, but this approach is no substitute for a hardware accelerator if higher-precision results are required. This paper includes instruction profiles for the basic floating-point operations and evaluates the usefulness of some aspects of the instruction set.

  18. The Software Improvement Process - Tools And Rules To Encourage Quality

    CERN Document Server

    Sigerud, K

    2011-01-01

    The Applications section of the CERN accelerator Controls group has decided to apply a systematic approach to quality assurance (QA), the “Software Improvement Process”, SIP. This process focuses on three areas: the development process itself, suitable QA tools, and how to practically encourage developers to do QA. For each stage of the development process we have agreed on the recommended activities and deliverables, and identified tools to automate and support the task. For example we do more code reviews. As peer reviews are resource-intensive, we only do them for complex parts of a product. As a complement, we are using static code checking tools, like FindBugs and Checkstyle. We also encourage unit testing and have agreed on a minimum level of test coverage recommended for all products, measured using Clover. Each of these tools is well integrated with our IDE (Eclipse) and give instant feedback to the developer about the quality of their code. The major challenges of SIP have been to 1) agree on com...

  19. Implementation of the Quality Management System in the telecommunications companies

    Directory of Open Access Journals (Sweden)

    Menshikova Ekaterina

    2016-01-01

    Full Text Available This article describes the advantages of the implementing of the quality management system based on ISO 9001 in the telecommunications companies. The stages of the implementation of the quality management system in the telecommunications companies are discussed. The map of the main process and criteria of processes is developed. The expediency of implementing the quality management system based on the requirements of ISO 9001 in Telecommunications is showed.

  20. Implementation of the Quality Management System in the telecommunications companies

    OpenAIRE

    Menshikova Ekaterina; Khazanov Oleg; Styazhkin Mark

    2016-01-01

    This article describes the advantages of the implementing of the quality management system based on ISO 9001 in the telecommunications companies. The stages of the implementation of the quality management system in the telecommunications companies are discussed. The map of the main process and criteria of processes is developed. The expediency of implementing the quality management system based on the requirements of ISO 9001 in Telecommunications is showed.

  1. A BEHAVIOR-PRESERVING TRANSLATION FROM FBD DESIGN TO C IMPLEMENTATION FOR REACTOR PROTECTION SYSTEM SOFTWARE

    Directory of Open Access Journals (Sweden)

    JUNBEOM YOO

    2013-08-01

    Full Text Available Software safety for nuclear reactor protection systems (RPSs is the most important requirement for the obtainment of permission for operation and export from government authorities, which is why it should be managed with well-experienced software development processes. The RPS software is typically modeled with function block diagrams (FBDs in the design phase, and then mechanically translated into C programs in the implementation phase, which is finally compiled into executable machine codes and loaded on RPS hardware – PLC (Programmable Logic Controller. Whereas C Compilers are fully-verified COTS (Commercial Off-The-Shelf software, translators from FBDs to C programs are provided by PLC vendors. Long-term experience, experiments and simulations have validated their correctness and function safety. This paper proposes a behavior-preserving translation from FBD design to C implementation for RPS software. It includes two sets of translation algorithms and rules as well as a prototype translator. We used an example of RPS software in a Korean nuclear power plant to demonstrate the correctness and effectiveness of the proposed translation.

  2. A Behavior-Preserving Translation From FBD Design to C Implementation for Reactor Protection System Software

    Energy Technology Data Exchange (ETDEWEB)

    Yoo, Junbeom; Kim, Euisub [Konkuk Univ., Seoul (Korea, Republic of); Lee, Jangsoo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2013-08-15

    Software safety for nuclear reactor protection systems (RPSs) is the most important requirement for the obtainment of permission for operation and export from government authorities, which is why it should be managed with well-experienced software development processes. The RPS software is typically modeled with function block diagrams (FBDs) in the design phase, and then mechanically translated into C programs in the implementation phase, which is finally compiled into executable machine codes and loaded on RPS hardware - PLC (Programmable Logic Controller). Whereas C Compilers are fully-verified COTS (Commercial Off-The-Shelf) software, translators from FBDs to C programs are provided by PLC vendors. Long-term experience, experiments and simulations have validated their correctness and function safety. This paper proposes a behavior-preserving translation from FBD design to C implementation for RPS software. It includes two sets of translation algorithms and rules as well as a prototype translator. We used an example of RPS software in a Korean nuclear power plant to demonstrate the correctness and effectiveness of the proposed translation.

  3. Integrated software environment dedicated for implementation of control systems based on PLC controllers

    Directory of Open Access Journals (Sweden)

    Szymon SURMA

    2007-01-01

    Full Text Available Industrial processes’ control systems based on PLC controllers play today a very important role in all fields of transport, including also sea transport. Construction of control systems is the field of engineering, which has been continuously evolving towards maximum simplification of system design path. Up to now the time needed forthe system construction from the design to commissioning had to be divided into a few stages. A mistake made in an earlier stage caused that in most cases the next stages had to be restarted. Available debugging systems allows defect detection at an early stage of theproject implementation. The paper presents general characteristic of integrated software for implementation of complex control systems. The issues related to the software use for programming of the visualisation environment, control computer, selection oftransmission medium and transmission protocol as well as PLC controllers’ configuration, software and control have been analysed.

  4. Implementation of LTE SC-FDMA on the USRP2 Software Defined Radio Platform

    DEFF Research Database (Denmark)

    Jørgensen, Peter Bjørn; Hansen, Thomas Lundgaard; Sørensen, Troels Bundgaard;

    2011-01-01

    In this paper we discuss the implementation of a Single Carrier Frequency Division Multiple Access (SC-FDMA) transceiver running over the Universal Software Radio Peripheral 2 (USRP2). SC-FDMA is the air interface which has been selected for the uplink in the latest Long Term Evolution (LTE...

  5. Portal for Families Overcoming Neurodevelopmental Disorders (PFOND): Implementation of a Software Framework for Facilitated Community Website Creation by Nontechnical Volunteers.

    Science.gov (United States)

    Ye, Xin Cynthia; Ng, Isaiah; Seid-Karbasi, Puya; Imam, Tuhina; Lee, Cheryl E; Chen, Shirley Yu; Herman, Adam; Sharma, Balraj; Johal, Gurinder; Gu, Bobby; Wasserman, Wyeth W

    2013-08-06

    The Portal for Families Overcoming Neurodevelopmental Disorders (PFOND) provides a structured Internet interface for the sharing of information with individuals struggling with the consequences of rare developmental disorders. Large disease-impacted communities can support fundraising organizations that disseminate Web-based information through elegant websites run by professional staff. Such quality resources for families challenged by rare disorders are infrequently produced and, when available, are often dependent upon the continued efforts of a single individual. The project endeavors to create an intuitive Web-based software system that allows a volunteer with limited technical computer skills to produce a useful rare disease website in a short time period. Such a system should provide access to emerging news and research findings, facilitate community participation, present summary information about the disorder, and allow for transient management by volunteers who are likely to change periodically. The prototype portal was implemented using the WordPress software system with both existing and customized supplementary plug-in software modules. Gamification scoring features were implemented in a module, allowing editors to measure progress. The system was installed on a Linux-based computer server, accessible across the Internet through standard Web browsers. A prototype PFOND system was implemented and tested. The prototype system features a structured organization with distinct partitions for background information, recent publications, and community discussions. The software design allows volunteer editors to create a themed website, implement a limited set of topic pages, and connect the software to dynamic RSS feeds providing information about recent news or advances. The prototype was assessed by a fraction of the disease sites developed (8 out of 27), including Aarskog-Scott syndrome, Aniridia, Adams-Oliver syndrome, Cat Eye syndrome, Kabuki syndrome

  6. Software Simulation in GSM Environment and Hardware Implementation of Improved Multi-band Excitation Vocoder

    Institute of Scientific and Technical Information of China (English)

    1998-01-01

    The algorithm of the Improved Multi-Band Excitation (IMBE) vocoder is thoroughly studied, designed and implemented including software implementation on PC/DOS, SUN/UNIX workstation system and hardware real-time implementation on TMS320C31 DSP. In order to explore the performance of IMBE vocoder in GSM environment, a GSM radio interface software simulation platform is built and a series of tests are run on four languages (Chinese, English, German, Swedish) and different channel models (urban, hilly and rural areas) with different SNR. Finally simulation result is analyzed which is useful for the performance analysis of IMBE and the application of vocoders with bit rate of 4kbps order in GSM environment.

  7. A Study on a Software Quality Assurance of a Process and a Product

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Hyun Tae [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    2006-07-01

    Since nuclear plants and facilities have made increasing use of digital technology, the safety and reliability of software is a primary concern. Software errors are more difficult to detect and handle than hardware-related failures. It is crucial to consider the a process and a product of a software life cycle to increase the quality of a software. This paper discusses the quality assurance of a process and a product of a software life cycle based on two prominent standards, ISO 9001:2000 and CMMI.

  8. Understanding the State of Quality of Software on the basis of Time Gap, Quality Gap and Difference with Standard Model

    Directory of Open Access Journals (Sweden)

    Ekbal Rashid

    2013-06-01

    Full Text Available This paper tries to introduce a new mathematical model to understand the state of quality of software by calculating parameters such as the time gap and quality gap with relation to some predefinedstandard software quality or in relation to some chalked out software quality plan. The paper also suggests methods to calculate the difference in quality of the software being developed and the modelsoftware which has been decided upon as the criteria for comparison. These methods can be employed to better understand the state of quality as compared to other standards. In order to obtain the graphical representation of data we have used Microsoft office 2007 graphical chart. Which facilitate easy simulation of time and quality gap.

  9. Software engineering and Ada (Trademark) training: An implementation model for NASA

    Science.gov (United States)

    Legrand, Sue; Freedman, Glenn

    1988-01-01

    The choice of Ada for software engineering for projects such as the Space Station has resulted in government and industrial groups considering training programs that help workers become familiar with both a software culture and the intricacies of a new computer language. The questions of how much time it takes to learn software engineering with Ada, how much an organization should invest in such training, and how the training should be structured are considered. Software engineering is an emerging, dynamic discipline. It is defined by the author as the establishment and application of sound engineering environments, tools, methods, models, principles, and concepts combined with appropriate standards, guidelines, and practices to support computing which is correct, modifiable, reliable and safe, efficient, and understandable throughout the life cycle of the application. Neither the training programs needed, nor the content of such programs, have been well established. This study addresses the requirements for training for NASA personnel and recommends an implementation plan. A curriculum and a means of delivery are recommended. It is further suggested that a knowledgeable programmer may be able to learn Ada in 5 days, but that it takes 6 to 9 months to evolve into a software engineer who uses the language correctly and effectively. The curriculum and implementation plan can be adapted for each NASA Center according to the needs dictated by each project.

  10. 78 FR 53270 - Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2013-08-29

    ... AGENCY 40 CFR Part 52 Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality Management District; Stationary Source Permits AGENCY: Environmental Protection Agency... permitting rules submitted by California as a revision to the Sacramento Metropolitan Air Quality Management...

  11. 78 FR 10589 - Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2013-02-14

    ... AGENCY 40 CFR Part 52 Revision of Air Quality Implementation Plan; California; Sacramento Metropolitan Air Quality Management District; Stationary Source Permits AGENCY: Environmental Protection Agency... by California as a revision to the Sacramento Metropolitan Air Quality Management District (SMAQMD or...

  12. DLA-X Total Quality Management (TQM) Implementation Plan

    Science.gov (United States)

    1989-07-01

    PAGES TOM ( Total Quality Management ), Continuous Process Improvement.( .) 4L-- Administration 16. PRICE CODE 17. SECURITY CLASSIFICATION 18. SECURITY...NSN 7540-01-280-5500 Standard Form 298 (Rev. 2-89) Pr.-cr,bed by ANSI Std ,,fl.f 296-102 DLA-X TOTAL QUALITY MANAGEMENT (TQM) IMPLEMENTATION PLAN o...application of proven Total Quality Management techniques. Quality Policy: Responsibility for quality is delegated to every employee ;11 DLA-X. Every

  13. Total Quality Management: A Guide to Implementation

    Science.gov (United States)

    1989-08-01

    Total Quality Management (TQM) is a means for improving personal effectiveness and performance and for aligning and focusing all individual efforts...your individual effort and extending its effect and its importance throughout an organization and beyond. Total Quality Management is not a destination

  14. SQuAVisiT : A Software Quality Assessment and Visualisation Toolset

    NARCIS (Netherlands)

    Roubtsov, Serguei; Telea, Alexandru; Holten, Danny

    2007-01-01

    Software quality assessment of large COBOL industrial legacy systems, both for maintenance or migration purposes, mounts a serious challenge. We present the Software Quality Assessment and Visualisation Toolset (SQuAVisiT), which assists users in performing the above task. First, it allows a fully a

  15. Total Quality Management Office for Contracting Integrity Implementation Plan

    Science.gov (United States)

    1989-07-01

    REPORT______ANDDATESCOVERED 4. TITLE AND SUBTITLE S. FUNDING NUMBERS Total Quality Management Office for Contracting Integrity Implementatiun Plan 6. AUTHOR(S) 7...01-280-5500 Standard Form 298 (Rev. 2-89) P’,croed 1:, ANSI Std 3J9-16 29d. 102 4 TOTAL QUALITY MANAGEMENT OFFICE FOR CONTRACTING INTEGRITY...IMPLEMENTATION PLAN According to the Total Quality Management (TQM) Master Plan, each PSE head, supported by Working Groups, will implement the HQ DLA Master

  16. Teamwork Quality and Project Success in Software Development: A Survey of Agile Development Teams

    OpenAIRE

    Lindsjørn, Yngve; Sjøberg, Dag; Dingsøyr, Torgeir; Bergersen, Gunnar R.; Dybå, Tore

    2016-01-01

    Small, self-directed teams are central in agile development. This article investigates the effect of team- work quality on team performance, learning and work satisfaction in agile software teams, and whether this effect differs from that of traditional software teams. A survey was administered to 477 respondents from 71 agile software teams in 26 companies and analyzed using structural equation modeling. A posi- tive effect of teamwork quality on team performance was found when team members ...

  17. A Comprehensive Survey on Evaluation of Lightweight Symmetric Ciphers: Hardware and Software Implementation

    Directory of Open Access Journals (Sweden)

    Jaber Hosseinzadeh

    2016-07-01

    Full Text Available Low-resource devices like wireless sensor networks have some limitations on memory, power and energy. Using common encryption algorithms are not appropriate for these devices due to their hard limitations and leads to a waste of energy and power. Here, lightweight symmetric ciphers have been evaluated in hardware and software implementations. Comprehensive Evaluation of lightweight ciphers in this work is performed based on cost, speed, efficiency and balance criterion. In each of the criteria, evaluation is done based on a specific measure and the best ciphers have been introduced in each. Evaluation in terms of hardware and software implementation indicates the superiority of SPECK and SIMON ciphers. Evaluation in terms of speed in hardware implementation indicates the superiority of Trivium and Grain, and it shows the superiority of MASHA and SPECK in software implementation. Results of the Evaluation in terms of efficiency express the superiority of SIMON and SPECK. The results of these evaluations helps finding ciphers appropriate to the user based on requirements and restrictions. The user sets his desired system and then obtains the system needs; at the final step, based on the type of requirements, the results of our work help the system to select the appropriate cipher.

  18. ISO and software quality assurance - licensing and certification of software professionals

    Energy Technology Data Exchange (ETDEWEB)

    Hare, J.; Rodin, L.

    1997-11-01

    This report contains viewgraphs on licensing and certifing of software professionals. Discussed in this report are: certification programs; licensing programs; why became certified; certification as a condition of empolyment; certification requirements; and examination structures.

  19. Internal Quality Assurance System and Its Implementation in Kaunas College

    Science.gov (United States)

    Misiunas, Mindaugas

    2007-01-01

    The article discusses the internal system of quality assurance and its implementation methods in Kaunas College. The issues of quality assurance are reviewed in the context of the European higher education area covering the three levels: European, national and institutional. The importance of quality assurance and its links with external…

  20. Assessing and updating the reliability of concrete bridges subjected to spatial deterioration - principles and software implementation

    DEFF Research Database (Denmark)

    Schneider, Ronald; Fischer, Johannes; Bügler, Maximilian

    2015-01-01

    . The overall system reliability is computed by means of a probabilistic structural model coupled with the deterioration model. Inspection data are included in the system reliability calculation through Bayesian updating on the basis of the DBN model. As proof of concept, a software prototype is developed...... to implement the method presented here. The software prototype is applied to a typical highway bridge and the influence of inspection information on the system deterioration state and the structural reliability is quantified taking into account the spatial correlation of the corrosion process. This work...... is a step towards developing a software tool that can be used by engineering practitioners to perform reliability assessments of ageing concrete bridges and update their reliability with inspection and monitoring data....

  1. Validation and clinical implementation of commercial secondary check software with heterogeneity corrections.

    Science.gov (United States)

    Kuppusamy, Vijayalakshmi; Nagarajan, Vivekanandan; Murugan, Lavanya

    2016-01-01

    To validate and implement PTW diamond secondary check software (SCS) in a routine clinical use. The secondary independent monitor unit or dose calculation verifications have led to a significant increase in the workflow associated with QA treatments. Modelling, validation and commissioning are necessary steps thereby making it a useful tool for QA. PTW Diamond SCS is capable of calculating VMAT fields, based on modified Clarkson integration, accounting for multi-leaf collimators (MLC) transmission and measured collimator scatter factors. Validation for heterogeneity corrections is made using circular phantom with inserts of various density materials. 150 VMAT plans were compared using (i) plans calculated in homogeneous cylindrical phantom and (ii) VMAT plans calculated with heterogeneity corrections using electron density values for each organ. Diamond SCS calculated dose for homogeneous cylindrical phantom resulted in average deviation of (0.1 ± 2.14%) with Eclipse TPS calculated dose and (-2.0 ± 1.66%) with absolute measured dose. PTW's OCTAVIUS-4D phantom with 729 ion chamber detector array measurements agreed well with Eclipse TPS calculated dose showing an average deviation of (-1.69 ± 1.56%). Diamond SCS dose calculations were performed with heterogeneity corrections for 124 VMAT plans with isocentre at a region above -350 HU. The overall MU variations between Diamond SCS and TPS Acuros-XB algorithms were within ±5%. Hence, the Diamond SCS can be used as an additional tool along with phantom measurements for patient specific quality assurance of VMAT plans with heterogeneity corrections having isocentre at a region above -350 HU.

  2. elrm: Software Implementing Exact-Like Inference for Logistic Regression Models

    Directory of Open Access Journals (Sweden)

    David Zamar

    2007-09-01

    Full Text Available Exact inference is based on the conditional distribution of the sufficient statistics for the parameters of interest given the observed values for the remaining sufficient statistics. Exact inference for logistic regression can be problematic when data sets are large and the support of the conditional distribution cannot be represented in memory. Additionally, these methods are not widely implemented except in commercial software packages such as LogXact and SAS. Therefore, we have developed elrm, software for R implementing (approximate exact inference for binomial regression models from large data sets. We provide a description of the underlying statistical methods and illustrate the use of elrm with examples. We also evaluate elrm by comparing results with those obtained using other methods.

  3. Effective Defect Prevention Approach in Software Process for Achieving Better Quality Levels

    CERN Document Server

    Suma, V

    2010-01-01

    Defect prevention is the most vital but habitually neglected facet of software quality assurance in any project. If functional at all stages of software development, it can condense the time, overheads and wherewithal entailed to engineer a high quality product. The key challenge of an IT industry is to engineer a software product with minimum post deployment defects. This effort is an analysis based on data obtained for five selected projects from leading software companies of varying software production competence. The main aim of this paper is to provide information on various methods and practices supporting defect detection and prevention leading to thriving software generation. The defect prevention technique unearths 99% of defects. Inspection is found to be an essential technique in generating ideal software generation in factories through enhanced methodologies of abetted and unaided inspection schedules. On an average 13 % to 15% of inspection and 25% - 30% of testing out of whole project effort tim...

  4. Implementing reusable software components for SNOMED CT diagram and expression concept representations.

    Science.gov (United States)

    Bánfai, Balázs; Porció, Roland; Kovács, Tibor

    2014-01-01

    SNOMED CT is a vital component in the future of semantic interoperability in healthcare as it provides the meaning to EHRs via its semantically rich, controlled terminology. Communicating the concepts of this terminology to both humans and machines is crucial therefore formal guidelines for diagram and expression representations have been developed by the curators of SNOMED CT. This paper presents a novel, model-based approach to implementing these guidelines that allows simultaneous editing of a concept via both diagram and expression editors. The implemented extensible software component can be embedded both both desktop and web applications.

  5. Hardware-Software Co-implementation of H.264 Decoder in SoC

    Institute of Scientific and Technical Information of China (English)

    YANG Yu-hong; ZHANG Wen-jun; XIONG Lian-xue; RAO Zhen-ning

    2006-01-01

    With the increasing demand for flexible and efficient implementation of image and video processing algorithms, there should be a good tradeoff between hardware and software design method. This paper utilized the HW-SW codesign method to implement the H. 264 decoder in an SoC with an ARM core, a multimedia processor and a deblocking filter coprocessor. For the parallel processing features of the multimedia processor, clock cycles of decoding process can be dramatically reduced. And the hardware dedicated deblocking filter coprocessor can improve the efficiency a lot. With maximum clock frequency of 150 MHz, the whole system can achieve real time processing speed and flexibility.

  6. MVC Architecture Driven Design and Agile Implementation of a Web-Based Software System

    Directory of Open Access Journals (Sweden)

    Sandhya Prakash

    2013-11-01

    Full Text Available This paper reports design and implementation of a web based software system for storing and managinginformation related to time management and productivity of employees working on a project.The systemhas been designed and implemented with best principles from model viewcontrollerand agile development.Such system has practical use for any organization in terms of ease of use, efficiency, and cost savings. Themanuscript describes design of the system as well as its database and user interface. Detailed snapshots ofthe working system are provided too

  7. Software Implementation of a Secure Firmware Update Solution in an IOT Context

    Directory of Open Access Journals (Sweden)

    Lukas Kvarda

    2016-01-01

    Full Text Available The present paper is concerned with the secure delivery of firmware updates to Internet of Things (IoT devices. Additionally, it deals with the design of a safe and secure bootloader for a UHF RFID reader. A software implementation of a secure firmware update solution is performed. The results show there is space to integrate even more security features into existing devices.

  8. Affect 4.0: a free software package for implementing psychological and psychophysiological experiments.

    Science.gov (United States)

    Spruyt, Adriaan; Clarysse, Jeroen; Vansteenwegen, Debora; Baeyens, Frank; Hermans, Dirk

    2010-01-01

    We describe Affect 4.0, a user-friendly software package for implementing psychological and psychophysiological experiments. Affect 4.0 can be used to present visual, acoustic, and/or tactile stimuli in highly complex (i.e., semirandomized and response-contingent) sequences. Affect 4.0 is capable of registering response latencies and analog behavioral input with millisecond accuracy. Affect 4.0 is available free of charge.

  9. Virtual Machine-level Software Transactional Memory: Principles, Techniques, and Implementation

    Science.gov (United States)

    2015-08-13

    VM-managed environment. ByteSTM is built by modifying Jikes RVM [3], a Java research virtual machine implemented in Java , using the optimizing...project have been publicly released as open-source software and research papers published at international conferences. In the following we summarize them... Research (AFOSR)/ RTC Arlington, Virginia 22203 Air Force Research Laboratory Air Force Materiel Command REPORT DOCUMENTATION PAGE Form Approved OMB No

  10. Design and implementation air quality monitoring robot

    Science.gov (United States)

    Chen, Yuanhua; Li, Jie; Qi, Chunxue

    2017-01-01

    Robot applied in environmental protection can break through the limitations in working environment, scope and mode of the existing environmental monitoring and pollution abatement equipments, which undertake the innovation and improvement in the basin, atmosphere, emergency and pollution treatment facilities. Actually, the relevant technology is backward with limited research and investment. Though the device companies have achieved some results in the study on the water quality monitoring, pipeline monitoring and sewage disposal, this technological progress on the whole is still much slow, and the mature product has not been formed. As a result, the market urges a demand of a new type of device which is more suitable for environmental protection on the basis of robot successfully applied in other fields. This paper designs and realizes a tracked mobile robot of air quality monitoring, which can be used to monitor air quality for the pollution accident in industrial parks and regular management.

  11. Total Quality Management Implementation Plan: Defense Depot, Ogden

    Science.gov (United States)

    1989-07-01

    NUMBERS Total Quality Management Implementation Plan Defense Depot Ogden 6. AUTHOR(S) 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) 8. PERFORMING...PAGES TQM ( Total Quality Management ), Continuous Process Improvement, Depot Operations, Process Action Teams 16. PRICE CODE 17. SECURITY...034 A Message From The Commander On Total Quality Management i fully support the DLA aoproacii to Total Quality Management . As stated by General

  12. Quality Management System Implementation at KELTEC-A Case Study

    OpenAIRE

    Aniyan, S.

    2002-01-01

    The scenario of quality is expanding exponentially. In the present-day world, quality is now in the driver's seat. With this globalisation of industrial products, many problems and opportunities have risen, primarily due to inconsistent standards existing in the market. The International Standards Organisation has tried (and seem to be very successful in doing so) to standardise a minimum level of quality norms. These norms act as the foundation to implements total quality management (...

  13. Quality Systems Implementation in the Pharmaceutical Industry

    African Journals Online (AJOL)

    Nafiisah

    competition amongst themselves, and to survive, one of the important issues to consider is the ... that pharmaceutical products reaching patients comply with quality standards ... found in the Pharmacy Act 1983 of Mauritius. According to the ... delegations from several countries and have registered its product for safety,.

  14. 78 FR 44485 - Implementation of the 2008 National Ambient Air Quality Standards for Ozone: State Implementation...

    Science.gov (United States)

    2013-07-24

    ... AGENCY 40 CFR Parts 50, 51, 70 and 71 Implementation of the 2008 National Ambient Air Quality Standards... on this rulemaking, contact Dr. Karl Pepple, Office of Air Quality Planning and Standards, U.S... Stackhouse, Office of Air Quality Planning and Standards, U.S. Environmental Protection Agency, by phone...

  15. 75 FR 65572 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2010-10-26

    ... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality Standards AGENCY: Environmental Protection Agency (EPA). ACTION: Direct final rule... of Ohio's Ambient Air Quality Standards (AAQS) into Ohio's State Implementation Plan (SIP) under the...

  16. The Role and Quality of Software Safety in the NASA Constellation Program

    Science.gov (United States)

    Layman, Lucas; Basili, Victor R.; Zelkowitz, Marvin V.

    2010-01-01

    In this study, we examine software safety risk in the early design phase of the NASA Constellation spaceflight program. Obtaining an accurate, program-wide picture of software safety risk is difficult across multiple, independently-developing systems. We leverage one source of safety information, hazard analysis, to provide NASA quality assurance managers with information regarding the ongoing state of software safety across the program. The goal of this research is two-fold: 1) to quantify the relative importance of software with respect to system safety; and 2) to quantify the level of risk presented by software in the hazard analysis. We examined 154 hazard reports created during the preliminary design phase of three major flight hardware systems within the Constellation program. To quantify the importance of software, we collected metrics based on the number of software-related causes and controls of hazardous conditions. To quantify the level of risk presented by software, we created a metric scheme to measure the specificity of these software causes. We found that from 49-70% of hazardous conditions in the three systems could be caused by software or software was involved in the prevention of the hazardous condition. We also found that 12-17% of the 2013 hazard causes involved software, and that 23-29% of all causes had a software control. Furthermore, 10-12% of all controls were software-based. There is potential for inaccuracy in these counts, however, as software causes are not consistently scoped, and the presence of software in a cause or control is not always clear. The application of our software specificity metrics also identified risks in the hazard reporting process. In particular, we found a number of traceability risks in the hazard reports may impede verification of software and system safety.

  17. Measurement and Management of the Level of Quality Control Process in SoC (System on Chip Embedded Software Development

    Directory of Open Access Journals (Sweden)

    Ki-won Song

    2012-04-01

    quality control activities and it is desirable to create a quality process to integrally represent overall level of quality control activities performed while developing the software deliverables. With the quality process, it is possible to evaluate whether enough quality control activities are performed for the project officially and secure the quality of the software deliverables before it is delivered to the customers.

  18. Keys for successful implementation of total quality management in hospitals.

    Science.gov (United States)

    Carman, James M; Shortell, Stephen M; Foster, Richard W; Hughes, Edward F X; Boerstler, Heidi; O' Brien, James L; O'Connor, Edward J

    2010-01-01

    Editor's Note: This article reports the findings of an analysis of the implementation of continuous quality improvement (CQI) or total quality management (TQM) programs in 10 hospitals. This analysis is the result of a 2-year study designed to identify and assess the ingredients that lead to the successful implementation of CQI programs in acute care hospitals. This article first appeared in Health Care Management Review 21(1), 48-60. Copyright © 1996 Aspen Publishers, Inc. (Lippincott Williams & Wilkins).

  19. Implementing Total Quality Management in a University Setting.

    Science.gov (United States)

    Coate, L. Edwin

    1991-01-01

    Oregon State University implemented Total Quality Management in nine phases: exploration; establishing a pilot study team; defining customer needs; adopting the breakthrough planning process; performing breakthrough planning in divisions; forming daily management teams; initiating cross-functional pilot projects; implementing cross-functional…

  20. Implementing Total Quality Management in a University Setting.

    Science.gov (United States)

    Coate, L. Edwin

    1991-01-01

    Oregon State University implemented Total Quality Management in nine phases: exploration; establishing a pilot study team; defining customer needs; adopting the breakthrough planning process; performing breakthrough planning in divisions; forming daily management teams; initiating cross-functional pilot projects; implementing cross-functional…

  1. A proposed metamodel for the implementation of object oriented software through the automatic generation of source code

    Directory of Open Access Journals (Sweden)

    CARVALHO, J. S. C.

    2008-12-01

    Full Text Available During the development of software one of the most visible risks and perhaps the biggest implementation obstacle relates to the time management. All delivery deadlines software versions must be followed, but it is not always possible, sometimes due to delay in coding. This paper presents a metamodel for software implementation, which will rise to a development tool for automatic generation of source code, in order to make any development pattern transparent to the programmer, significantly reducing the time spent in coding artifacts that make up the software.

  2. Implementation of quality management and quality measurements in education

    DEFF Research Database (Denmark)

    Madsen, Ole Nørgaard; Carlsson, Ricci

    Gennemgang af program for implementering af kvalitetsledelse på en uddannelsesinstitution. I 7 faser præsenteres en fremgangsmåde baseret på den danske/europæiske model for selvevaluering som grundlag for kvalitetsbeskrivelse af virksomheden, prioritering af kvalitetsindsats, planlægning af kvali...

  3. Advanced quality prediction model for software architectural knowledge sharing

    NARCIS (Netherlands)

    Liang, Peng; Jansen, Anton; Avgeriou, Paris; Tang, Antony; Xu, Lai

    2011-01-01

    In the field of software architecture, a paradigm shift is occurring from describing the outcome of architecting process to describing the Architectural Knowledge (AK) created and used during architecting. Many AK models have been defined to represent domain concepts and their relationships, and the

  4. An empirical study of software architectures' effect on product quality

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Jonasson, Kristjan; Neukirchen, Helmut

    2011-01-01

    Software architectures shift the focus of developers from lines-of-code to coarser-grained components and their interconnection structure. Unlike 2ne-grained objects, these components typically encompass business functionality and need to be aware of the underlying business processes. Hence, the ...

  5. Implementing Kanban for agile process management within the ALMA Software Operations Group

    Science.gov (United States)

    Reveco, Johnny; Mora, Matias; Shen, Tzu-Chiang; Soto, Ruben; Sepulveda, Jorge; Ibsen, Jorge

    2014-07-01

    After the inauguration of the Atacama Large Millimeter/submillimeter Array (ALMA), the Software Operations Group in Chile has refocused its objectives to: (1) providing software support to tasks related to System Integration, Scientific Commissioning and Verification, as well as Early Science observations; (2) testing the remaining software features, still under development by the Integrated Computing Team across the world; and (3) designing and developing processes to optimize and increase the level of automation of operational tasks. Due to their different stakeholders, each of these tasks presents a wide diversity of importances, lifespans and complexities. Aiming to provide the proper priority and traceability for every task without stressing our engineers, we introduced the Kanban methodology in our processes in order to balance the demand on the team against the throughput of the delivered work. The aim of this paper is to share experiences gained during the implementation of Kanban in our processes, describing the difficulties we have found, solutions and adaptations that led us to our current but still evolving implementation, which has greatly improved our throughput, prioritization and problem traceability.

  6. 78 FR 19990 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2013-04-03

    ... From the Federal Register Online via the Government Publishing Office ENVIRONMENTAL PROTECTION AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality Standards; Correction AGENCY: Environmental Protection Agency (EPA). ACTION: Final...

  7. Thirty Meter Telescope: observatory software requirements, architecture, and preliminary implementation strategies

    Science.gov (United States)

    Silva, David R.; Angeli, George; Boyer, Corinne; Sirota, Mark; Trinh, Thang

    2008-07-01

    The Thirty Meter Telescope (TMT) will be a ground-based, 30-m optical-IR alt-az telescope with a highly segmented primary mirror located in a remote location. Efficient science operations require the asynchronous coordination of many different sub-systems including telescope mount, three independent active optics sub-systems, adaptive optics, laser guide stars, and user-configured science instrument. An important high-level requirement is target acquisition and observatory system configuration must be completed in less than 5 minutes (or 10 minutes if moving to a new instrument). To meet this coordination challenge and target acquisition time requirement, a distributed software architecture is envisioned consisting of software components linked by a service-based software communications backbone. A master sequencer coordinates the activities of mid-layer sequencers for the telescope, adaptive optics, and selected instrument. In turn, these mid-layer sequencers coordinate the activities of groups of sub-systems. In this paper, TMT observatory requirements are presented in more detail, followed by a description of the design reference software architecture and a discussion of preliminary implementation strategies.

  8. Implementation of total quality management Case study: British Airways

    Directory of Open Access Journals (Sweden)

    Anca MADAR

    2015-06-01

    Full Text Available Total Quality Management (TQM has an important role in any company, since the implementation of this program companies can continuously improve their performance. Thus, organizations will be able to considerably meet their internal and external clients' needs in terms of services and products quality, and they will also be able to develop an efficient and profitable business. The paper analyzes how this concept was implemented by the British Airways airline, since it is among the first companies that have implemented such a program.

  9. A Framework for Evaluating the Software Product Quality of Pregnancy Monitoring Mobile Personal Health Records.

    Science.gov (United States)

    Idri, Ali; Bachiri, Mariam; Fernández-Alemán, José Luis

    2016-03-01

    Stakeholders' needs and expectations are identified by means of software quality requirements, which have an impact on software product quality. In this paper, we present a set of requirements for mobile personal health records (mPHRs) for pregnancy monitoring, which have been extracted from literature and existing mobile apps on the market. We also use the ISO/IEC 25030 standard to suggest the requirements that should be considered during the quality evaluation of these mPHRs. We then go on to design a checklist in which we contrast the mPHRs for pregnancy monitoring requirements with software product quality characteristics and sub-characteristics in order to calculate the impact of these requirements on software product quality, using the ISO/IEC 25010 software product quality standard. The results obtained show that the requirements related to the user's actions and the app's features have the most impact on the external sub-characteristics of the software product quality model. The only sub-characteristic affected by all the requirements is Appropriateness of Functional suitability. The characteristic Operability is affected by 95% of the requirements while the lowest degrees of impact were identified for Compatibility (15%) and Transferability (6%). Lastly, the degrees of the impact of the mPHRs for pregnancy monitoring requirements are discussed in order to provide appropriate recommendations for the developers and stakeholders of mPHRs for pregnancy monitoring.

  10. Component-based Software Design and Implementation for Network Security System

    Directory of Open Access Journals (Sweden)

    Jianchao Han

    2009-12-01

    Full Text Available A computer network intrusion detection and prevention system consists of collecting network traffic data, discovering user behavior patterns as intrusion detection rules, and applying these rules to prevent malicious and misuse. Many commercial off-the-shelf (COTS products have been developed to perform each of these tasks. In this paper, the component-based software engineering approach is exploited to integrate these COTS products as components into a computerized system to automatically detect intrusion rules from network traffic data and setup IPTables to prevent future potential attacks. The component- based software architecture of this kind of system is designed, COTS components are analyzed and selected, adaptor components to connect COTS products are developed, the system implementation is illustrated, and the preliminary system experiment is presented.

  11. Total Quality Management Implementation at the Defense Technical Information Center

    Science.gov (United States)

    1989-09-01

    improvement programs. -- 7- 14. SUBJECT TERMS 15. NUMBER OF PAGES TOM ( Total Quality Management ), Continuous Process Improvement, ________ Collection and...TECHNICAL INFORMATION CENTER September 1989 Approved for Public Release; Distribution is Unlimited. I8 . 22 ~89 9 29 0 22 TOTAL QUALITY MANAGEMENT IMPLEMENTATION...technical information support, services, and products to the DoD research and development community. The DTIC Total Quality Management (TQM

  12. Total Quality Management Implementation Plan for Military Personnel Management

    Science.gov (United States)

    1989-09-01

    2050.. )ATE 3. REPORT TYPE AND DATES CO VERED 4. TITLE AND SUBTITLE 5,rrmir18 . FUNDING NUMBERS Total Quality Management Implementation Plan for...SUBJECT TERMS 15. NUMBER OF PAGES TQM ( Total Quality Management ), Military Personnel Management, Continuous Process Improvement 16. PRICE CODE 17. SECURITY...UNCLASSIFIED UNCLASSIFIED UL NSN 7540-01-280-5500 Standard Form 298 (Rev. 2-89) Prescribed by ANSI Std Z39-16 296-102 TOTAL QUALITY MANAGEMENT I

  13. ThermoData Engine (TDE) software implementation of the dynamic data evaluation concept. 7. Ternary mixtures.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Muzny, Chris D; Kazakov, Andrei F; Kroenlein, Kenneth; Magee, Joseph W; Abdulagatov, Ilmutdin; Kang, Jeong Won; Frenkel, Michael

    2012-01-23

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported in this journal. The present paper describes the first application of this concept to the evaluation of thermophysical properties for ternary chemical systems. The method involves construction of Redlich-Kister type equations for individual properties (excess volume, thermal conductivity, viscosity, surface tension, and excess enthalpy) and activity coefficient models for phase equilibrium properties (vapor-liquid and liquid-liquid equilibrium). Constructed ternary models are based on those for the three pure component and three binary subsystems evaluated on demand through the TDE software algorithms. All models are described in detail, and extensions to the class structure of the program are provided. Reliable evaluation of properties for the binary subsystems is essential for successful property evaluations for ternary systems, and algorithms are described to aid appropriate parameter selection and fitting for the implemented activity coefficient models (NRTL, Wilson, Van Laar, Redlich-Kister, and UNIQUAC). Two activity coefficient models based on group contributions (original UNIFAC and NIST-KT-UNIFAC) are also implemented. Novel features of the user interface are shown, and directions for future enhancements are outlined.

  14. Principles for Evaluating the Quality Attributes of a Software Architecture.

    Science.gov (United States)

    1997-03-01

    gay , lesbian and bisexual students from receiving ROTC scholarships or serving in the military. Nevertheless, all ROTC classes at Carnegie Mellon...attributes. Some are measured by system activities (e.g., latency, availability), some are measured by inspection activities (e.g., coupling , cohesion... Italy , October 11-13,1995. New York: Springer, 1995. Fenton, N. "How Effective Are Software Engineering Methods?" 295-305. Proceedings ofAQUIS

  15. Handbook of software quality assurance techniques applicable to the nuclear industry

    Energy Technology Data Exchange (ETDEWEB)

    Bryant, J.L.; Wilburn, N.P.

    1987-08-01

    Pacific Northwest Laboratory is conducting a research project to recommend good engineering practices in the application of 10 CFR 50, Appendix B requirements to assure quality in the development and use of computer software for the design and operation of nuclear power plants for NRC and industry. This handbook defines the content of a software quality assurance program by enumerating the techniques applicable. Definitions, descriptions, and references where further information may be obtained are provided for each topic.

  16. E-learning Materials Development: Applying and Implementing Software Reuse Principles and Granularity Levels in the Small

    Directory of Open Access Journals (Sweden)

    Nabil Arman

    2010-06-01

    Full Text Available E-learning materials development is typically acknowledged as an expensive, complicated, and lengthy process, often producing materials that are of low quality and difficult to adaptand maintain. It has always been a challenge to identify proper e-learning materials that can be reused at a reasonable cost and effort. In this paper, software engineering reuse principlesare applied to e-learning materials development process. These principles are then applied and implemented in a prototype that is integrated with an open-source course management systems. The reuse of existing e-learning materials is beneficial in improving developers of elearning materials productivity. E-learning material reuse is performed, in this research, based on construct’s granularity rather than on unified constructs of one size.

  17. Pre-Hardware Optimization of Spacecraft Image Processing Software Algorithms and Hardware Implementation

    Science.gov (United States)

    Kizhner, Semion; Flatley, Thomas P.; Hestnes, Phyllis; Jentoft-Nilsen, Marit; Petrick, David J.; Day, John H. (Technical Monitor)

    2001-01-01

    Spacecraft telemetry rates have steadily increased over the last decade presenting a problem for real-time processing by ground facilities. This paper proposes a solution to a related problem for the Geostationary Operational Environmental Spacecraft (GOES-8) image processing application. Although large super-computer facilities are the obvious heritage solution, they are very costly, making it imperative to seek a feasible alternative engineering solution at a fraction of the cost. The solution is based on a Personal Computer (PC) platform and synergy of optimized software algorithms and re-configurable computing hardware technologies, such as Field Programmable Gate Arrays (FPGA) and Digital Signal Processing (DSP). It has been shown in [1] and [2] that this configuration can provide superior inexpensive performance for a chosen application on the ground station or on-board a spacecraft. However, since this technology is still maturing, intensive pre-hardware steps are necessary to achieve the benefits of hardware implementation. This paper describes these steps for the GOES-8 application, a software project developed using Interactive Data Language (IDL) (Trademark of Research Systems, Inc.) on a Workstation/UNIX platform. The solution involves converting the application to a PC/Windows/RC platform, selected mainly by the availability of low cost, adaptable high-speed RC hardware. In order for the hybrid system to run, the IDL software was modified to account for platform differences. It was interesting to examine the gains and losses in performance on the new platform, as well as unexpected observations before implementing hardware. After substantial pre-hardware optimization steps, the necessity of hardware implementation for bottleneck code in the PC environment became evident and solvable beginning with the methodology described in [1], [2], and implementing a novel methodology for this specific application [6]. The PC-RC interface bandwidth problem for the

  18. Proceso de pruebas para productos de software en un laboratorio de calidad /Testing process for software products at a quality laboratory

    Directory of Open Access Journals (Sweden)

    Dalila Jústiz-Núñez

    2014-04-01

    Full Text Available La calidad en sentido general, tanto de software como de otros tipos de productos, es un elemento que cada día se tiene más en cuenta a nivel mundial y su logro se relaciona directamente con el proceso que se emplee para obtenerla. Este trabajo presenta una propuesta de proceso de pruebas de software, para un Laboratorio de Calidad, inmerso en un ambiente universitario. Se detallan las actividades de los procesos fundamentales y los artefactos de salida, los niveles de prueba que se aplican y otros elementos de interés. Además se muestra una experiencia práctica de aplicación del proceso y los resultados de varios casos de estudio. Esta propuesta incluye la definición de los aspectos metodológicos y la selección de herramientas que automaticen el proceso. ABSTRACT In general terms, the quality of the software as of other products, is an element of increasing importance worldwide and it is strongly linked to its obtaining process. This work presents a proposal of a software testing process for a Quality Laboratory, integrated into an academic environment. The activities of the main processes and the output artifacts were detailed, as well as the testing levels applied, among other elements of interest. It was also showed a practical experience related to the process implementation and the results of several study cases. This proposal includes the definition of the methodological issues and the selection of the tools for the process automation.

  19. Implementation of Programmatic Quality and the Impact on Safety

    Science.gov (United States)

    Huls, Dale Thomas; Meehan, Kevin

    2005-01-01

    The purpose of this paper is to discuss the implementation of a programmatic quality assurance discipline within the International Space Station Program and the resulting impact on safety. NASA culture has continued to stress safety at the expense of quality when both are extremely important and both can equally influence the success or failure of a Program or Mission. Although safety was heavily criticized in the media after Colimbiaa, strong case can be made that it was the failure of quality processes and quality assurance in all processes that eventually led to the Columbia accident. Consequently, it is possible to have good quality processes without safety, but it is impossible to have good safety processes without quality. The ISS Program quality assurance function was analyzed as representative of the long-term manned missions that are consistent with the President s Vision for Space Exploration. Background topics are as follows: The quality assurance organizational structure within the ISS Program and the interrelationships between various internal and external organizations. ISS Program quality roles and responsibilities with respect to internal Program Offices and other external organizations such as the Shuttle Program, JSC Directorates, NASA Headquarters, NASA Contractors, other NASA Centers, and International Partner/participants will be addressed. A detailed analysis of implemented quality assurance responsibilities and functions with respect to NASA Headquarters, the JSC S&MA Directorate, and the ISS Program will be presented. Discussions topics are as follows: A comparison of quality and safety resources in terms of staffing, training, experience, and certifications. A benchmark assessment of the lessons learned from the Columbia Accident Investigation (CAB) Report (and follow-up reports and assessments), NASA Benchmarking, and traditional quality assurance activities against ISS quality procedures and practices. The lack of a coherent operational

  20. Implementation of Programmatic Quality and the Impact on Safety

    Science.gov (United States)

    Huls, Dale Thomas; Meehan, Kevin

    2005-01-01

    The purpose of this paper is to discuss the implementation of a programmatic quality assurance discipline within the International Space Station Program and the resulting impact on safety. NASA culture has continued to stress safety at the expense of quality when both are extremely important and both can equally influence the success or failure of a Program or Mission. Although safety was heavily criticized in the media after Colimbiaa, strong case can be made that it was the failure of quality processes and quality assurance in all processes that eventually led to the Columbia accident. Consequently, it is possible to have good quality processes without safety, but it is impossible to have good safety processes without quality. The ISS Program quality assurance function was analyzed as representative of the long-term manned missions that are consistent with the President s Vision for Space Exploration. Background topics are as follows: The quality assurance organizational structure within the ISS Program and the interrelationships between various internal and external organizations. ISS Program quality roles and responsibilities with respect to internal Program Offices and other external organizations such as the Shuttle Program, JSC Directorates, NASA Headquarters, NASA Contractors, other NASA Centers, and International Partner/participants will be addressed. A detailed analysis of implemented quality assurance responsibilities and functions with respect to NASA Headquarters, the JSC S&MA Directorate, and the ISS Program will be presented. Discussions topics are as follows: A comparison of quality and safety resources in terms of staffing, training, experience, and certifications. A benchmark assessment of the lessons learned from the Columbia Accident Investigation (CAB) Report (and follow-up reports and assessments), NASA Benchmarking, and traditional quality assurance activities against ISS quality procedures and practices. The lack of a coherent operational

  1. ThermoData Engine (TDE): software implementation of the dynamic data evaluation concept. 3. Binary mixtures.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Kazakov, Andrei F; Muzny, Chris D; Frenkel, Michael

    2009-02-01

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported recently in this journal. The present paper describes the first application of this concept to the evaluation of thermophysical properties for binary chemical systems. Five activity-coefficient models have been implemented for representation of phase-equilibrium data (vapor-liquid, liquid-liquid, and solid-liquid equilibrium): NRTL, UNIQUAC, Van Laar, Margules/Redlich-Kister, and Wilson. Implementation of these models in TDE is fully described. Properties modeled individually are densities, surface tensions, critical temperatures, critical pressures, excess enthalpies, and the transport properties-viscosity and thermal conductivity. Extensions to the class structure of the program are described with emphasis on special features allowing close linkage between mixture and pure-component properties required for implementation of the models. Details of gas-phase models used in conjunction with the activity-coefficient models are shown. Initial implementation of the dynamic data evaluation concept for reactions is demonstrated with evaluation of enthalpies of formation for compounds containing carbon, hydrogen, oxygen, and nitrogen. Directions for future enhancements are outlined.

  2. Design and Implementation of Visual Dynamic Display Software of Gene Expression Based on GTK

    Institute of Scientific and Technical Information of China (English)

    JIANG Wei; MENG Fanjiang; LI Yong; YU Xiao

    2009-01-01

    The paper presented an implement method for a dynamic gene expression display software based on the GTK. This method established the dynamic presentation system of gene expression which according to gene expression data from gene chip hybridize at different time, adopted a linearity combination model and Pearson correlation coefficient algorithm. The system described the gene expression changes in graphic form, the gene expression changes with time and the changes in characteristics of the gene expression, also the changes in relations of the gene expression and regulation relationships among genes. The system also provided an integrated platform for analysis on gene chips data, especially for the research on the network of gene regulation.

  3. The SIFT computer and its development. [Software Implemented Fault Tolerance for aircraft control

    Science.gov (United States)

    Goldberg, J.

    1981-01-01

    Software Implemented Fault Tolerance (SIFT) is an aircraft control computer designed to allow failure probability of less than 10 to the -10th/hour. The system is based on advanced fault-tolerance computing and validation methodology. Since confirmation of reliability by observation is essentially impossible, system reliability is estimated by a Markov model. A mathematical proof is used to justify the validity of the Markov model. System design is represented by a hierarchy of abstract models, and the design proof comprises mathematical proofs that each model is, in fact, an elaboration of the next more abstract model.

  4. Modeling and software implementation of flight system for simulator of a new fighter

    Institute of Scientific and Technical Information of China (English)

    SUI Cheng-cheng; YANG Yong-tian; JIA Rong-zhen

    2004-01-01

    Real-time modeling and simulation of flight system are the key parts of simulator. After describing the architecture of simulator for a newer fighter, author presents the composition of flight system and its mathematic models. In this paper, aircraft is regarded as an elastic flight body. And a new integrated algorithm which can remedy the shortcoming of Euler method and four-element method is used to calculate the Eulerian angles of aircraft. Finally, the software implementation of the flight system is given in the paper.

  5. Maximizing the utilization and impact of medical educational software by designing for local area network (LAN) implementation.

    Science.gov (United States)

    Stevens, R; Reber, E

    1993-01-01

    The design, development and implementation of medical education software often occurs without sufficient consideration of the potential benefits that can be realized by making the software network aware. These benefits can be considerable and can greatly enhance the utilization and potential impact of the software. This article details how multiple aspects of the IMMEX problem solving project have benefited from taking maximum advantage of LAN resources.

  6. THE EXISTING BARRIERS IN IMPLEMENTING TOTAL QUALITY MANAGEMENT

    Directory of Open Access Journals (Sweden)

    Salagean Horatiu Catalin

    2014-07-01

    Full Text Available In the current market economy, companies are constantly struggling to achieve a sustained competitive advantage that will enable them to improve performance, which results in increased competitiveness, and of course, profit. Among the few competitive advantages that can become sustainable competitive advantages, quality plays a crucial role. Recent research shows that about 90% of buyers in the international market, consider quality as having at least equal importance with price in making the decision to purchase. In the opinion of some specialists in economic theory and practice, total quality refers to the holistic approach of quality, which actually means, addressing all aspects of economic and social development and technical of quality. Thus, the holistic approach of quality at organisation-wide involves procedural approach of quality, in this respect, the study focuses on this type of quality approach, i.e. the procedural approach, taking into account the strategic aspects of the continuous improvement of quality, which means in fact, the quality management. Total Quality Management is seen as a way to transform the economies of some countries to be more competitive than others. However, Total Quality Management brings not and will not produce results overnight, it is not a panacea for all the problems facing the organization. Total Quality Management requires a change in organizational culture, which must focus on meeting customer expectations and increasing the involvement of all employees to meet this objective, as an expression of the ethics of continuous improvement. In general, research on quality aiming identify why an organization should adopt the principles of total quality management, but attempts to identify the failing companies' attempts to implement total quality management principles are not so visible. Concerns companies to introduce quality management systems are becoming more pronounced, therefore, in this study we try to

  7. 75 FR 65594 - Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient Air Quality...

    Science.gov (United States)

    2010-10-26

    ... From the Federal Register Online via the Government Publishing Office ENVIRONMENTAL PROTECTION AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Ohio; Ohio Ambient... consolidation of Ohio's Ambient Air Quality Standards (AAQS) into Ohio's State Implementation Plan (SIP)...

  8. 78 FR 30829 - Approval and Promulgation of Air Quality Implementation Plans; Illinois; Air Quality Standards...

    Science.gov (United States)

    2013-05-23

    ... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Illinois; Air Quality Standards Revision AGENCY: Environmental Protection Agency (EPA). ACTION: Proposed rule. SUMMARY: EPA is proposing to approve a revision to the Illinois State Implementation Plan (SIP) to reflect...

  9. Quality Management System Implementation at KELTEC-A Case Study

    Directory of Open Access Journals (Sweden)

    S. Aniyan

    2002-10-01

    Full Text Available The scenario of quality is expanding exponentially. In the present-day world, quality is now in the driver's seat. With this globalisation of industrial products, many problems and opportunities have risen, primarily due to inconsistent standards existing in the market. The International Standards Organisation has tried (and seem to be very successful in doing so to standardise a minimum level of quality norms. These norms act as the foundation to implements total quality management (TQM. TQM has been identified as a philosophy of interdependent characteristics, such as customer focus, commitment, measurement, systematic support, and continuous improvement. This paper presents a case study on the implementation of ISO 9002: 1994 at Kerala Hitech Industries (KELTEC, with the benefits, hurdles faced, and shortcomings. Upgradation of KELTEC's quality management system to the requirements of ISO 9001: 2000 is currently under progress.

  10. Implementation of Shifted Periodic Boundary Conditions in the Large-Scale Atomic/Molecular Massively Parallel Simulator (LAMMPS) Software

    Science.gov (United States)

    2015-08-01

    Atomic /Molecular Massively Parallel Simulator (LAMMPS) Software by N Scott Weingarten and James P Larentzos Approved for...0687 ● AUG 2015 US Army Research Laboratory Implementation of Shifted Periodic Boundary Conditions in the Large-Scale Atomic /Molecular...Shifted Periodic Boundary Conditions in the Large-Scale Atomic /Molecular Massively Parallel Simulator (LAMMPS) Software 5a. CONTRACT NUMBER 5b

  11. THE QUALITY MANAGEMENT SYSTEM AT THE UNIVERSITY. PROBLEMS IN IMPLEMENTING

    Directory of Open Access Journals (Sweden)

    L. I. Nazina

    2014-01-01

    Full Text Available Summary. Beginning of the XXI century marked the active integration of Russia and the world community, including entry into the world educational space. The international activities of the leading universities of the country depends on the level of cooperation with foreign countries in the field of scientific and educational activities. Today topical issue recognition of the quality of Russian education abroad. One way to confirm this is the certification of the quality management system of the university. As long-term, mostly foreign experience, achieving the required level of quality and continue to improve the most productive is accomplished by implementing a quality management system in accordance with ISO 9000. In the modern system of higher education a lot of attention paid to quality, to ensure the competitiveness of the institution in the education market by training qualified specialists. Implementation of the quality management system is to move the university to process management prepared in accordance with approved and enacted by the QMS documents. Workplaces should determine the duties, powers and responsibilities of each employee within the quality management system, develop and implement the job descriptions, staff verify documented procedures. Implementation of the certification must be accompanied by specific and persistent administration by management in conjunction with the training and retraining of all managers and employees involved in the work on the implementation of the QMS in the organization. This article describes the goals, objectives in the formation and implementation of QMS in higher education arising from this problem and barriers (information, methodological, psychological, multiple ways to address them.

  12. Exploring Issues of Implementation, Equity, and Student Achievement With Educational Software in the DC Public Schools

    Directory of Open Access Journals (Sweden)

    June Ahn

    2016-08-01

    Full Text Available In this article, we present analyses from a researcher-practitioner partnership in the District of Columbia Public Schools, where we are exploring the impact of educational software on students’ academic achievement. We analyze a unique data set that combines student-level information from the district with data of student usage of a mathematics game platform: First in Math (FIM. These data offer a window into long-standing issues in the educational technology literature around implementation, equity, and student achievement. We show that time spent in FIM was correlated with improved future performance on standardized math assessments for students in Grades 4–8. However, student time spent using FIM was highly related to factors such as race, gender, and prior achievement. Such observations from data are helpful for school districts and researchers to inform equitable implementation of new technologies and maximize benefits to learners.

  13. A Case of Engineering Quality for Mobile Healthcare Applications Using Augmented Personal Software Process Improvement

    Directory of Open Access Journals (Sweden)

    Shahbaz Ahmed Khan Ghayyur

    2016-01-01

    Full Text Available Mobile healthcare systems are currently considered as key research areas in the domain of software engineering. The adoption of modern technologies, for mobile healthcare systems, is a quick option for industry professionals. Software architecture is a key feature that contributes towards a software product, solution, or services. Software architecture helps in better communication, documentation of design decisions, risks identification, basis for reusability, scalability, scheduling, and reduced maintenance cost and lastly it helps to avoid software failures. Hence, in order to solve the abovementioned issues in mobile healthcare, the software architecture is integrated with personal software process. Personal software process has been applied successfully but it is unable to address the issues related to architectural design and evaluation capabilities. Hence, a new technique architecture augmented personal process is presented in order to enhance the quality of the mobile healthcare systems through the use of architectural design with integration of personal software process. The proposed process was validated by case studies. It was found that the proposed process helped in reducing the overall costs and effort. Moreover, an improved architectural design helped in development of high quality mobile healthcare system.

  14. Assessing quality in software development: An agile methodology approach

    National Research Council Canada - National Science Library

    V Rodríguez-Hernández; M C Espino-Gudiño; J L González-Pérez; J Gudiño-Bazaldúa; V M Castaño

    2015-01-01

      A novel methodology, result of 10 years of in-field testing, which makes possible the convergence of different types of models and quality standards for Engineering and Computer Science Faculties, is presented...

  15. A Simple Stream Water Quality Modelling Software for Educational and Training Purposes

    OpenAIRE

    Erturk, Ali

    2010-01-01

    Water quality models are important decision support system tools for water pollution control, study of the health of aquatic ecosystems and assessment of the effects of point and diffuse pollution. However, water quality models are usually comprehensive software, which are usually not easy to learn and apply. Thus extensive training is needed before scientists and engineers can use most of the water quality models effectively. In this study; a new, easy to use and simple stream water quality ...

  16. Fuzzy system for risk analysis in software projects through the attributes of quality standards iso 25000

    Directory of Open Access Journals (Sweden)

    Chau Sen Shia

    2014-02-01

    Full Text Available With the growth in demand for products and services in the IT area, companies encounter difficulties in establishing a metric or measure of quality of services to address measurably qualitative values in their planning. In this work fuzzy logic, standard SQuaRE (measurement of the quality of software products, Likertscale, GQM method (Goal-Question-Metric -indicator of quality of Software and the project risk analysis model of Boehm were used to assess the quality of services and decision-making, according to your demand and requests for software development. With the aim of improving the quality in the provision of services, the application is used to integrate the team and follow the life cycle of a project from its initial phase, and to assist in the comparison with the proposed schedule during the requirements elicitation.

  17. Implementing an integrative multi-agent clinical decision support system with open source software.

    Science.gov (United States)

    Sayyad Shirabad, Jelber; Wilk, Szymon; Michalowski, Wojtek; Farion, Ken

    2012-02-01

    Clinical decision making is a complex multi-stage process. Decision support can play an important role at each stage of this process. At present, the majority of clinical decision support systems have been focused on supporting only certain stages. In this paper we present the design and implementation of MET3-a prototype multi-agent system providing an integrative decision support that spans over the entire decision making process. The system helps physicians with data collection, diagnosis formulation, treatment planning and finding supporting evidence. MET3 integrates with external hospital information systems via HL7 messages and runs on various computing platforms available at the point of care (e.g., tablet computers, mobile phones). Building MET3 required sophisticated and reliable software technologies. In the past decade the open source software movement has produced mature, stable, industrial strength software systems with a large user base. Therefore, one of the decisions that should be considered before developing or acquiring a decision support system is whether or not one could use open source technologies instead of proprietary ones. We believe MET3 shows that the answer to this question is positive.

  18. Architecture and Implementation of OpenPET Firmware and Embedded Software.

    Science.gov (United States)

    Abu-Nimeh, Faisal T; Ito, Jennifer; Moses, William W; Peng, Qiyu; Choong, Woon-Seng

    2016-04-01

    OpenPET is an open source, modular, extendible, and high-performance platform suitable for multi-channel data acquisition and analysis. Due to the flexibility of the hardware, firmware, and software architectures, the platform is capable of interfacing with a wide variety of detector modules not only in medical imaging but also in homeland security applications. Analog signals from radiation detectors share similar characteristics - a pulse whose area is proportional to the deposited energy and whose leading edge is used to extract a timing signal. As a result, a generic design method of the platform is adopted for the hardware, firmware, and software architectures and implementations. The analog front-end is hosted on a module called a Detector Board, where each board can filter, combine, timestamp, and process multiple channels independently. The processed data is formatted and sent through a backplane bus to a module called Support Board, where 1 Support Board can host up to eight Detector Board modules. The data in the Support Board, coming from 8 Detector Board modules, can be aggregated or correlated (if needed) depending on the algorithm implemented or runtime mode selected. It is then sent out to a computer workstation for further processing. The number of channels (detector modules), to be processed, mandates the overall OpenPET System Configuration, which is designed to handle up to 1,024 channels using 16-channel Detector Boards in the Standard System Configuration and 16,384 channels using 32-channel Detector Boards in the Large System Configuration.

  19. Global Crisis as Enterprise Software Motivator: from Lifecycle Optimization to Efficient Implementation Series

    Directory of Open Access Journals (Sweden)

    Sergey V. Zykov

    2012-04-01

    Full Text Available It is generally known that software system development lifecycle (SSDL should be managed adequately. The global economy crisis and subsequent depression have taught us certain lessons on the subject, which is so vital for enterprises. The paper presents the adaptive methodology of enterprise SSDL, which allows to avoid "local crises" while producing large-scale software. The methodology is based on extracting common ERP module level patterns and applying them to series of heterogeneous implementations. The approach includes a lifecycle model, which extends conventional spiral model by formal data representation/management models and DSL-based "low-level" CASE tools supporting the formalisms. The methodology has been successfully implemented as a series of portal-based ERP systems in ITERA oil-and-gas corporation, and in a number of trading/banking enterprise applications for other enterprises. Semantic network-based airline dispatch system, and a 6D-model-driven nuclear power plant construction support system are currently in progress. Combining various SSDL models is discussed. Terms-and-cost reduction factors are examined. Correcting SSDL according to project size and scope is overviewed. The so-called “human factor errors” resulting from non-systematic SSDL approach, and their influencing crisis and depression, are analyzed. The ways to systematic and efficient SSDL are outlined. Troubleshooting advises are given for the problems concerned.

  20. SOFTWARE IMPLEMENTATION OF CULTURAL VIEW PROFESSIONAL SELF-DETERMINATION SENIOR PUPILS

    Directory of Open Access Journals (Sweden)

    Илья Фёдорович Исаев

    2014-04-01

    Full Text Available Professional self-determination is a public high school graduates of scientific and pedagogical problem and an important factor for socio-economic development of the country. In the process of self-affirmation and self-actualization is the individual. The mechanism for the implementation of professional self-schooler is a dialogue with them, reflecting the integration of the processes taking place in the cultural and social spaces. This raises the question of creating culturogical concept of self-determination of the graduate school and its software, the purpose of which is to create a culture of self-determination of the individual school student.For the development of this software has been used provision of cultural, systemic, activity-oriented, student-centered, competency, axiological, and polysubject reflexive approaches.On the basis of their programs were developed: "Self-determination of high school students in the area of high-collar occupations", "Design strategies in life and professional career", "Culture of self-determination of schoolchildren in blue-collar occupations", "Professional career. Success in life", "The culture of self-determination in science", "Psycho-pedagogical support of senior professional self-determination" for teachers, heads of educational institutions, students 9-11 years, undergraduate and graduate students. As an illustrative example, the article is a description of the last program. Development and application of the software implementation of the concept of self-determination culturogical school students provided the process is science-based teaching materials, increased the effectiveness of social partnership of educational institutions and universities. DOI: http://dx.doi.org/10.12731/2218-7405-2013-9-72

  1. Implementation of dose monitoring software in the clinical routine. First experiences

    Energy Technology Data Exchange (ETDEWEB)

    Heilmaier, C.; Zuber, N.; Bruijns, B.; Weishaupt, D. [Stadtspital Triemli, Zurich (Switzerland). Dept. of Radiology and Nuclear Medicine; Ceyrolle, C. [DoseWatch, GE Healthcare, Buc (France)

    2016-01-15

    Radiation exposure of the public as a result of medical imaging has significantly increased during the last decades. To have a tool to register and control patient dose exposure, we implemented dose monitoring software at our institution and first connected our computed tomography (CT) scanners. CT dose data from July 2014 to February 2015 was retrospectively analyzed using dose monitoring software. We evaluated a number of scans above predefined dose thresholds (''alerts''), assessed reasons for alerts and compared data of two CT scanners, one located close to the emergency room (''emergency CT scanner'') and one mainly used on an outpatient basis (''clinical routine CT scanner''). To check for statistically significant differences between scanners, chi-square-tests were performed. A total of 8883 scans were acquired (clinical routine CT scanner, n = 3415; emergency CT scanner, n = 5468) during which 316 alerts were encountered (alert quota, 4 %). The overall alert quota ranged from 2 - 5 % with significantly higher values for the clinical routine CT scanner. Reasons for alerts were high BMI (51 %), patient off-centering (24 %), scan repetition (11 %), orthopedic hardware (9 %), or other (5 %). Scan repetition was necessary significantly more often with the emergency CT scanner (p = 0.019), while high BMI, off-centering and orthopedic hardware were more frequently seen with the clinical routine CT scanner (for all, p < 0.05). There was a good correlation between high body weight and dose above threshold (r = 0.585). Implementation of dose monitoring software in the clinical routine was successfully accomplished and provides important information regarding patient radiation protection.

  2. New tools for digital medical image processing implemented in DIP software

    Energy Technology Data Exchange (ETDEWEB)

    Araujo, Erica A.C.; Santana, Ivan E. [Instituto Federal de Educacao, Ciencia e Tecnologia de Pernambuco, Recife, PE (Brazil); Lima, Fernando R.A., E-mail: falima@cnen.gov.b [Centro Regional de Ciencias Nucleares, (CRCN/NE-CNEN-PE), Recife, PE (Brazil); Viera, Jose W. [Escola Politecnica de Pernambuco, Recife, PE (Brazil)

    2011-07-01

    The anthropomorphic models used in computational dosimetry, also called phantoms, are mostly built from stacks of images CT (Computed Tomography) or MRI (Magnetic Resonance Imaging) obtained from scans of patients or volunteers. The construction of voxel phantoms requires computational processing for transforming image formats, dimensional image compression (2D) to form three-dimensional arrays (3D), quantization, resampling, enhancement, restoration and image segmentation, among others. Hardly the computational dosimetry researcher finds all these skills into a single software and often it results in a decreased development of their research or inadequate use of alternative tools. The need to integrate the various tasks of the original digital image processing to obtain an image that can be used in a computational model of exposure led to the development of software DIP (Digital Image Processing). This software reads, writes and edits binary files containing the 3D matrix corresponding to a stack of cross-sectional images of a given geometry that can be a human body or other volume of interest. It can also read any type of computer image and do conversions. When the task involves only one output image, it is saved in the JPEG standard Windows. When it involves a stack of images, the binary output file is called SGI (Interactive Graphic Simulations, a symbol already used in other publications of the Research Group in Numerical Dosimetry). The following paper presents the third version of the DIP software and emphasizes the new tools it implemented. Currently it has the menus Basics, Views, Spatial Domain, Frequency Domain, Segmentations and Study. Each menu contains items and subitems with features that generally require an image as input and produce an image or an attribute in the output. (author)

  3. C++ Software Quality in the ATLAS experiment: Tools and Experience

    CERN Document Server

    Martin-Haugh, Stewart; The ATLAS collaboration

    2017-01-01

    In this paper we explain how the C++ code quality is managed in ATLAS using a range of tools from compile-time through to run time testing and reflect on the substantial progress made in the last two years largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other available code analysis tools are also discussed, as is the role of unit testing with an example of how the GoogleTest framework can be applied to our codebase.

  4. C++ Software Quality in the ATLAS Experiment: Tools and Experience

    CERN Document Server

    Kluth, Stefan; The ATLAS collaboration; Obreshkov, Emil; Roe, Shaun; Seuster, Rolf; Snyder, Scott; Stewart, Graeme

    2016-01-01

    The ATLAS experiment at CERN uses about six million lines of code and currently has about 420 developers whose background is largely from physics. In this paper we explain how the C++ code quality is managed using a range of tools from compile-time through to run time testing and reflect on the great progress made in the last year largely through the use of static analysis tools such as Coverity®, an industry-standard tool which enables quality comparison with general open source C++ code. Other tools including cppcheck, Include-What-You-Use and run-time 'sanitizers' are also discussed.

  5. A data quality monitoring software framework for the BESⅢ experiment

    Institute of Scientific and Technical Information of China (English)

    HU Ji-Feng; ZHENG Yang-Heng; SUN Xiao-Dong; JI Xiao-Bin

    2012-01-01

    Data quality monitoring (DQM) plays an important role in data taking at the BESⅢ experiments.DQM is used to monitor detector status and data quality.A DQM framework (DQMF) has been developed to make it possible to reuse the BESⅢ offline reconstruction system in the online environment.In this framework,the DQMF can also simulate a virtual data taking environment,transfer events to the event display,publish histograms to a histogram presenter in a fixed interval,and dump histograms into a ROOT file.The DQMF has been stably running throughout BESⅢ data taking.

  6. Achieving dependable software through Continuous Delivery and Quality Monitoring

    CERN Document Server

    CERN. Geneva

    2015-01-01

    The idea for the presentation is to present our implementation of the CI/CD paradigms and explain on real live examples advantages and drawbacks of the current solution. During the presentation we will try to cover all the required steps which should automatically triggered by a developer’s commit. The presentation should give users a good hands-on experience on basic CI/CD principles and allow them to design and i...

  7. 77 FR 12482 - Approval and Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality...

    Science.gov (United States)

    2012-03-01

    ... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Indiana; Lead Ambient Air Quality Standards AGENCY: Environmental Protection Agency (EPA). ACTION: Direct final rule... Clean Air Act (CAA). This submittal incorporates the National Ambient Air Quality Standards (NAAQS) for...

  8. 77 FR 52277 - Approval of Air Quality Implementation Plans; California; South Coast Air Quality Management...

    Science.gov (United States)

    2012-08-29

    ... submitted for the South Coast Air Quality Management District (District) portion of the California State... AGENCY 40 CFR Part 52 Approval of Air Quality Implementation Plans; California; South Coast Air Quality Management District; Prevention of Significant Deterioration; Greenhouse Gases AGENCY: Environmental...

  9. Software quality and process improvement in scientific simulation codes

    Energy Technology Data Exchange (ETDEWEB)

    Ambrosiano, J.; Webster, R. [Los Alamos National Lab., NM (United States)

    1997-11-01

    This report contains viewgraphs on the quest to develope better simulation code quality through process modeling and improvement. This study is based on the experience of the authors and interviews with ten subjects chosen from simulation code development teams at LANL. This study is descriptive rather than scientific.

  10. Implementation of Quality Management System in Sweco Norge

    OpenAIRE

    Hansen, Hedda Heier

    2013-01-01

    Good and appropriate quality management is a necessity for achieving competitive advantage within most industries, including the engineering consulting industry. In this case study factors contributing to or hindering quality management system implementation success in Sweco Norge AS have been examined. The study has shown that the benefits that the organization wishes to achieve by system use, how the system is designed to achieve those objectives and how different mechanisms that promote sy...

  11. Saasquality - A Method for Quality Evaluation of Software as a Service (Saas

    Directory of Open Access Journals (Sweden)

    Nemesio Freitas Duarte Filho

    2013-07-01

    Full Text Available The market for software products offered as a service (SaaS is growing steadily and has attractedsuppliers from different segments of the global IT market. However, the use of the SaaS products brings arange of challenges,both in the organizational, cultural and technological areas. A difficulty that existstoday is the lack of methods and models for assessing the quality of these products. This document presentsa method to assess the quality of a software product offeredas a service, named SaaSQuality. Theproposed method has a quality model appropriate to the SaaS context, based on standards and models ofsoftware quality (ISO 9126 and models for IT management (ITIL and COBIT. The experimental resultsobtained througha case study show that the method offers suitable assessment practices for Software as aService.

  12. Introductory Molecular Orbital Theory: An Honors General Chemistry Computational Lab as Implemented Using Three-Dimensional Modeling Software

    Science.gov (United States)

    Ruddick, Kristie R.; Parrill, Abby L.; Petersen, Richard L.

    2012-01-01

    In this study, a computational molecular orbital theory experiment was implemented in a first-semester honors general chemistry course. Students used the GAMESS (General Atomic and Molecular Electronic Structure System) quantum mechanical software (as implemented in ChemBio3D) to optimize the geometry for various small molecules. Extended Huckel…

  13. Total Quality Management Implementation Plan: DLA-N

    Science.gov (United States)

    1989-07-01

    e Wastimto , n. Othe 20 Seato3 4. TITLE AND SUBTITLE S. FUNDING NUMBERS DLA-N Total Quality Management 6. AUTHOR(S) 7. PERFORMING ORGANIZATION NAME(S...PAGES TQM ( Total Quality Management ), Continuous Process Improvement.(; , Defense National Stockpile 16. PRICE CODEI17. SECURITY CLASSIFICATION 18...IUNCLASSIFIED UL NSN 7540-01-280-5500 Standard Form 298 (Rev. 2-89) pr"!Cbed ty ANSI Std Z39’B6 296-102 DLA - N TOTAL QUALITY MANAGEMENT IMPLEMENTATION PLAN I

  14. Building better guidelines with BRIDGE-Wiz: development and evaluation of a software assistant to promote clarity, transparency, and implementability

    Science.gov (United States)

    Michel, George; Rosenfeld, Richard M; Davidson, Caryn

    2011-01-01

    Objective To demonstrate the feasibility of capturing the knowledge required to create guideline recommendations in a systematic, structured, manner using a software assistant. Practice guidelines constitute an important modality that can reduce the delivery of inappropriate care and support the introduction of new knowledge into clinical practice. However, many guideline recommendations are vague and underspecified, lack any linkage to supporting evidence or documentation of how they were developed, and prove to be difficult to transform into systems that influence the behavior of care providers. Methods The BRIDGE-Wiz application (Building Recommendations In a Developer's Guideline Editor) uses a wizard approach to address the questions: (1) under what circumstances? (2) who? (3) ought (with what level of obligation?) (4) to do what? (5) to whom? (6) how and why? Controlled natural language was applied to create and populate a template for recommendation statements. Results The application was used by five national panels to develop guidelines. In general, panelists agreed that the software helped to formalize a process for authoring guideline recommendations and deemed the application usable and useful. Discussion Use of BRIDGE-Wiz promotes clarity of recommendations by limiting verb choices, building active voice recommendations, incorporating decidability and executability checks, and limiting Boolean connectors. It enhances transparency by incorporating systematic appraisal of evidence quality, benefits, and harms. BRIDGE-Wiz promotes implementability by providing a pseudocode rule, suggesting deontic modals, and limiting the use of ‘consider’. Conclusion Users found that BRIDGE-Wiz facilitates the development of clear, transparent, and implementable guideline recommendations. PMID:21846779

  15. Implementing Quality Service-Learning Programs in Community Colleges

    Science.gov (United States)

    Vaknin, Lauren Weiner; Bresciani, Marilee J.

    2013-01-01

    This cross-case comparative study at Western Community College and the University of the Coast explored through a constructive lens the characteristics that lead to sustainable, high quality service-learning programs and how they are implemented at institutions of higher education. The researchers determined that both Western Community College and…

  16. Implementing Quality Service-Learning Programs in Community Colleges

    Science.gov (United States)

    Vaknin, Lauren Weiner; Bresciani, Marilee J.

    2013-01-01

    This cross-case comparative study at Western Community College and the University of the Coast explored through a constructive lens the characteristics that lead to sustainable, high quality service-learning programs and how they are implemented at institutions of higher education. The researchers determined that both Western Community College and…

  17. Determinants of quality management systems implementation in hospitals

    NARCIS (Netherlands)

    Wardhani, Viera; Utarini, Adi; van Dijk, Jitse Pieter; Post, Doeke; Groothoff, Johan Willem

    2009-01-01

    Objective: To identify the problems and facilitating factors in the implementation of quality management system (QMS) in hospitals through a systematic review. Method: A search strategy was pet-formed on the Medline database for articles written in English published between 1992 and early 2006. Usin

  18. A New Model for Software Engineering Systems Quality Improvement

    Directory of Open Access Journals (Sweden)

    Ahmad A. Al-Rababah

    2014-04-01

    Full Text Available In the continuing effort to improve the system analysis and design process, several different approaches have been developed. This study will propose a new process methodology solves some problems in traditional system development methodologies it will study the strength and limitation of existing system development methodologies from traditional waterfall to iterative model including (Prototyping, Spiral, Rapid Application Development, XP and RUP to Agility. Propose a new methodology focus on produce a high quality product and suitable for all kind of project. Compare the new methodology with others to view some features that is differentiating it from previous methodologies.

  19. 企业ERP软件实施的能力成熟度研究%Enterprise ERP Software Implementation Capability Maturity Studies

    Institute of Scientific and Technical Information of China (English)

    吴鹏

    2011-01-01

    通过总结实地调查30多家企业ERP实施的成功因素和失败教训,借鉴国内外ERP实施的成功因素法,把软件开发过程的能力成熟度模型SW-CMM引入到我国企业实施ERP的实施能力测评中,提供企业一个基于过程改进的"ERP实施的能力成熟度模型",对此进行了论述。%Due to the implementation of ERP application enterprise involved in the existing management mode and its experience with ERP demand,ERP software quality,ERP software vendors internal management,ERP implementation service and management consulting to suppo

  20. Space Shuttle Program Primary Avionics Software System (PASS) Success Legacy - Quality and Reliability Date

    Science.gov (United States)

    Orr, James K.; Peltier, Daryl

    2010-01-01

    Thsi slide presentation reviews the avionics software system on board the space shuttle, with particular emphasis on the quality and reliability. The Primary Avionics Software System (PASS) provides automatic and fly-by-wire control of critical shuttle systems which executes in redundant computers. Charts given show the number of space shuttle flights vs time, PASS's development history, and other charts that point to the reliability of the system's development. The reliability of the system is also compared to predicted reliability.

  1. Implementation and validation of a commercial portal dosimetry software for intensity-modulated radiation therapy pre-treatment verification

    Directory of Open Access Journals (Sweden)

    Varatharaj C

    2010-01-01

    Full Text Available Electronic portal imaging devices (EPIDs are extensively used for obtaining dosimetric information of pre-treatment field verification and in-vivo dosimetry for intensity-modulated radiotherapy (IMRT. In the present study, we have implemented the newly developed portal dosimetry software using independent dose prediction algorithm EPIDose TM and evaluated this new tool for the pre-treatment IMRT plan quality assurance of Whole Pelvis with Simultaneous Integrated Boost (WP-SIB-IMRT of prostate cases by comparing with routine two-dimensional (2D array detector system (MapCHECK TM . We have investigated 104 split fields using g-distributions in terms of predefined g frequency parameters. The mean γ values are found to be 0.42 (SD: 0.06 and 0.44 (SD: 0.06 for the EPIDose and MapCHECK TM , respectively. The average g∆ for EPIDose and MapCHECK TM are found as 0.51 (SD: 0.06 and 0.53 (SD: 0.07, respectively. Furthermore, the percentage of points with g < 1, γ < 1.5, and γ > 2 are 97.4%, 99.3%, and 0.56%, respectively for EPIDose and 96.4%, 99.0% and 0.62% for MapCHECK TM . Based on our results obtained with EPIDose and strong agreement with MapCHECK TM , we may conclude that the EPIDose portal dosimetry system has been successfully implemented and validated with our routine 2D array detector

  2. Implementation Of Code And Carrier Tracking Loops For Software GPS Receivers

    Directory of Open Access Journals (Sweden)

    Win Kay Khaing

    2015-06-01

    Full Text Available Abstract GPS is playing in very important role in our modern mobile societies. Software approach is very flexible rather than the traditional hardware receivers. The soft-GPS receiver includes two portions hardware and software. In hardware portion an antenna filter down-converter from RF Radio Frequency to IF Intermediate Frequency and an ADC Analog to Digital Converter are included. In software portion signal processing such as acquisition tracking and navigation that runs on general purpose processor is included. The GPS signal is taken from N-FUELS Full Educational Library of Signals for Navigation signal simulator. The heart of soft-GPS receiver is the synchronization processes such as acquisition and tracking. In tracking there are two main loops for code and carrier tracking. The objective of this paper is to analyse and find the optimum discriminator function for the code tracking loop in soft-GPS receivers. The delay lock loop DLL is a well-known technique to track the codes for GNSS spread spectrum systems. This paper also presents non-coherent square law DLLs and the impacts of some parameters on DLL discriminators such as number of samples per chip early-late spacing different C No values where C denotes the signal power and No is the noise spectral density and the impact of with or without front-end device. The results of discriminator outputs are illustrated by using S-curves. Testing results with the real GPS signal are also described. This optimized discriminator functions can be implemented in any soft-GPS receivers.

  3. ThermoData engine (TDE): software implementation of the dynamic data evaluation concept. 4. Chemical reactions.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Kazakov, Andrei F; Muzny, Chris D; Frenkel, Michael

    2009-12-01

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported recently in this journal. This paper describes the first application of this concept to the evaluation of thermodynamic properties for chemical reactions. Reaction properties evaluated are the enthalpies, entropies, Gibbs energies, and thermodynamic equilibrium constants. Details of key considerations in the critical evaluation of enthalpies of formation and of standard entropies for organic compounds are discussed in relation to their application in the calculation of reaction properties. Extensions to the class structure of the program are described that allow close linkage between the derived reaction properties and the underlying pure-component properties. Derivation of pure-component enthalpies of formation and of standard entropies through the use of directly measured reaction properties (enthalpies of reaction and equilibrium constants) is described. Directions for future enhancements are outlined.

  4. Measuring the impact of computer resource quality on the software development process and product

    Science.gov (United States)

    Mcgarry, Frank; Valett, Jon; Hall, Dana

    1985-01-01

    The availability and quality of computer resources during the software development process was speculated to have measurable, significant impact on the efficiency of the development process and the quality of the resulting product. Environment components such as the types of tools, machine responsiveness, and quantity of direct access storage may play a major role in the effort to produce the product and in its subsequent quality as measured by factors such as reliability and ease of maintenance. During the past six years, the NASA Goddard Space Flight Center has conducted experiments with software projects in an attempt to better understand the impact of software development methodologies, environments, and general technologies on the software process and product. Data was extracted and examined from nearly 50 software development projects. All were related to support of satellite flight dynamics ground-based computations. The relationship between computer resources and the software development process and product as exemplified by the subject NASA data was examined. Based upon the results, a number of computer resource-related implications are provided.

  5. Software for evaluating magnetic induction field generated by power lines: implementation of a new algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Comelli, M.; Benes, M.; Bampo, A.; Villalta, R. [Regional Environment Protection Agency of Friuli Venezia Giulia (ARPA FVG), Environmental Physics, Udine (Italy)

    2006-07-01

    The Regional Environment Protection Agency of Friuli Venezia Giulia (A.R.P.A. F.V.G., Italy) has performed an analysis on existing software designed to calculate magnetic induction field generated by power lines. As far as the agency requirements are concerned the tested programs display some difficulties in the immediate processing of electrical and geometrical data supplied by plant owners, and in certain cases turn out to be inadequate in representing complex configurations of power lines. Furthermore, none of them is preset for cyclic calculus to determine the time evolution of induction in a certain exposure area. Finally, the output data are not immediately importable by ArcView, the G.I.S. used by A.R.P.A. F.V.G., and it is not always possible to implement the territory orography to determine the field at specified heights above the ground. P.h.i.d.e.l., an innovative software, tackles and works out al l the above mentioned problems. The power line wires interested in its implementation are represented by poly lines, and the field is analytically calculated, with no further approximation, not even when more power lines are concerned. Therefore, the obtained results, when compared with those of other programs, are the closest to experimental measurements. The output data can be employed both in G.I.S. and Excel environments, allowing the immediate overlaying of digital cartography and the determining of the 3 and 10 {mu}T bands, in compliance with the Italian Decree of the President of the Council of Ministers of 8 July 2003. (authors)

  6. Implementation Of Quality Management System For Irradiation Processing Services

    Science.gov (United States)

    Lungu, Ion-Bogdan; Manea, Maria-Mihaela

    2015-07-01

    In today's market, due to an increasing competitiveness, quality management has set itself as an indispensable tool and a reference point for every business. It is ultimately focused on customer satisfaction which is a stringent factor for every business. Implementing and maintaining a QMS is a rather difficult, time consuming and expensive process which must be done with respect of many factors. The aim of this paper is to present a case study for implementing QMS ISO 9001 in a gamma irradiation treatment service provider. The research goals are the identification of key benefits, reasons, advantages, disadvantages, drawbacks etc for a successful QMS implementation and use. Finally, the expected results focus on creating a general framework for implementing an efficient QMS plan that can be easily adapted to other kind of services and markets.

  7. ALGORITHM OF PLACEMENT OF VIDEO SURVEILLANCE CAMERAS AND ITS SOFTWARE IMPLEMENTATION

    Directory of Open Access Journals (Sweden)

    Loktev Alexey Alexeevich

    2012-10-01

    Full Text Available Comprehensive distributed safety, control, and monitoring systems applied by companies and organizations of different ownership structure play a substantial role in the present-day society. Video surveillance elements that ensure image processing and decision making in automated or automatic modes are the essential components of new systems. This paper covers the modeling of video surveillance systems installed in buildings, and the algorithm, or pattern, of video camera placement with due account for nearly all characteristics of buildings, detection and recognition facilities, and cameras themselves. This algorithm will be subsequently implemented as a user application. The project contemplates a comprehensive approach to the automatic placement of cameras that take account of their mutual positioning and compatibility of tasks. The project objective is to develop the principal elements of the algorithm of recognition of a moving object to be detected by several cameras. The image obtained by different cameras will be processed. Parameters of motion are to be identified to develop a table of possible options of routes. The implementation of the recognition algorithm represents an independent research project to be covered by a different article. This project consists in the assessment of the degree of complexity of an algorithm of camera placement designated for identification of cases of inaccurate algorithm implementation, as well as in the formulation of supplementary requirements and input data by means of intercrossing sectors covered by neighbouring cameras. The project also contemplates identification of potential problems in the course of development of a physical security and monitoring system at the stage of the project design development and testing. The camera placement algorithm has been implemented as a software application that has already been pilot tested on buildings and inside premises that have irregular dimensions. The

  8. Improving Code Quality of the Compact Muon Solenoid Electromagnetic Calorimeter Control Software to Increase System Maintainability

    CERN Multimedia

    Holme, Oliver; Dissertori, Günther; Djambazov, Lubomir; Lustermann, Werner; Zelepoukine, Serguei

    2013-01-01

    The Detector Control System (DCS) software of the Electromagnetic Calorimeter (ECAL) of the Compact Muon Solenoid (CMS) experiment at CERN is designed primarily to enable safe and efficient operation of the detector during Large Hadron Collider (LHC) data-taking periods. Through a manual analysis of the code and the adoption of ConQAT [1], a software quality assessment toolkit, the CMS ECAL DCS team has made significant progress in reducing complexity and improving code quality, with observable results in terms of a reduction in the effort dedicated to software maintenance. This paper explains the methodology followed, including the motivation to adopt ConQAT, the specific details of how this toolkit was used and the outcomes that have been achieved. [1] ConQAT, Continuous Quality Assessment Toolkit; https://www.conqat.org/

  9. Software Quality Validation for Web Applications Developed Using Geographically Distributed Human Resources

    Directory of Open Access Journals (Sweden)

    Mihai GHEORGHE

    2015-01-01

    Full Text Available Developing web applications using Geographically Distributed Team Members has seen an increased popularity during the last years mainly because the rise of Open Source technologies, fast penetration of the Internet in emerging economies, the continuous quest for reduced costs as well for the fast adoption of online platforms and services which successfully address project planning, coordination and other development tasks. This paper identifies general software process stages for both collocated and distributed development and analyses the impact the use of planning, management and testing online services has on the duration, cost and quality of each stage. Given that Quality Assurance is one of the most important concerns in Geographically Distributed Software Development (GDSD, the focus is on Software Quality Validation.

  10. The challenge of quality for Social Third Sector. Cases analysis of EFQM model implementation

    Directory of Open Access Journals (Sweden)

    Emma Juaneda Ayensa

    2013-05-01

    Full Text Available Nowadays, the difficulties of the Governments to maintain the Welfare Systems have promoted the private management in the welfare services sector. The Third Sector organizations, as welfare services suppliers, have to adapt to the changing social demands and try to encourage their participation in the system configuration but being able to demonstrate the efficiency with which they operate. Faced with this challenge, Quality Management Systems offer an opportunity for the Third Sector Organizations, although with some reluctance due to the industrial origin of the model.The main aim of this work is to study the implementation process of a Total Quality Management system in the Third Social Sector organizations. Our research has been done through the analysis of multiple cases, selecting four organizations in different stages of quality management implementation. The data analysis was done using qualitative analysis software ATLAS.ti and for the evaluation of each case was done following the EFQM assessment system software to analyze the different elements of the model enablers. The main results are the relevant meaning of alliance development, the user focus as well as the person focus, elements system which the organizations consider as key issues. This research work shows the process for four organizations but suggest the necessity to research deeply about relationship among the enablers as well as a causal analysis as future lines.

  11. ISO 17025: practical benefits of implementing a quality system.

    Science.gov (United States)

    Honsa, Julie D; McIntyre, Deborah A

    2003-01-01

    As a laboratory certified to ISO 9001:2000 and accredited to ISO 17025, rtech laboratories has incorporated an overall system for technical and quality management, which results in benefits observed in daily laboratory practices. Technical requirements were updated to include the addition of formal personnel training plans and detailed records, method development and validation procedures, measurement of method uncertainty, and a defined equipment calibration and maintenance program. In addition, a stronger definition of the sample preparation process was documented to maintain consistency in sampling, and a more rigorous quality control monitoring program was implemented for chemistry and microbiology. Management quality improvements focused on document control to maintain consistent analytical processes, improved monitoring of supplier performance, a contract review process for documenting customer requirements, and a system for handling customer comments and complaints, with continuous improvement through corrective and preventive action procedures and audits. Quarterly management review of corrective actions, nonconforming testing, and proficiency testing aid in determining long-term trending. The practical benefits of these technical and management quality improvements are seen on a daily basis in the laboratory. Faster identification and resolution of issues regarding methods, personnel or equipment, improved customer satisfaction, meeting quality requirements of specialized customers, and overall increased laboratory business are all the result of implementing an effective quality system.

  12. COST OF QUALITY MODELS AND THEIR IMPLEMENTATION IN MANUFACTURING FIRMS

    Directory of Open Access Journals (Sweden)

    N.M. Vaxevanidis

    2009-03-01

    Full Text Available In order to improve quality, an organization must take into account the costs associated with achieving quality since the objective of continuous improvement programs is not only to meet customer requirements, but also to do it at the lowest, possible, cost. This can only obtained by reducing the costs needed to achieve quality, and the reduction of these costs is only possible if they are identified and measured. Therefore, measuring and reporting the cost of quality (CoQ should be considered an important issue for achieving quality excellence. To collect quality costs an organization needs to adopt a framework to classify costs; however, there is no general agreement on a single broad definition of quality costs. CoQ is usually understood as the sum of conformance plus non-conformance costs, where cost of conformance is the price paid for prevention of poor quality (for example, inspection and quality appraisal and cost of non-conformance is the cost of poor quality caused by product and service failure (for example, rework and returns. The objective of this paper is to give a survey of research articles on the topic of CoQ; it opens with a literature review focused on existing CoQ models; then, it briefly presents the most common CoQ parameters and the metrics (indices used for monitoring CoQ. Finally, the use of CoQ models in practice, i.e., the implementation of a quality costing system and cost of quality reporting in companies is discussed, with emphasis in cases concerning manufacturing firms.

  13. Lightweight and Continuous Architectural Software Quality Assurance using the aSQA Technique

    DEFF Research Database (Denmark)

    Christensen, Henrik Bærbak; Hansen, Klaus Marius; Lindstrøm, Bo

    2010-01-01

    In this paper, we present a novel technique for assessing and prioritizing architectural quality in large-scale software development projects. The technique can be applied with relatively little effort by software architects and thus suited for agile development in which quality attributes can...... be assessed and prioritized, e.g., within each development sprint. We outline the processes and metrics embodied in the technique, and report initial experiences on the benefits and liabilities. In conclusion, the technique is considered valuable and a viable tool, and has benefits in an architectural......, technical, context, as well as in a business and people context....

  14. Total Quality Management Implementation and Guest Satisfaction in Hospitality

    Directory of Open Access Journals (Sweden)

    Miroslav Knežević

    2017-02-01

    Full Text Available Total quality management (TQM has become a modern system of constant improvement of the quality of all company activities. The purpose of this study is to measure the expectations and satisfaction of the guests concerning the attribute quality of the hotel product. Furthermore obtained results were compared in such a way as to analyse particularly the reviews of hotels which have implemented TQM and have the ISO 9001 certificates with reviews from hotels which have not implemented TQM and do not have the ISO 9001 certificates. The conducted analysis included 55 hotels in Serbia belonging to the 4- and 5-star categories, i.e. 1308 guests who have stayed in them. The results show that between the observed groups of guests there are fewer differences in expectations than in perception, and that generally speaking guests who have stayed in the hotels that have implemented TQM are more satisfied. The biggest difference concerning the guest satisfaction with the quality of service in the observed hotels is noticeable in relation to the employees and the value-for-money.

  15. Quality assurance for CORAL and COOL within the LCG software stack for the LHC experiments

    CERN Document Server

    CERN. Geneva

    2015-01-01

    CORAL and COOL are software packages used by the LHC experiments for managing different categories of physics data using a variety of relational database technologies. The core components are written in C++, but Python bindings are also provided. CORAL is a generic relational access layer, while COOL includes the implementation of a specific relational data model and optimization of SQL queries for "conditions data". The software is the result of more than 10 years of development in colaboration between the IT department and the LHC experiments. The packages are built and released within the LCG software stack, for which automatic nightly builds and release installations are provided by PH-SFT (cmake, jenkins, cdash) for many different platforms, compilers and software version configurations. Test-driven development and functional tests of both C++ and Python components (CppUnit, unittest) have been key elements in the success of the projects. Dedicated test suites have also been prepared to commission and ma...

  16. Some key principles in implementing knowledge management: the views of employees in a small software company

    Directory of Open Access Journals (Sweden)

    Clair Wagner

    2002-09-01

    Full Text Available As the Knowledge Era replaces the Information Age the challenge is to transform information into knowledge and use this knowledge to create and sustain the competitive advantage of organisations. In this article four key principles for implementing knowledge management are discussed. These principles were derived from a survey and discussions with employees at a small software company. The four principles are: the need for a common definition of and vision for knowledge management, the interdependence between the human and technology track, the way employees punctuate ‘knowing’ and the structure of a reward system for sharing knowledge. Opsomming Organisasies word gekonfronteer met unieke uitdagings soos wat die Kennisera die Inligtingsera vervang. Een van die grootste uitdagings is om inligting te verander na betekenisvolle kennis en om die kennis te gebruik om ‘n kompeterende voorsprong te skep en te bou. In hierdie artikel word vier kern beginsels vir die implementering van kennisbestuur bespreek. Die beginsels is geskoei op ‘n opname van en onderhoude met werknemers in ‘n klein sagteware besigheid. Die vier beginsels is: die behoefte aan ‘n definisie en visie van kennisbestuur, die interafhanklikheid tussen mense en tegnologie, die wyse waarop werknemers kennis punktueer en die skep van ‘n beloningstruktuur vir die deel van kennis.

  17. Implementation of New Process Models for Tailored Polymer Composite Structures into Processing Software Packages

    Energy Technology Data Exchange (ETDEWEB)

    Nguyen, Ba Nghiep; Jin, Xiaoshi; Wang, Jin; Phelps, Jay; Tucker III, Charles L.; Kunc, Vlastimil; Bapanapalli, Satish K.; Smith, Mark T.

    2010-02-23

    This report describes the work conducted under the Cooperative Research and Development Agreement (CRADA) (Nr. 260) between the Pacific Northwest National Laboratory (PNNL) and Autodesk, Inc. to develop and implement process models for injection-molded long-fiber thermoplastics (LFTs) in processing software packages. The structure of this report is organized as follows. After the Introduction Section (Section 1), Section 2 summarizes the current fiber orientation models developed for injection-molded short-fiber thermoplastics (SFTs). Section 3 provides an assessment of these models to determine their capabilities and limitations, and the developments needed for injection-molded LFTs. Section 4 then focuses on the development of a new fiber orientation model for LFTs. This model is termed the anisotropic rotary diffusion - reduced strain closure (ARD-RSC) model as it explores the concept of anisotropic rotary diffusion to capture the fiber-fiber interaction in long-fiber suspensions and uses the reduced strain closure method of Wang et al. to slow down the orientation kinetics in concentrated suspensions. In contrast to fiber orientation modeling, before this project, no standard model was developed to predict the fiber length distribution in molded fiber composites. Section 5 is therefore devoted to the development of a fiber length attrition model in the mold. Sections 6 and 7 address the implementations of the models in AMI, and the conclusions drawn from this work is presented in Section 8.

  18. Implementation of Quality Management System for Historical Building Conservation

    Directory of Open Access Journals (Sweden)

    Zahari N.F.

    2014-01-01

    Full Text Available The main objectives of this study are twofold. Firstly, to identify the implementation of ISO 9001 procedures being used as references for conservation works and the development of Quality Management System (QMS guidelines. Data were solicited from three (3 conservation areas. The analysis involved of descriptive approach and statistical methods. The findings revealed that QMS is not structurally established, implemented and enforced as part of conservation practice in Malaysia. From the findings, the authors hope to give clear perception to the reader on current preservation practice and the existence of QMS with reference to ISO 9001 for future conservation mechanism.

  19. Perceptron system to assist in decision making and monitoring of quality of software development in Information Technology environments

    Directory of Open Access Journals (Sweden)

    Chau Sen Shia

    2016-07-01

    Full Text Available - Classical methods for software development require high costs and problems of communication between development teams, project risks and delays in the delivery of its services. In this work, the aim is to develop a perceptron model to demonstrate the ability to control, service quality assessment and decision-making in IT (Information Technology. For this, we used perceptron network techniques, standards of service quality measures and risk analysis model of applied projects in software engineering. With the network perceptron model implementation was possible to simulate the application of development in several requests for applications for software, in order to meet the management of schedules in all phases of the life cycle of the projects carried out. The tests with the perceptron model were applied in it environments to meet service requests from various fields. The results and analyses presented in these projects demonstrate that communication between development teams were more consistent. It was also possible to predict with more accuracy the delivery of services, decision making and risk reduction projects.

  20. 75 FR 40726 - Revisions to the California State Implementation Plan, Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2010-07-14

    ... Quality Management District (SCAQMD) portions of the California State Implementation Plan (SIP). These... AGENCY 40 CFR Part 52 Revisions to the California State Implementation Plan, Sacramento Metropolitan Air Quality Management District and South Coast Air Quality Management District AGENCY: Environmental...

  1. Software: our quest for excellence. Honoring 50 years of software history, progress, and process

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-06-01

    The Software Quality Forum was established by the Software Quality Assurance (SQA) Subcommittee, which serves as a technical advisory group on software engineering and quality initiatives and issues for DOE`s quality managers. The forum serves as an opportunity for all those involved in implementing SQA programs to meet and share ideas and concerns. Participation from managers, quality engineers, and software professionals provides an ideal environment for identifying and discussing issues and concerns. The interaction provided by the forum contributes to the realization of a shared goal--high quality software product. Topics include: testing, software measurement, software surety, software reliability, SQA practices, assessments, software process improvement, certification and licensing of software professionals, CASE tools, software project management, inspections, and management`s role in ensuring SQA. The bulk of this document consists of vugraphs. Selected papers have been indexed separately for inclusion in the Energy Science and Technology Database.

  2. Digital radiography: optimization of image quality and dose using multi-frequency software

    Energy Technology Data Exchange (ETDEWEB)

    Precht, H. [University College Lillebelt, Conrad Research Center, Odense (Denmark); Gerke, O. [Odense University Hospital, Department of Nuclear Medicine, Odense (Denmark); University of Southern Denmark, Research Unit of Health Economics, Odense (Denmark); Rosendahl, K. [Haukeland University Hospital, Section of Pediatric Radiology, Bergen (Norway); University of Bergen, Institute of Surgical Sciences, Bergen (Norway); Tingberg, A. [Skaane University Hospital, Lund University (Sweden); Medical Radiation Physics, Department of Clinical Sciences, Malmoe (Sweden); Waaler, D. [Gjoevik University College, Gjoevik (Norway)

    2012-09-15

    New developments in processing of digital radiographs (DR), including multi-frequency processing (MFP), allow optimization of image quality and radiation dose. This is particularly promising in children as they are believed to be more sensitive to ionizing radiation than adults. To examine whether the use of MFP software reduces the radiation dose without compromising quality at DR of the femur in 5-year-old-equivalent anthropomorphic and technical phantoms. A total of 110 images of an anthropomorphic phantom were imaged on a DR system (Canon DR with CXDI-50 C detector and MLT[S] software) and analyzed by three pediatric radiologists using Visual Grading Analysis. In addition, 3,500 images taken of a technical contrast-detail phantom (CDRAD 2.0) provide an objective image-quality assessment. Optimal image-quality was maintained at a dose reduction of 61% with MLT(S) optimized images. Even for images of diagnostic quality, MLT(S) provided a dose reduction of 88% as compared to the reference image. Software impact on image quality was found significant for dose (mAs), dynamic range dark region and frequency band. By optimizing image processing parameters, a significant dose reduction is possible without significant loss of image quality. (orig.)

  3. Flexible Implementation of Multiphysics and Discretizations in PyLith Crustal Deformation Modeling Software

    Science.gov (United States)

    Aagaard, B.; Knepley, M.; Williams, C. A.

    2016-12-01

    We are creating a flexible implementation of multiphysics and finite-element discretizations in PyLith, a community, open-source code (http://geodynamics.org/cig/software/pylith/) for modeling quasi-static and dynamic crustal deformation with an emphasis on earthquake faulting. The goals include expanding the current suite of elastic, viscoelastic, and elastoplastic bulk rheologies to include poroelasticity, thermoelasticity, and incompressible elasticity. We cast the governing equations in a form that involves the product of the finite-element basis function or its derivatives with pointwise functions that look very much like the strong form of the governing equation. This allows the finite-element integration to be decomposed into a routine for the numerical integration over cells and boundaries of the finite-element mesh and simple routines implementing the physics (pointwise functions). The finite-element integration routine works in any spatial dimension with an arbitrary number of physical fields (e.g., displacement, temperature, and fluid pressure). It also makes it much easier optimize the finite-element integrations for proper vectorization, tiling, and other traversal optimization on multiple architectures (e.g., CUDA and OpenCL) independent of the pointwise functions. Users can easily extend the code by adding new routines for the pointwise functions to implement different rheologies and/or governing equations. Tight integration with the Portable, Extensible Toolkit for Scientific Computation (PETSc) provides support for a wide range of linear and nonlinear solvers and time-stepping algorithms so that a wide variety of governing equations can be solved efficiently.

  4. An efficient implementation of reconfigurable interpolation rootraised cosine FIR filter for software-defined radio application

    Indian Academy of Sciences (India)

    S C PRASANNA; S P JOY VASANTHA RANI

    2016-10-01

    This brief presents an efficient binary common subexpression elimination (BCSE)-based approach for designing reconfigurable interpolation root-raised cosine (RRC) finite-impulse-response (FIR) filter, whose coefficients change during runtime for multistandard wireless communication system called software-defined radio (SDR). Reconfiguration can be done conveniently by storing the coded coefficients in the lookup tables (LUTs), and loading the required coefficient set over the interpolation filter. In the proposed method based on 4-bit BCSE algorithm, first the number of binary common subexpressions (BCSs) formed in the coefficients is reduced. Hence, multiplexers, shifters, and adders in the multiplier structure are reduced, which results in theimprovement of operating frequency. The number of addition operations is further reduced using programmable adders and an efficient polyphase interpolation structure is implemented to reduce the hardware cost. The proposed design has 49.5% less area-delay product and 28.6% improved frequency of operation when comparedto a 2-bit BCSE-based technique reported earlier when implemented on Xilinx field-programmable gate array (FPGA) device XC2VP4FF672-6. Similarly, the proposed design supports 93.14 MHz operating frequency, which is 59.2% and 74.2% greater when compared to 2-bit BCSE- and 3-bit BCSE-based approach when implemented on XC2V3000FF1152-4. The proposed structure also shows improved performance in terms of speed and area when compared to distributed arithmetic (DA)-based and multiply-accumulate (MAC)-based approaches.

  5. Release Process on Quality Improvement in Open Source Software Project Management

    Directory of Open Access Journals (Sweden)

    S. Chandra Kumar Mangalam

    2012-01-01

    Full Text Available Problem statement: The Software Industry has changed and developed as a consequence of the impact of Open Source Software (OSS since 1990s. Over a period of time, OSS has evolved in an integrated manner and most of the participants in OSS activity are volunteers. Approach: This coordination form of development has produced a considerable quantity of software; and often, the development method has been viewed as an unorganized and unstructured method of development. Few existing researches deal with the Open Source Software phenomenon from a quality perception point of view and studies where enhancements are possible in the development process. Results: Release Process in OSS plays a key role in most of the OSS projects. As this process is related to the evolution of a quality software from the community of OSS developers, this research attempts to explore the process practices which are employed by OSS developers and examines the problems associated with the development process. The scope of the study is mainly confined to process management in OSS. “Prototype development and iterative development process” approaches were adapted as a methodology. Conclusion/Recommendations: The major finding and conclusion drawn is ‘lack of coordination among developers’ who are geographically isolated. Hence, the study suggests the need for coordination among developers to line up their development process for achieving the goal of the software release process.

  6. Food Safety and the Implementation of Quality System in Food

    Directory of Open Access Journals (Sweden)

    Noveria Sjafrina

    2013-01-01

    Full Text Available One of the goals the development of the food sector in Indonesia is food secured the release of which is characterized by the type of food that are harmful to health. In some way of avoiding the kind of food that is harmful to health, strengthen institutional food sector, and increase the number of food industry comply with regulations. Implementation of Good Handling Pratice (GHP and Good Manufacturing Pratice (GMP and Hazard Analysis Critical Control Point (HACCP are a responsibility and awareness of manufacturers and distributors. Some of the food safety regulations have been issued by the government of Indonesia such as Law. 18 on Food in 2012 to provide protection to consumers and food producers will be healthy, safe and lawful. Development of quality systems and food safety and implementation quality system in the food industry are a shared responsibility between government, industry includes producers of raw materials, food industry and distributors, and consumers.

  7. IMPLEMENTATION OF INTEGRAL SYSTEM OF QUALITY MANAGEMENT IN TOURISM

    Directory of Open Access Journals (Sweden)

    Vidoje Vujic

    2007-12-01

    Full Text Available Theory and practice have corroborated the need and usefulness of an integral approach to quality management in tourist destinations, since the destination and its touristic offer define a number of disparate participants and interested parties. An integral system of quality management is one of the models of touristic management that with the implementation of contemporary principles strives to achieve business excellence and competitory advantage. The paper determines and projects the need and importance of implementing an integral system and accordingly seeks to form a model for its development. By studying the oneness of the whole we have established the dependence and firm connections between particular norms and elements, and by analyzing them in this paper we describe the structure and associated characteristics of the whole.

  8. A systematic literature review of open source software quality assessment models.

    Science.gov (United States)

    Adewumi, Adewole; Misra, Sanjay; Omoregbe, Nicholas; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    Many open source software (OSS) quality assessment models are proposed and available in the literature. However, there is little or no adoption of these models in practice. In order to guide the formulation of newer models so they can be acceptable by practitioners, there is need for clear discrimination of the existing models based on their specific properties. Based on this, the aim of this study is to perform a systematic literature review to investigate the properties of the existing OSS quality assessment models by classifying them with respect to their quality characteristics, the methodology they use for assessment, and their domain of application so as to guide the formulation and development of newer models. Searches in IEEE Xplore, ACM, Science Direct, Springer and Google Search is performed so as to retrieve all relevant primary studies in this regard. Journal and conference papers between the year 2003 and 2015 were considered since the first known OSS quality model emerged in 2003. A total of 19 OSS quality assessment model papers were selected. To select these models we have developed assessment criteria to evaluate the quality of the existing studies. Quality assessment models are classified into five categories based on the quality characteristics they possess namely: single-attribute, rounded category, community-only attribute, non-community attribute as well as the non-quality in use models. Our study reflects that software selection based on hierarchical structures is found to be the most popular selection method in the existing OSS quality assessment models. Furthermore, we found that majority (47%) of the existing models do not specify any domain of application. In conclusion, our study will be a valuable contribution to the community and helps the quality assessment model developers in formulating newer models and also to the practitioners (software evaluators) in selecting suitable OSS in the midst of alternatives.

  9. Reliability of adaptive multivariate software sensors for sewer water quality monitoring

    DEFF Research Database (Denmark)

    Vezzaro, Luca; Mikkelsen, Peter Steen

    2015-01-01

    mainly originated from wastewater sources (such as ammonia) than when used for pollutants affected by several processes (such as TSS). Overall, this study provides a first insight in the application of multivariate methods for software sensors, highlighting drawback and potential development areas......This study investigates the use of a multivariate approach, based on Principal Component Analysis PCA), as software sensor for fault detection and reconstruction of missing measurements in on-line monitoring of sewer water quality. The analysis was carried out on a 16-months dataset of five...... validation. However, the study also showed a dramatic drop in predictive capability of the software sensor when used for reconstructing missing values, with performance quickly deteriorating after 1 week since parameter estimation. The software sensor provided better results when used to estimate pollutants...

  10. Assessing climate model software quality: a defect density analysis of three models

    Directory of Open Access Journals (Sweden)

    J. Pipitone

    2012-08-01

    Full Text Available A climate model is an executable theory of the climate; the model encapsulates climatological theories in software so that they can be simulated and their implications investigated. Thus, in order to trust a climate model, one must trust that the software it is built from is built correctly. Our study explores the nature of software quality in the context of climate modelling. We performed an analysis of defect reports and defect fixes in several versions of leading global climate models by collecting defect data from bug tracking systems and version control repository comments. We found that the climate models all have very low defect densities compared to well-known, similarly sized open-source projects. We discuss the implications of our findings for the assessment of climate model software trustworthiness.

  11. Secure software practices among Malaysian software practitioners: An exploratory study

    Science.gov (United States)

    Mohamed, Shafinah Farvin Packeer; Baharom, Fauziah; Deraman, Aziz; Yahya, Jamaiah; Mohd, Haslina

    2016-08-01

    Secure software practices is increasingly gaining much importance among software practitioners and researchers due to the rise of computer crimes in the software industry. It has become as one of the determinant factors for producing high quality software. Even though its importance has been revealed, its current practice in the software industry is still scarce, particularly in Malaysia. Thus, an exploratory study is conducted among software practitioners in Malaysia to study their experiences and practices in the real-world projects. This paper discusses the findings from the study, which involved 93 software practitioners. Structured questionnaire is utilized for data collection purpose whilst statistical methods such as frequency, mean, and cross tabulation are used for data analysis. Outcomes from this study reveal that software practitioners are becoming increasingly aware on the importance of secure software practices, however, they lack of appropriate implementation, which could affect the quality of produced software.

  12. ThermoData Engine (TDE): software implementation of the dynamic data evaluation concept. 9. Extensible thermodynamic constraints for pure compounds and new model developments.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Muzny, Chris D; Kazakov, Andrei F; Kroenlein, Kenneth; Magee, Joseph W; Abdulagatov, Ilmutdin; Frenkel, Michael

    2013-12-23

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported in this journal. The present article describes the background and implementation for new additions in latest release of TDE. Advances are in the areas of program architecture and quality improvement for automatic property evaluations, particularly for pure compounds. It is shown that selection of appropriate program architecture supports improvement of the quality of the on-demand property evaluations through application of a readily extensible collection of constraints. The basis and implementation for other enhancements to TDE are described briefly. Other enhancements include the following: (1) implementation of model-validity enforcement for specific equations that can provide unphysical results if unconstrained, (2) newly refined group-contribution parameters for estimation of enthalpies of formation for pure compounds containing carbon, hydrogen, and oxygen, (3) implementation of an enhanced group-contribution method (NIST-Modified UNIFAC) in TDE for improved estimation of phase-equilibrium properties for binary mixtures, (4) tools for mutual validation of ideal-gas properties derived through statistical calculations and those derived independently through combination of experimental thermodynamic results, (5) improvements in program reliability and function that stem directly from the recent redesign of the TRC-SOURCE Data Archival System for experimental property values, and (6) implementation of the Peng-Robinson equation of state for binary mixtures, which allows for critical evaluation of mixtures involving supercritical components. Planned future developments are summarized.

  13. Nursing Leader Collaboration to Drive Quality Improvement and Implementation Science.

    Science.gov (United States)

    Ryan, Rosemary W; Harris, Karen K; Mattox, Lisa; Singh, Olivine; Camp, Melanie; Shirey, Maria R

    2015-01-01

    Nursing leadership opportunities to improve quality and align resources in health care exist. An estimated 18% of United States gross domestic product is spent on health care delivery systems that produce poor outcomes. The purpose of this article was to describe how quality improvement and implementation science initiatives enhance outcomes using nursing leadership strategies that play an integral role in aligning key colleagues to drive the collaborative process. A critical appraisal of the literature was conducted, which supports the importance of evidenced-based practice improvement, collaborative change process, and professional role of nursing leadership. Limited evidence exists related to practice strategies for nursing leaders to implement sustainable change at the unit level for successful alignment of resources. Strategies based on Rogers' Diffusion of Innovation Theory are recommended to address the gap in the literature. The strategies aim to increase meaningful knowledge or the "why," create a tipping point, and implement sustainable change starting with the end in mind. Nurse leaders are a central component for driving alignment and implementing change at the unit level. Uses of the described evidenced-based strategies have implications for nursing practice, education, and scholarship.

  14. Quality Assurance Framework Implementation Guide for Isolated Community Power Systems

    Energy Technology Data Exchange (ETDEWEB)

    Esterly, Sean R. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Baring-Gould, Edward I. [National Renewable Energy Lab. (NREL), Golden, CO (United States); Burman, Kari A. [National Renewable Energy Lab. (NREL), Golden, CO (United States)

    2017-08-15

    This implementation guide is a companion document to the 'Quality Assurance Framework for Mini-Grids' technical report. This document is intended to be used by one of the many stakeholder groups that take part in the implementation of isolated power systems. Although the QAF could be applied to a single system, it was designed primarily to be used within the context of a larger national or regional rural electrification program in which many individual systems are being installed. This guide includes a detailed overview of the Quality Assurance Framework and provides guidance focused on the implementation of the Framework from the perspective of the different stakeholders that are commonly involved in expanding energy development within specific communities or regions. For the successful long-term implementation of a specific rural electrification program using mini-grid systems, six key stakeholders have been identified that are typically engaged, each with a different set of priorities 1. Regulatory agency 2. Governmental ministry 3. System developers 4. Mini-utility 5. Investors 6. Customers/consumers. This document is broken into two distinct sections. The first focuses on the administrative processes in the development and operation of community-based mini-grid programs, while the second focuses on the process around the installation of the mini-grid project itself.

  15. Rapid algorithm prototyping and implementation for power quality measurement

    Science.gov (United States)

    Kołek, Krzysztof; Piątek, Krzysztof

    2015-12-01

    This article presents a Model-Based Design (MBD) approach to rapidly implement power quality (PQ) metering algorithms. Power supply quality is a very important aspect of modern power systems and will become even more important in future smart grids. In this case, maintaining the PQ parameters at the desired level will require efficient implementation methods of the metering algorithms. Currently, the development of new, advanced PQ metering algorithms requires new hardware with adequate computational capability and time intensive, cost-ineffective manual implementations. An alternative, considered here, is an MBD approach. The MBD approach focuses on the modelling and validation of the model by simulation, which is well-supported by a Computer-Aided Engineering (CAE) packages. This paper presents two algorithms utilized in modern PQ meters: a phase-locked loop based on an Enhanced Phase Locked Loop (EPLL), and the flicker measurement according to the IEC 61000-4-15 standard. The algorithms were chosen because of their complexity and non-trivial development. They were first modelled in the MATLAB/Simulink package, then tested and validated in a simulation environment. The models, in the form of Simulink diagrams, were next used to automatically generate C code. The code was compiled and executed in real-time on the Zynq Xilinx platform that combines a reconfigurable Field Programmable Gate Array (FPGA) with a dual-core processor. The MBD development of PQ algorithms, automatic code generation, and compilation form a rapid algorithm prototyping and implementation path for PQ measurements. The main advantage of this approach is the ability to focus on the design, validation, and testing stages while skipping over implementation issues. The code generation process renders production-ready code that can be easily used on the target hardware. This is especially important when standards for PQ measurement are in constant development, and the PQ issues in emerging smart

  16. Software architects’ experiences of quality requirements : what we know and what we do not know?

    NARCIS (Netherlands)

    Daneva, Maia; Buglione, Luigi; Herrmann, Andrea; Doerr, J.; Opdahl, A.

    2013-01-01

    [Context/motivation] Quality requirements (QRs) are a concern of both requirement engineering (RE) specialists and software architects (SAs). However, the majority of empirical studies on QRs take the RE analysts’/clients’ perspectives, and only recently very few included the SAs’ perspective. As a

  17. Reliability of adaptive multivariate software sensors for sewer water quality monitoring

    DEFF Research Database (Denmark)

    Vezzaro, Luca; Mikkelsen, Peter Steen

    2015-01-01

    This study investigates the use of a multivariate approach, based on Principal Component Analysis PCA), as software sensor for fault detection and reconstruction of missing measurements in on-line monitoring of sewer water quality. The analysis was carried out on a 16-months dataset of five...

  18. Software Quality and Testing: What DoD Can Learn from Commercial Practices

    Science.gov (United States)

    1992-08-31

    Defects, and Correction of Processes PROCESS IMPROVEMENT ...... I,- DEVELOPMENT ] STEST"ING -- [ PROCESO IMPROVEMENT 1 Figure 1. Software Quality Control...Furthermore, these goals are not necessarily mutually exclusive . Code can often satisfy user requirements (as defined) and still contain faults. The

  19. Possible Application of Quality Function Deployment in Software Systems Development in the United States Air Force

    Science.gov (United States)

    1991-12-01

    his cooperation in acquiring QFD Designer. I also wish to thank Mr Allen Chartier of the American Supplieri Institute for his help in identifying...and What Didn’t," Transactions from the Symposium on Quality Function Deployment. 305-335. Dearborn MI: ASI Press, 1989. Pressman, Roger S. Software

  20. An empirical evaluation of software quality assurance practices and challenges in a developing country: a comparison of Nigeria and Turkey.

    Science.gov (United States)

    Sowunmi, Olaperi Yeside; Misra, Sanjay; Fernandez-Sanz, Luis; Crawford, Broderick; Soto, Ricardo

    2016-01-01

    The importance of quality assurance in the software development process cannot be overemphasized because its adoption results in high reliability and easy maintenance of the software system and other software products. Software quality assurance includes different activities such as quality control, quality management, quality standards, quality planning, process standardization and improvement amongst others. The aim of this work is to further investigate the software quality assurance practices of practitioners in Nigeria. While our previous work covered areas on quality planning, adherence to standardized processes and the inherent challenges, this work has been extended to include quality control, software process improvement and international quality standard organization membership. It also makes comparison based on a similar study carried out in Turkey. The goal is to generate more robust findings that can properly support decision making by the software community. The qualitative research approach, specifically, the use of questionnaire research instruments was applied to acquire data from software practitioners. In addition to the previous results, it was observed that quality assurance practices are quite neglected and this can be the cause of low patronage. Moreover, software practitioners are neither aware of international standards organizations or the required process improvement techniques; as such their claimed standards are not aligned to those of accredited bodies, and are only limited to their local experience and knowledge, which makes it questionable. The comparison with Turkey also yielded similar findings, making the results typical of developing countries. The research instrument used was tested for internal consistency using the Cronbach's alpha, and it was proved reliable. For the software industry in developing countries to grow strong and be a viable source of external revenue, software assurance practices have to be taken seriously

  1. CLaSS Computer Literacy Software: From Design to Implementation - A Three Year Student Evaluation

    Directory of Open Access Journals (Sweden)

    Ian Cole

    2006-12-01

    Full Text Available Both computer literacy and information retrieval techniques are required to undertake studies in higher education in the United Kingdom. This paper considers the research, development and the 3-year student evaluation of a piece of learning technology in computer and information literacy (CLaSS software. Students completed a questionnaire to examine their own assessment of knowledge and competence in computer and information literacy and based on this assessment CLaSS software was created to assist nursing students with computer and information literacy. This paper draws on existing literature and applies a specific learning model to the software while considering software engineering and user-centered design methodologies. The technical processes involved in designing and creating the software are briefly considered with software development data analysis discussed. A 3-year student evaluation of the software after it's release was undertaken to consider the long-term validity and usefulness of this software with the results analysed and discussed.

  2. Design and implementation of a mobile phone locator using software defined radio

    OpenAIRE

    Larsen, Ian Paul

    2007-01-01

    This thesis presents an approach for generating, detecting, and decoding a Global System for Mobile Communications (GSM) signal using software defined radio and commodity computer hardware. Using software designed by the GNU freesoftware project as a base, standard GSM packets were transmitted and received over the air, and their arrival times detected. A method is provided to use software analysis of multiple receivers to locate an emitter based on the information received by the softwar...

  3. Implementation of quality systems by Mexican exporters of processed meat.

    Science.gov (United States)

    Maldonado-Siman, E; Bernal-Alcántara, R; Cadena-Meneses, J A; Altamirano-Cárdenas, J R; Martinez-Hernández, P A

    2014-12-01

    Requirements of hazard analysis and critical control points (HACCP) are becoming essential for international trade in food commodities as a safety assurance component. This research reports the level of the adoption of ISO 9000 and the HACCP system by Federal Inspection Type (TIF) pork-exporting enterprises. Implementation and operating costs are reported as well as the benefits involved in this food industry process. In Mexico, there are 97 companies classified as TIF enterprises, and 22 are registered as exporters of processed pork with the National Services for Safety and Quality and Animal Health of the Secretariat of Agriculture, Livestock, Rural Development, Fisheries and Food. Surveys were administered to 22 companies, with a 95.2% response rate. Enterprise characteristics were evaluated, as well as their operating activities. Fieldwork consisted of administering structured questionnaires to TIF exporters. All the surveyed enterprises had implemented HACCP, whereas the ISO 9000 regulation was applied in only 30%. Of total production, 75% is exported to 13 countries, and 25% goes to the Mexican market niche. Results indicate that the main factors for adopting HACCP are related to accessibility to international markets, improving quality, and reducing product quality audits by customers. The results also indicated that staff training was the most important issue. Microbiological testing was the highest cost of the operation. The main benefits reported were related to better access to international markets and a considerable reduction in microbial counts. This study shows the willingness of Mexican pork processors to implement food safety protocols for producing safe and quality products to compete in the international food trade.

  4. Microsatellite Attitude Determination and Control Subsystem Design and Implementation: Software-in-the-Loop Approach

    Directory of Open Access Journals (Sweden)

    Ho-Nien Shou

    2014-01-01

    Full Text Available The paper describes the development of a microsatellite attitude determination and control subsystem (ADCS and verification of its functionality by software-in-the-loop (SIL method. The role of ADCS is to provide attitude control functions, including the de-tumbling and stabilizing the satellite angular velocity, and as well as estimating the orbit and attitude information during the satellite operation. In Taiwan, Air Force Institute of Technology (AFIT, dedicating for students to design experimental low earth orbit micro-satellite, called AFITsat. For AFITsat, the operation of the ADCS consists of three modes which are initialization mode, detumbling mode, and normal mode, respectively. During the initialization mode, ADCS collects the early orbit measurement data from various sensors so that the data can be downlinked to the ground station for further analysis. As particularly emphasized in this paper, during the detumbling mode, ADCS implements the thrusters in plus-wide modulation control method to decrease the satellite angular velocity. ADCS provides the attitude determination function for the estimation of the satellite state, during normal mode. The three modes of microsatellite adopted Kalman filter algorithm estimate microsatellite attitude. This paper will discuss using the SIL validation ADCS function and verify its feasibility.

  5. Implementation of the ATLAS trigger within the multi-threaded software framework AthenaMT

    CERN Document Server

    Wynne, Benjamin; The ATLAS collaboration

    2016-01-01

    We present an implementation of the ATLAS High Level Trigger that provides parallel execution of trigger algorithms within the ATLAS multi­threaded software framework, AthenaMT. This development will enable the ATLAS High Level Trigger to meet future challenges due to the evolution of computing hardware and upgrades of the Large Hadron Collider, LHC, and ATLAS Detector. During the LHC data­taking period starting in 2021, luminosity will reach up to three times the original design value. Luminosity will increase further, to up to 7.5 times the design value, in 2026 following LHC and ATLAS upgrades. This includes an upgrade of the ATLAS trigger architecture that will result in an increase in the High Level Trigger input rate by a factor of 4 to 10 compared to the current maximum rate of 100 kHz. The current ATLAS multiprocess framework, AthenaMP, manages a number of processes that process events independently, executing algorithms sequentially in each process. AthenaMT will provide a fully multi­threaded env...

  6. Implementing a DVB-T/H Receiver on a Software-Defined Radio Platform

    Directory of Open Access Journals (Sweden)

    Yong Jiang

    2009-01-01

    Full Text Available Digital multimedia broadcasting is available in more and more countries with various forms. One of the most successful forms is Digital Video Broadcasting for Terrestrial (DVB-T, which has been deployed in most countries of the world for years. In order to bring the digital multimedia broadcasting services to battery-powered handheld receivers in a mobile environment, Digital Video Broadcasting for Handheld (DVB-H has been formally adopted by ETSI. More advanced and complex digital multimedia broadcasting systems are under development, for example, the next generation of DVB-T, a.k.a. DVB-T2. Current commercial DVB-T/H receivers are usually built upon dedicated application-specific integrated circuits (ASICs. However, ASICs are not flexible for incoming evolved standards and less overall-area efficient since they cannot be efficiently reused and shared among different radio standards, when we integrate a DVB-T/H receiver into a mobile phone. This paper presents an example implementation of a DVB-T/H receiver on the prototype of Infineon Technologies' Software-Defined Radio (SDR platform called MuSIC (Multiple SIMD Cores, which is a DSP-centered and accelerator-assisted architecture and aims at battery-powered mass-market handheld terminals.

  7. Automatic Rotational Sky Quality Meter (R-SQM) Design and Software for Astronomical Observatories

    Science.gov (United States)

    Dogan, E.; Ozbaldan, E. E.; Shameoni, Niaei M.; Yesilyaprak, C.

    2016-12-01

    We have presented the new design of Sky Quality Meter (SQM) device that is an automatic rotational model of sky quality meter (R-SQM) carried out by DAG (Eastern Anatolia Observatory) Technical Team. R-SQM is required for determining the long-term changes of sky quality of an astronomical observatory and consists of four SQM devices mounted on a rotating shaft with different angles for scanning all sky. This system is controlled by a Raspberry Pi control card and a step motor with its driver and a special software.

  8. Efficient FPGA Implementation of a STBC-OFDM Combiner for an IEEE 802.16 Software Radio Receiver

    DEFF Research Database (Denmark)

    Cattoni, Andrea Fabio; Le Moullec, Yannick; Sacchi, Claudio

    2014-01-01

    In this paper, an efficient FPGA implementation of a 4x4 Space-Time Block Coding (STBC) combiner for MIMO-OFDM software radio receivers is considered. The proposed combiner is based on a low-complexity algorithm which reduces the interference due to the Quasi-Orthogonality of the STBC decoding. I...

  9. Design: Vital step for Quality management system implementation.

    Directory of Open Access Journals (Sweden)

    Paula Evenilda Cruz Esmoris

    2009-12-01

    Full Text Available Sancti Spíritus´ Acopio, Beneficio y Comercialización de Productos Apícolas organization developed a Quality management system (QMS according to NC-ISO 9001:2008 ”Quality management systems - Requirements”. In this organization, QMS implementation is not only a strategy to increase competitiveness and clients satisfaction, but a requirement to improve its performance and constitutes a demand of international market because honey is an exportable product. This article presents QMS design according to the description of its key elements. Design was based on legal and regulatory requirements from corresponding organizations and technical orientations issued by the Empresa Nacional Apícola. Items taking into account were related with the definition of: scope (products, processes and areas; QMS processes and their relations (map; product related processes and activities; needed resources; responsibilities with QMS; barriers favorable forces to QMS implementation and documentation to support QMS processes. Use of a correct design allows a successful next steps implementation of QMS and complying with the Instituto Nacional de Normalización methodology.

  10. 78 FR 21582 - Revisions to the California State Implementation Plan, Butte County Air Quality Management...

    Science.gov (United States)

    2013-04-11

    ... AGENCY 40 CFR Part 52 Revisions to the California State Implementation Plan, Butte County Air Quality Management District and Sacramento Metropolitan Air Quality Management District AGENCY: Environmental... County Air Quality Management District (BCAQMD) and Sacramento Metropolitan Air Quality...

  11. 75 FR 40762 - Revisions to the California State Implementation Plan, Sacramento Metropolitan Air Quality...

    Science.gov (United States)

    2010-07-14

    ... AGENCY 40 CFR Part 52 Revisions to the California State Implementation Plan, Sacramento Metropolitan Air Quality Management District and South Coast Air Quality Management District AGENCY: Environmental... Sacramento Metropolitan Air Quality Management District (SMAQMD) and South Coast Air Quality Management...

  12. [Implementation of a software for acquiring and analysing myoelectric potential signals in ergonomical research of space manual system].

    Science.gov (United States)

    Li, J; Yang, J Q; Yao, S Y; Xu, Y Z; Wang, H X

    2000-10-01

    Objective. A software used for acquiring and analyzing signals was developed for ergonomical research on Human Workload in space manual system. Method. As an important part of the whole experimental equipment and being developed in PC, the software is composed of acquisition and display program, analysis and processing program and data files. Result. The software is capable of making realtime acquisition and display of four channels of myoelectric potential signals and one channel of operation signal synchronously. The signals are then analyzed and processed off-line. Conclusion. During the development, its realtime feature was implemented by means of some technical methods, such as that different machine assigned to view-scenery display and signal acquisition, different frequency assigned to EMG signal and operation ones, and different cycle assigned to acquisition, display and storage. Above all, with friendly man-machine interface and high accuracy of data, the software was even reliable.

  13. The Role of Information Systems in Implementing Total Quality Management

    Directory of Open Access Journals (Sweden)

    Ahmad Z. Siam

    2012-01-01

    Full Text Available Problem statement: The aim of this study is to identify the extent to which Information Technology (IT has been used to support TQM in order to identify the role of IT in implementing Total Quality Management (TQM. The primary emphasis of this study was on how IT provides help and supports in TQM processes rather than on TQM performance. This has many policy implications for different firms that are welling to enhance the role of IT in different aspects of their TQM. Approach: The questionnaire survey was used as the data collection method. Data were analysed using SPSS programme. This study is limited only to the managers of operations departments in different companies at Sohar Industrial Estate in Oman. Results: The results of the research indicate that the usage of IT is not the same among the different TQM dimensions. The highest level of IT support was used in information and analysis, output quality assurance and important innovation. While the lowest level of IT support was used in quality results and supplier quality assurance. Conclusion/Recommendations: As IT role should be enhanced in modifying work processes in order to improve product quality and productivity. Value and international implications: This study is the first to be applied in Gulf area, an area that exhibits arapid groth and presence of international companies.

  14. Design and implementation of the STKOS quality-control system

    Institute of Scientific and Technical Information of China (English)

    Haixia; SUN; Danya; LI; Junlian; LI; Sizhu; WU; Tiejun; HU; Qing; QIAN

    2015-01-01

    Purpose: The aim of this study is to develop and implement a quality-control system to ensure authority control of the different knowledge units for the Scientific & Technological Knowledge Organization Systems(STKOS).Design/methodology/approach: First, we analyzed quality-control requirements based on the construction of the STKOS Metathesaurus. Then we designed a quality-control framework, the task management and transfer mechanism, and a service model. Afterwards, we carried out the experiments to check the rules and algorithms used in the system. Finally, the system was developed, and gradually optimized during its service.Findings: The quality-control system supports collaborative knowledge construction, as well as consistency checks of knowledge units with different granularity levels, including terminologies, relationships, and concepts. The system can be flexibly configured.Research limitations: The system is oriented to an English-language knowledge organization system, and may not perform well with Chinese-language systems and ontologies.Practical implications: The system can be used to support the construction of a single knowledge organization system, as well as data warehousing and interoperable knowledge organization systems. Originality/value: The STKOS quality-control system not only focuses on content building for the knowledge system, but also supports collaborative task management.

  15. 软件体系结构质量评价概述%An Introduction to Software Architecture Quality Evaluation

    Institute of Scientific and Technical Information of China (English)

    周欣; 黄璜; 孙家骕; 燕小荣

    2003-01-01

    Software quality is one of the most important characteristics of software system and impacts on thesystem's effect, cost and efficiency. As is well known, it's better to improve the quality as early as possible, whichcan reduce the cost in following development and maintenance. Software architecture is the first activity from problemspace to software solution space, therefore, the decisions made during this process are significant to software quality.Software architecture quality evaluation analyzes and predicts qualityfrom architecture level, helping make proper ar-chitectural decisions and detecting derivation during following development. This paper summarizes the researches inthis area, introducing the purpose, content, state of the art and application status, analyzing the difficulties and dis-cussing the future directions.

  16. SOFTWARE PROCESS MATURITY AND CAPABILITY EVALUATION AND ITS IMPLEMENTATION IN THE AREAS OF E-LEARNING ORGANIZATIONS

    Directory of Open Access Journals (Sweden)

    Senthilkumar Chennakrishnan

    2014-01-01

    Full Text Available A detailed literature review revealed that, most of the organizations develop their projects and associated software engineering activities using in-house established methods. There are discrete references in the literature on application of broad based scientific methods in designing and developing process capability and maturity models by software organizations. Documentation and extent of adherence to a well defined organized scientific method plays a major role in standardizing the systems. Capability Maturity Model Integration (CMMI is a recognized model for providing process maturity and capability in managing the software engineering activities. Further, the area of development of process models for optimization and implementation of thrust areas were not given due attention by researchers. To address above issues, we made an attempt in this study (i to review the extent to which software organization make use of software process capability models, (ii to develop an organized software process modeling and (iii to evaluate its performance through a case study conducted in an IT industry developing E-learning maturity models and to arrive at optimized model for E-Learning thrust areas. The basis for above work is drawn by comparing the existing methods and tools used, with international models. The major contribution of present work is that, the proposed process capability model will enable organizations to follow broad based scientific methods for pre modeling their software activities and will also enable them to form levels that result in bench mark approach for E-Learning modules.

  17. Quality control of CarboEurope flux data - Part 2: Inter-comparison of eddy-covariance software

    NARCIS (Netherlands)

    Mauder, M.; Foken, T.; Clement, R.; Elbers, J.A.; Eugster, W.; Grunwald, T.; Heusinkveld, B.G.; Kolle, O.

    2008-01-01

    As part of the quality assurance and quality control activities within the CarboEurope-IP network, a comparison of eddy-covariance software was conducted. For four five-day datasets, CO2 flux estimates were calculated by seven commonly used software packages to assess the uncertainty of CO2 flux est

  18. Commissioning, clinical implementation and quality assurance of Siemen's Virtual Wedge.

    Science.gov (United States)

    Rathee, S; Kwok, C B; MacGillivray, C; Mirzaei, M

    1999-01-01

    This report presents the results of commissioning, clinical implementation and quality assurance of Siemens Virtual Wedge. Our measurements show that: (1) wedge factors are within 2% of unity, (2) percentage depth doses are within 1% of open beam data, and (3) wedged beam profiles can be modeled similarly to a physical wedge and follow a well defined equation to facilitate modeling of an arbitrary wedge angle. The gantry angle dependence of wedge profiles is similar to open beam profiles. The output of wedged fields is linear with delivered monitor units within 1%. Quality assurance results indicate the wedge profiles are very stable over time. Day to day variations of two points measured along the wedge gradient direction are within 1.5%.

  19. Design and Implementation of Real-Time Software Radio for Anti-Interference GPS/WAAS Sensors

    Directory of Open Access Journals (Sweden)

    Per Enge

    2012-10-01

    Full Text Available Adaptive antenna array processing is widely known to provide significant anti-interference capabilities within a Global Navigation Satellite Systems (GNSS receiver. A main challenge in the quest for such receiver architecture has always been the computational/processing requirements. Even more demanding would be to try and incorporate the flexibility of the Software-Defined Radio (SDR design philosophy in such an implementation. This paper documents a feasible approach to a real-time SDR implementation of a beam-steered GNSS receiver and validates its performance. This research implements a real-time software receiver on a widely-available x86-based multi-core microprocessor to process four-element antenna array data streams sampled with 16-bit resolution. The software receiver is capable of 12 channels all-in-view Controlled Reception Pattern Antenna (CRPA array processing capable of rejecting multiple interferers. Single Instruction Multiple Data (SIMD instructions assembly coding and multithreaded programming, the key to such an implementation to reduce computational complexity, are fully documented within the paper. In conventional antenna array systems, receivers use the geometry of antennas and cable lengths known in advance. The documented CRPA implementation is architected to operate without extensive set-up and pre-calibration and leverages Space-Time Adaptive Processing (STAP to provide adaptation in both the frequency and space domains. The validation component of the paper demonstrates that the developed software receiver operates in real time with live Global Positioning System (GPS and Wide Area Augmentation System (WAAS L1 C/A code signal. Further, interference rejection capabilities of the implementation are also demonstrated using multiple synthetic interferers which are added to the live data stream.

  20. Design and implementation of real-time software radio for anti-interference GPS/WAAS sensors.

    Science.gov (United States)

    Chen, Yu-Hsuan; Juang, Jyh-Ching; Seo, Jiwon; Lo, Sherman; Akos, Dennis M; De Lorenzo, David S; Enge, Per

    2012-01-01

    Adaptive antenna array processing is widely known to provide significant anti-interference capabilities within a Global Navigation Satellite Systems (GNSS) receiver. A main challenge in the quest for such receiver architecture has always been the computational/processing requirements. Even more demanding would be to try and incorporate the flexibility of the Software-Defined Radio (SDR) design philosophy in such an implementation. This paper documents a feasible approach to a real-time SDR implementation of a beam-steered GNSS receiver and validates its performance. This research implements a real-time software receiver on a widely-available x86-based multi-core microprocessor to process four-element antenna array data streams sampled with 16-bit resolution. The software receiver is capable of 12 channels all-in-view Controlled Reception Pattern Antenna (CRPA) array processing capable of rejecting multiple interferers. Single Instruction Multiple Data (SIMD) instructions assembly coding and multithreaded programming, the key to such an implementation to reduce computational complexity, are fully documented within the paper. In conventional antenna array systems, receivers use the geometry of antennas and cable lengths known in advance. The documented CRPA implementation is architected to operate without extensive set-up and pre-calibration and leverages Space-Time Adaptive Processing (STAP) to provide adaptation in both the frequency and space domains. The validation component of the paper demonstrates that the developed software receiver operates in real time with live Global Positioning System (GPS) and Wide Area Augmentation System (WAAS) L1 C/A code signal. Further, interference rejection capabilities of the implementation are also demonstrated using multiple synthetic interferers which are added to the live data stream.

  1. The SFXC software correlator for Very Long Baseline Interferometry: Algorithms and Implementation

    CERN Document Server

    Keimpema, A; Pogrebenko, S V; Campbell, R M; Cimó, G; Duev, D A; Eldering, B; Kruithof, N; van Langevelde, H J; Marchal, D; Calvés, G Molera; Ozdemir, H; Paragi, Z; Pidopryhora, Y; Szomoru, A; Yang, J

    2015-01-01

    In this paper a description is given of the SFXC software correlator, developed and maintained at the Joint Institute for VLBI in Europe (JIVE). The software is designed to run on generic Linux-based computing clusters. The correlation algorithm is explained in detail, as are some of the novel modes that software correlation has enabled, such as wide-field VLBI imaging through the use of multiple phase centres and pulsar gating and binning. This is followed by an overview of the software architecture. Finally, the performance of the correlator as a function of number of CPU cores, telescopes and spectral channels is shown.

  2. Characteristics for Software Optimization Projects

    Directory of Open Access Journals (Sweden)

    Iulian NITESCU

    2008-01-01

    Full Text Available The increasing of the software systems complexity imposes the identification and implementation of some methods and techniques in order to manage it. The software optimization project is a way in which the software complexity is controlled. The software optimization project must face to the organization need to earn profit. The software optimization project is an integrated part of the application cycle because share same resources, depends on other stages and influences next phases. The optimization project has some particularities because it works on an finished product around its quality. The process is quality and performance oriented and it assumes that the product life cycle is almost finished.

  3. STATIC CODE ANALYSIS FOR SOFTWARE QUALITY IMPROVEMENT: A CASE STUDY IN BCI FRAMEWORK DEVELOPMENT

    Directory of Open Access Journals (Sweden)

    Indar Sugiarto

    2008-01-01

    Full Text Available This paper shows how the systematic approach in software testing using static code analysis method can be used for improving the software quality of a BCI framework. The method is best performed during the development phase of framework programs. In the proposed approach, we evaluate several software metrics which are based on the principles of object oriented design. Since such method is depending on the underlying programming language, we describe the method in term of C++ language programming whereas the Qt platform is also currently being used. One of the most important metric is so called software complexity. Applying the software complexity calculation using both McCabe and Halstead method for the BCI framework which consists of two important types of BCI, those are SSVEP and P300, we found that there are two classes in the framework which have very complex and prone to violation of cohesion principle in OOP. The other metrics are fit the criteria of the proposed framework aspects, such as: MPC is less than 20; average complexity is around value of 5; and the maximum depth is below 10 blocks. Such variables are considered very important when further developing the BCI framework in the future.

  4. The Data Quality Monitoring Software for the CMS experiment at the LHC

    CERN Document Server

    Borrello, Laura

    2014-01-01

    The Data Quality Monitoring (DQM) Software is a central tool in the CMS experiment. Its flexibility allows for integration in several key environments Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release-Validation, to constantly validate the functionalities and the performance of the reconstruction software; in Monte Carlo productions.Since the end of data taking at a center of mass energy of 8 TeV, the environment in which the DQM lives has undergone fundamental changes. In turn, the DQM system has made significant upgrades in many areas to respond to not only the changes in infrastructure, but also the growing specialized needs of the collaboration with an emphasis on more sophisticated methods for evaluating dataquality, as well as advancing the DQM system to provide quality assessments of various Monte Carlo simulations versus data distributions, monitoring changes in physical effects due to modifications of algorithms or framework, and enabling reg...

  5. [Recommendations for implementing the quality policy and organisation of a quality management system].

    Science.gov (United States)

    Daunizeau, A

    2013-06-01

    Preliminary issues to implement a quality management system are described. They include the definition of the structure, a hierarchical and functional organization chart and the engagement of the whole personnel to apply the requirements of the standard EN ISO 15189. The policy has to be translated into objectives.

  6. Implementation of Quality Management System in Iranian Medical Laboratories

    Directory of Open Access Journals (Sweden)

    P Dahim

    2009-03-01

    Full Text Available "nReference health Laboratory of Iran (RHL has been established under the Secretary of Health in 2007. This organization has been developed from integration of Reference laboratories of Iran, Directorate of General Affairs and Laboratory de­partment of CDC. In accordance with policies of Ministry of Health and Medical Education, also Millennium Development Goals and requirements of IHR the RHL has highlighted his main goal as "Quality Improvement of Medical Laboratory Net­work". In order to achieve this goal as soon and efficient as possible, RHL has tried to launch Quality Management sys­tem in medical laboratories. After studying and analyzing the situation in the country RHL has tried by using ISO 15189 which is an successful international standard to draw national standards and write necessary manuals, guidelines and check­lists. For a successful implementation and detailed managerial observance, three phases were designed: Training, Stepwise imple­mentation of standard requirements in the laboratories, accreditation of laboratories. After one year of training and send­ing out national standards and guidelines, the RHL is in the middle of the second phase "implementation".

  7. Implementation of quality by design toward processing of food products.

    Science.gov (United States)

    Rathore, Anurag S; Kapoor, Gautam

    2017-05-28

    Quality by design (QbD) is a systematic approach that begins with predefined objectives and emphasizes product and process understanding and process control. It is an approach based on principles of sound science and quality risk management. As the food processing industry continues to embrace the idea of in-line, online, and/or at-line sensors and real-time characterization for process monitoring and control, the existing gaps with regard to our ability to monitor multiple parameters/variables associated with the manufacturing process will be alleviated over time. Investments made for development of tools and approaches that facilitate high-throughput analytical and process development, process analytical technology, design of experiments, risk analysis, knowledge management, and enhancement of process/product understanding would pave way for operational and economic benefits later in the commercialization process and across other product pipelines. This article aims to achieve two major objectives. First, to review the progress that has been made in the recent years on the topic of QbD implementation in processing of food products and second, present a case study that illustrates benefits of such QbD implementation.

  8. Software productivity improvement through software engineering technology

    Science.gov (United States)

    Mcgarry, F. E.

    1985-01-01

    It has been estimated that NASA expends anywhere from 6 to 10 percent of its annual budget on the acquisition, implementation and maintenance of computer software. Although researchers have produced numerous software engineering approaches over the past 5-10 years; each claiming to be more effective than the other, there is very limited quantitative information verifying the measurable impact htat any of these technologies may have in a production environment. At NASA/GSFC, an extended research effort aimed at identifying and measuring software techniques that favorably impact productivity of software development, has been active over the past 8 years. Specific, measurable, software development technologies have been applied and measured in a production environment. Resulting software development approaches have been shown to be effective in both improving quality as well as productivity in this one environment.

  9. Does the implementation of hardware need software? A longitudinal study on fluoride-removal filter use in Ethiopia.

    Science.gov (United States)

    Sonego, Ina L; Huber, Alexandra C; Mosler, Hans-Joachim

    2013-11-19

    Evidence suggests that the effectiveness of technology designed to provide safe and healthy water is dependent on the degree of its use. In addition to providing safe water "hardware" (i.e., new infrastructure or equipment) to populations at risk, it might be necessary to also provide suitable "software" programs (behavior change strategies) to support use. A longitudinal survey was conducted in rural Ethiopia following the distribution of fluoride-removal household filters. Three intervention groups were evaluated. Group 1 only received the hardware, i.e., the fluoride-removal filter. Groups 2 and 3 also received software in the form of two evidence-based psychological interventions: a planning and social prompts intervention and an educational workshop with pledging. Group 2 received both software interventions, and Group 3 only received the educational workshop. The effects of the hardware and software on behavior and thus filter use were analyzed along with specific psychological factors. The results showed that the provision of the hardware alone (the fluoride-removal filter) was not enough to ensure sufficient use of the equipment. The addition of a software component in the form of psychological interventions increased filter use up to 80%. An increase in filter use was measured following each intervention resulting in the health-risk being minimized. We conclude that it is necessary that the implementation of hardware of this nature is accompanied by evidence-based intervention software.

  10. Improving Quality of Perception (QoP, Quality of Experience (QoE, and Quality of Service (QoS in agile development using Cleanroom Software Engineering (CSE

    Directory of Open Access Journals (Sweden)

    Sana e Zainab

    2014-10-01

    Full Text Available Pioneering ideas from the software engineering discipline have factually affected every sphere of life. Agile software development approach has been promoted since its commencement and stipulates strategies that improve the quality of software product. To consummate fast and reliable development processes, several agile approaches are charted and are quite popular. For quality improvement and to achieve defect free system, the concept of Cleanroom Software Engineering (CSE is ingrained into agile development life cycle. For embedding users concerns, it is important to distinguish three approaches to quality: Quality of Service (QoS, User-perceived QoP, and Quality of Experience (QoE. QoS is technology centered approach, so by using Incremental Planning of CSE, it shall facilitate the customer's clarification of system requirements and will control the technical complexity. Usage Specification and Usage Modelling will be used during the Certification phase of CSE which will help to achieve QoP and QoE, being user centered approaches. Results collected from Survey conducted, explains above mentioned factors improvement.

  11. POLITO- A new open-source, platform independent software for generating high-quality lithostratigraphic columns

    Directory of Open Access Journals (Sweden)

    Cipran C. Stremtan

    2010-08-01

    Full Text Available POLITO is a free, open-source, and platform-independent software which can automatically generate lithostratigraphic columns from field data. Its simple and easy to use interface allows users to manipulate large datasets and create high-quality graphical outputs, either in editable vector or raster format, or as PDF files. POLITO uses USGS standard lithology patterns and can be downloaded from its Sourceforge project page (http://sourceforge.net/projects/polito/.

  12. Implementation of a quality assurance program for computerized treatment planning systems.

    Science.gov (United States)

    Camargo, Priscilla R T L; Rodrigues, Laura N; Furnari, Laura; Rubo, Rodrigo A

    2007-07-01

    In the present investigation, the necessary tests for implementing a quality assurance program for a commercial treatment planning system (TPS), recently installed at Sao Paulo University School of Medicine Clinicas Hospital-Brazil, was established and performed in accordance with the new IAEA publication TRS 430, and with AAPM Task Group 53. The tests recommended by those documents are classified mainly into acceptance, commissioning (dosimetric and nondosimetric), periodic quality assurance, and patient specific quality assurance tests. The recommendations of both IAEA and AAPM documents are being implemented at the hospital for photon beams produced by two linear accelerators. A Farmer ionization chamber was used in a 30 x 30 x 30 cm3 phantom with a dose rate of 320 monitor unit (MU)/min and 50 MU in the case of the dosimetric tests. The acceptance tests verified hardware, network systems integration, data transfer, and software parameters. The results obtained are in good agreement with the specifications of the manufacturer. For the commissioning dosimetric tests, the absolute dose was measured for simple geometries, such as square and rectangular fields, up to more complex geometries such as off-axis hard wedges and for behavior in the build up region. Results were analysed by the use of confidence limit as proposed by Venselaar et al. [Radio Ther. Oncol. 60, 191-201 (2001)]. Criteria of acceptability had been applied also for the comparison between the values of MU calculated manually and MU generated by TPS. The results of the dosimetric tests show that work can be reduced by choosing to perform only those that are more crucial, such as oblique incidence, shaped fields, hard wedges, and buildup region behavior. Staff experience with the implementation of the quality assurance program for a commercial TPS is extremely useful as part of a training program.

  13. Analyst Tools and Quality Control Software for the ARM Data System

    Energy Technology Data Exchange (ETDEWEB)

    Moore, Sean; Hughes, Gary

    2008-07-31

    Mission Research develops analyst tools and automated quality control software in order to assist the Atmospheric Radiation Measurement (ARM) Data Quality Office with their data inspection tasks. We have developed web-based data analysis and visualization tools such as the interactive plotting program NCVweb, various diagnostic plot browsers, and a datastream processing status application. These tools allow even novice ARM researchers to be productive with ARM data with only minimal effort. We also contribute to the ARM Data Quality Office by analyzing ARM data streams, developing new quality control metrics, new diagnostic plots, and integrating this information into DQ HandS - the Data Quality Health and Status web-based explorer. We have developed several ways to detect outliers in ARM data streams and have written software to run in an automated fashion to flag these outliers. We have also embarked on a system to comprehensively generate long time-series plots, frequency distributions, and other relevant statistics for scientific and engineering data in most high-level, publicly available ARM data streams. Furthermore, frequency distributions categorized by month or by season are made available to help define valid data ranges specific to those time domains. These statistics can be used to set limits that when checked, will improve upon the reporting of suspicious data and the early detection of instrument malfunction. The statistics and proposed limits are stored in a database for easy reporting, refining, and for use by other processes. Web-based applications to view the results are also available.

  14. Projeto Seis Sigma para a implementação de software de programação Six Sigma project for scheduling software implementation

    Directory of Open Access Journals (Sweden)

    Rogério Cerávolo Calia

    2005-12-01

    Full Text Available O artigo visa analisar a eficácia organizacional da metodologia Seis Sigma na gestão de projetos para a redução de atrasos e redução de estoques na manufatura, por meio da implementação de um software com algoritmos da Teoria das Restrições. Inicialmente, é apresentada uma revisão bibliográfica sobre a gestão de projetos na perspectiva da gestão da mudança organizacional nos processos de negócios. Em seguida, são revistos os conceitos sobre a metodologia Seis Sigma para a gestão de projetos e sobre os algoritmos da Teoria das Restrições. Então, são descritos os estudos de caso em dois projetos de implementação do software da Teoria das Restrições, sendo que apenas uma das implementações utilizou-se da metodologia Seis Sigma para a gestão do projeto. Na análise dos resultados, busca-se compreender os motivos de o projeto com a metodologia Seis Sigma ter reduzido inventário três vezes mais rápido do que o projeto sem o Seis Sigma.The article aims to analyze the organizational effectiveness of the Six Sigma methodology for project management to reduce delays and to reduce inventory in manufacture, by the implementation of software with Theory of Constraints algorithms. Initially, the article presents a bibliographic revision on project management and its impact on the organizational change management for improving business processes. Then, the article revises the concepts about the Six Sigma methodology for project management and about the Theory of Constraints algorithms. It follows, the case studies descriptions on two implementation projects of the Theory of Constraints software, in which only one of these implementations adopted the Six Sigma methodology in the project management. In the results analyzes, the article discusses the reasons why the project with the Six Sigma methodology was three times faster than the other project.

  15. Air Quality in Mexico City: Policies Implemented for its Improvement

    Science.gov (United States)

    Paramo, V.

    2007-12-01

    stringent emission levels of the gasoline fleet; update the detention of pollutant vehicles program; partial exemption of the inspection and maintenance program for cleaner and or highly efficient vehicles; substitution of 3,000 microbuses, 40,000 taxis and 1,200 buses; commissioning of the first Bus Rapid Transit system; implementation of a program for the emissions reduction for the 300 most polluted industrial facilities; and continuous update of the air quality environmental management programs. To continue improving the air quality in the MCMA, the environmental authorities will continue the implementation of the 2002-2010 Air Quality Improvement Program. In 2007 the Green Program was started, this includes those actions that have proven to be effective reduction of pollutant emissions and incorporates new actions for the reduction of local and global pollutant emissions. The most important of these new actions are: substitution of 9,500 microbuses; renewal of all the taxis fleet; commissioning of 10 Bus Rapid Transit lines; commissioning of Line 12 of the underground system; schedules and routes limitations to the cargo fleet; increase 5 percent the number of non-motorized trips (bicycling and walking); regulation of the private public transport passenger stops; requirement of private schools to provide school transport; regulation of non-occupied taxis in circulation; modifications to the circulation of 350 critical crossing points in the city; adoption of intelligent traffic lights systems; complete substitution of the local government vehicle's fleet; implement the inspection and maintenance of the cargo fleet; introduction of low- sulfur diesel, among other measures.

  16. Quality assurance applied to mammographic equipments using phantoms and software for its evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Mayo, Patricia, E-mail: p.mayo@titaniast.co [Titania Servicios Tecnologicos S.L., Grupo Dominguis, Apartado 46015, Valencia (Spain); Rodenas, Francisco [Departamento de Matematica Aplicada, Universidad Politecnica de Valencia, Apartado 46022, Valencia (Spain); Manuel Campayo, Juan [Hospital Clinico Universitario de Valencia, Avda. Blasco Ibanez, Apartado 46017, Valencia (Spain); Verdu, Gumersido [Departamento de Ingenieria Quimica y Nuclear, Universidad Politecnica de Valencia, Apartado 46022, Valencia (Spain)

    2010-07-21

    The image quality assessment in radiographic equipments is a very important item for a complete quality control of the radiographic image chain. The periodic evaluation of the radiographic image quality must guarantee the constancy of this quality to carry out a suitable diagnosis. Mammographic phantom images are usually used to study the quality of images obtained by determined mammographic equipment. The digital image treatment techniques allow to carry out an automatic analysis of the phantom image. In this work we apply some techniques of digital image processing to analyze in an automatic way the image quality of mammographic phantoms, namely CIRS SP01 and RACON for different varying conditions of the mammographic equipment. The CIRS SP01 phantom is usually used in analogic mammographic equipments and the RACON phantom has been specifically developed by authors to be applied to acceptance and constancy tests of the image quality in digital radiographic equipments following recommendations of international associations. The purpose of this work consists in analyzing the image quality for both phantoms by means of an automatic software utility. This analysis allows us to study the functioning of the image chain of the mammographic system in an objective way, so an abnormal functioning of the radiographic equipment might be detected.

  17. Quality assurance applied to mammographic equipments using phantoms and software for its evaluation

    Science.gov (United States)

    Mayo, Patricia; Rodenas, Francisco; Manuel Campayo, Juan; Verdú, Gumersido

    2010-07-01

    The image quality assessment in radiographic equipments is a very important item for a complete quality control of the radiographic image chain. The periodic evaluation of the radiographic image quality must guarantee the constancy of this quality to carry out a suitable diagnosis. Mammographic phantom images are usually used to study the quality of images obtained by determined mammographic equipment. The digital image treatment techniques allow to carry out an automatic analysis of the phantom image. In this work we apply some techniques of digital image processing to analyze in an automatic way the image quality of mammographic phantoms, namely CIRS SP01 and RACON for different varying conditions of the mammographic equipment. The CIRS SP01 phantom is usually used in analogic mammographic equipments and the RACON phantom has been specifically developed by authors to be applied to acceptance and constancy tests of the image quality in digital radiographic equipments following recommendations of international associations. The purpose of this work consists in analyzing the image quality for both phantoms by means of an automatic software utility. This analysis allows us to study the functioning of the image chain of the mammographic system in an objective way, so an abnormal functioning of the radiographic equipment might be detected.

  18. Implementation of Similarity Based Kriging in Open Source Software and Application to Uncertainty Quantification and Reduction in Hydrogeological Inversion

    Science.gov (United States)

    Komara, R.; Ginsbourger, D.

    2014-12-01

    We present the implementation of Similarity Based Kriging (SBK). This approach extends Gaussian process regression (GPR) methods, typically restricted to Euclidean spaces, to spaces that are non-Euclidean or perhaps even non-metric. SBK was inspired by problems in aquifer modeling, where inputs of numerical simulations are typically curves and parameter fields, and predicting scalar or vector outputs by Kriging with such very high-dimensional inputs may seem not feasible at first. SBK combines ideas from the distance-based set-up of Scheidt and Caers (2009) with GPR and allows calculating Kriging predictions based only on similarities between inputs rather than on their high-dimensional representation. Written in open source code, this proposed approach includes automated construction of SBK models and provides diagnostics to assess model quality both in terms of covariance fitting and internal/external prediction validation. Covariance hyperparameters can be estimated both by maximum likelihood and leave-one-out cross validation relying in both cases on efficient formulas and a hybrid genetic optimization algorithm using derivatives. The determination of the best dimension for Classical multidimensional scaling (MDS) and non-metric MDS of the data will be investigated. Application of this software to real life data examples in Euclidean and non-Euclidean (dis)similarity settings will be covered and touch on aquifer modeling, hydrogeological forecasting, and sequential inverse problem solving. In the last case, a novel approach where a variant of the expected improvement criterion is used for choosing several points at a time will be presented. This part of the method and the previous covariance hyperparameter estimation parallelize naturally and we demonstrate how to save computation time by optimally distributing function evaluations over multiple cores or processors.

  19. Implementation of a software for REmote COMparison of PARticlE and photon treatment plans: ReCompare.

    Science.gov (United States)

    Löck, Steffen; Roth, Klaus; Skripcak, Tomas; Worbs, Mario; Helmbrecht, Stephan; Jakobi, Annika; Just, Uwe; Krause, Mechthild; Baumann, Michael; Enghardt, Wolfgang; Lühr, Armin

    2015-09-01

    To guarantee equal access to optimal radiotherapy, a concept of patient assignment to photon or particle radiotherapy using remote treatment plan exchange and comparison - ReCompare - was proposed. We demonstrate the implementation of this concept and present its clinical applicability. The ReCompare concept was implemented using a client-server based software solution. A clinical workflow for the remote treatment plan exchange and comparison was defined. The steps required by the user and performed by the software for a complete plan transfer were described and an additional module for dose-response modeling was added. The ReCompare software was successfully tested in cooperation with three external partner clinics and worked meeting all required specifications. It was compatible with several standard treatment planning systems, ensured patient data protection, and integrated in the clinical workflow. The ReCompare software can be applied to support non-particle radiotherapy institutions with the patient-specific treatment decision on the optimal irradiation modality by remote treatment plan exchange and comparison. Copyright © 2015. Published by Elsevier GmbH.

  20. Design and Implementation of Adaptive Turbo Encoder for Quantized Software Defined Low-Power DVB-RCS Radios

    Directory of Open Access Journals (Sweden)

    S. H. Elramly

    2010-11-01

    Full Text Available Turbo codes are employed in every robust wireless digital communications system. Those codes have been adopted for the satellite return channel in DVB-RCS (Return Channel via Satellite standard. In Software Defined Radios (SDRs, Field Programmable Gate Array technology (FPGA is considered a highly configurable option for implementing many sophisticated signal processing tasks. The implementation for such codes is complex and dissipates a large amount of power. This paper studies the efficient implementation of quantized DVB-RCS turbo coding. Also, a low-power, turbo encoder for DVB-RCS is described using a VHDL code. The proposed encoder design is implemented on Xilinx Virtex-II Pro, XC2vpx30 FPGA chip. FPGA Advantage Pro package provided by Mentor Graphics is used for VHDL description and ISE 10.1 by Xilinx is used for synthetization.

  1. 78 FR 21540 - Revisions to the California State Implementation Plan, Butte County Air Quality Management...

    Science.gov (United States)

    2013-04-11

    ... Management District and Sacramento Metropolitan Air Quality Management District AGENCY: Environmental... revisions to the Butte County Air Quality Management District (BCAQMD) and Sacramento Metropolitan Air Quality Management District (SMAQMD) portions of the California State Implementation Plan (SIP)....

  2. THE DESIGN AND IMPLEMENTATION OF B96 COMPILER AND B96/8096/8098 SOFTWARE EMULATING PLATFORM IDDEE

    Institute of Scientific and Technical Information of China (English)

    陆建德; 卢维亮; 张玉华

    1995-01-01

    A new high-level langusge,B96,which is compiling-type and MCS-96 single chip computer series oriented ,has recently been designed and implemented by the authors. This paper discusses its design thought and its implementation. The paper also accounts for the design of B96/8096/8098 IDDEE,a software Integrated Debugging & Developing Environment of Emulation,which is used to debug and develop 8096/8098 assemble language and/or B96 high-level language source program on PC-based emulating platform.

  3. Structural monitoring system with fiber Bragg grating sensors: Implementation and software solution

    CERN Document Server

    Fedorov, Aleksey; Makhrov, Ilya; Pozhar, Nikolay; Anufriev, Maxim; Pnev, Alexey; Karasik, Valeriy

    2014-01-01

    We present a structural health monitoring system for nondestructive testing of composite materials based on the fiber Bragg grating sensors and specialized software solution. Developed structural monitoring system has potential applications for preliminary tests of novel composite materials as well as real-time structural health monitoring of industrial objects. The software solution realizes control for the system, data processing and alert of an operator.

  4. Neuromorphic implementation of a software-defined camera that can see through fire and smoke in real-time

    Science.gov (United States)

    Cha, Jae H.; Abbott, A. Lynn; Szu, Harold H.; Willey, Jefferson; Landa, Joseph; Krapels, Keith A.

    2014-05-01

    Software-defined Cameras (SDC) based on Boltzmann's molecular thermodynamics can "see" through visually-degraded fields such as fire, fog, and dust in some situations. This capability is possible by means of unsupervised learning implemented on a neuromorphic algo-tecture. This paper describes the SDC algorithm design strategy with respect to nontrivial solutions, stability, and accuracy. An example neuromorphic learning algorithm is presented along with unsupervised learning stopping criteria.

  5. Implementing a Data Quality Strategy to Simplify Access to Data

    Science.gov (United States)

    Druken, K. A.; Trenham, C. E.; Evans, B. J. K.; Richards, C. J.; Wang, J.; Wyborn, L. A.

    2016-12-01

    To ensure seamless programmatic access for data analysis (including machine learning), standardization of both data and services is vital. At the Australian National Computational Infrastructure (NCI) we have developed a Data Quality Strategy (DQS) that currently provides processes for: (1) the consistency of data structures in the underlying High Performance Data (HPD) platform; (2) quality control through compliance with recognized community standards; and (3) data quality assurance through demonstrated functionality across common platforms, tools and services. NCI hosts one of Australia's largest repositories (10+ PBytes) of research data collections spanning datasets from climate, coasts, oceans and geophysics through to astronomy, bioinformatics and the social sciences. A key challenge is the application of community-agreed data standards to the broad set of Earth systems and environmental data that are being used. Within these disciplines, data span a wide range of gridded, ungridded (i.e., line surveys, point clouds), and raster image types, as well as diverse coordinate reference projections and resolutions. By implementing our DQS we have seen progressive improvement in the quality of the datasets across the different subject domains, and through this, the ease by which the users can programmatically access the data, either in situ or via web services. As part of its quality control procedures, NCI has developed a compliance checker based upon existing domain standards. The DQS also includes extensive Functionality Testing which include readability by commonly used libraries (e.g., netCDF, HDF, GDAL, etc.); accessibility by data servers (e.g., THREDDS, Hyrax, GeoServer), validation against scientific analysis and programming platforms (e.g., Python, Matlab, QGIS); and visualization tools (e.g., ParaView, NASA Web World Wind). These tests ensure smooth interoperability between products and services as well as exposing unforeseen requirements and

  6. Increasing quality and managing complexity in neuroinformatics software development with continuous integration

    Directory of Open Access Journals (Sweden)

    Yury V. Zaytsev

    2013-01-01

    Full Text Available High quality neuroscience research requires accurate, reliable and well maintained neuroinformatics applications. As software projects become larger, offering more functionality and developing a denser web of interdependence between their component parts, we need more sophisticated methods to manage their complexity. If complexity is allowed to get out of hand, either the quality of the software or the speed of development suffer, and in many cases both. To address this issue, here we develop a scalable, low-cost and open source solution for continuous integration (CI, a technique which ensures the quality of changes to the code base during the development procedure, rather than relying on a pre-release integration phase. We demonstrate that a CI based workflow, due to rapid feedback about code integration problems and tracking of code health measures, enabled substantial increases in productivity for a major neuroinformatics project and additional benefits for three further projects. Beyond the scope of the current study, we identify multiple areas in which CI can be employed to further increase the quality of neuroinformatics projects by improving development practices and incorporating appropriate development tools. Finally, we discuss what measures can be taken to lower the barrier for developers of neuroinformatics applications to adopt this useful technique.

  7. PREDICTION OF SMARTPHONES’ PERCEIVED IMAGE QUALITY USING SOFTWARE EVALUATION TOOL VIQET

    Directory of Open Access Journals (Sweden)

    Pinchas ZOREA

    2016-12-01

    Full Text Available A great deal of resources and efforts have been made in recent years to assess how the smartphones users perceived the image quality. Unfortunately, only limited success has been achieved and the image quality assessment still based on many physical human visual test. The paper describes the new model proposed for perceived quality based on human visual tests compared with image analysis by the software application tool. The values of parameters of perceived image quality (brightness, contrast, color saturation and sharpness were calibrated based on results from human visual experiments.PREDICŢIA CALITĂŢII PERCEPUTE A IMAGINILOR AFIȘATE DE SMARTPHONE-URI UTILIZÂND APLICAŢIA DE EVALUARE VIQETÎn ultimii ani au fost depuse eforturi semnificative pentru a evalua modul în care utilizatorii de smartphone  percep calitatea imaginilor. Din păcate, a fost atins doar un progres limitat, evaluarea calităţii imaginiilor bazându-se încă pe multiple teste vizuale umane. În lucrare este descris un nou model al calităţii percepute pe baza testelor vizuale umane, comparate cu analiza imaginii efectuate cu o aplicaţie software. Valorile parametrilor calităţii  percepute a imaginii (lu­minozitate, contrast, saturaţia culorilor şi claritatea au fost calibrate pe baza rezultatelor experimentelor vizuale umane.

  8. Specific developed phantoms and software to assess radiological equipment image quality

    Energy Technology Data Exchange (ETDEWEB)

    Verdu, G., E-mail: gverdu@iqn.upv.es [Universidad Politecnica de Valencia (Spain). Dept. de Ingenieria Quimica y Nuclear; Mayo, P., E-mail: p.mayo@titaniast.com [TITANIA Servicios Teconologicos, Valencia (Spain); Rodenas, F., E-mail: frodenas@mat.upv.es [Universidad Politecnica de Valencia (Spain). Dept. de Matematica Aplicada; Campayo, J.M., E-mail: j.campayo@lainsa.com [Logistica y Acondicionamientos Industriales S.A.U (LAINSA), Valencia (Spain)

    2011-07-01

    The use of radiographic phantoms specifically designed to evaluate the operation of the radiographic equipment lets the study of the image quality obtained by this equipment in an objective way. In digital radiographic equipment, the analysis of the image quality can be automatized because the acquisition of the image is possible in different technologies that are, computerized radiography or phosphor plate and direct radiography or detector. In this work we have shown an application to assess automatically the constancy quality image in the image chain of the radiographic equipment. This application is integrated by designed radiographic phantoms which are adapted to conventional, dental equipment and specific developed software for the automatic evaluation of the phantom image quality. The software is based on digital image processing techniques that let the automatic detection of the different phantom tests by edge detector, morphological operators, threshold histogram techniques, etc. The utility developed is enough sensitive to the radiographic equipment of operating conditions of voltage (kV) and charge (mAs). It is a friendly user programme connected with a data base of the hospital or clinic where it has been used. After the phantom image processing the user can obtain an inform with a resume of the imaging system state with accepting and constancy results. (author)

  9. Investigating the Impact of Learning Organization on Implementing Total Quality Management in Karaj Telecommunication Company

    Directory of Open Access Journals (Sweden)

    Arash Shahin

    2013-09-01

    Full Text Available The aim of this study is to investigate the impact of learning organization on implementing Total Quality Management (TQM. For this purpose, 110 people out of the statistical population of 1210 employees of Karaj Telecommunication Company have been selected using classified probable sampling to respond to the questionnaire of this survey. The tools used in this survey include (Marsick and Watkins, 2003 standard questionnaire of learning organization dimensions and Zhang’s justified questionnaire related to TQM. The validity and reliability of utilized tools have been approved. In order to analyze the results, the SPSS software has been used. Correlation and multi regression analysis have been utilized for data analysis. Major findings imply that there is a positive and significant correlation between learning organization dimensions and seven dimensions considered for TQM, indicating that learning organization significantly impacts on TQM.

  10. Quality Criteria Implementation for Brain Computed Tomography Examinations

    Energy Technology Data Exchange (ETDEWEB)

    Calzado, A.; Rodriguez, R.; Munoz, A

    1998-07-01

    The main aim of this study was to implement the quality criteria proposed by the European Commission for brain computed tomography (CT) examinations. The proposed criteria were evaluated for 102 brain CT examinations, with a special emphasis on the diagnostic and radiation dose requirements. The examinations were selected at random from brain examinations performed over a period of a month at eight scanners of the CT Pace range. Achievement of image criteria was evaluated by two independent observers. The most important preliminary findings concerning image criteria fulfilment or lack of it and disagreements between observers are presented and discussed. The mean values of the proposed dosimetric quantities are compared to the reference values. The examinations performed with and without injection of a contrast agent are also analysed in relation to dosimetric quantities and criteria fulfilment. (author)

  11. ThermoData Engine (TDE): software implementation of the dynamic data evaluation concept. 5. Experiment planning and product design.

    Science.gov (United States)

    Diky, Vladimir; Chirico, Robert D; Kazakov, Andrei F; Muzny, Chris D; Magee, Joseph W; Abdulagatov, Ilmutdin; Kang, Jeong Won; Kroenlein, Kenneth; Frenkel, Michael

    2011-01-24

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported recently in this journal. In the present paper, we describe development of an algorithmic approach to assist experiment planning through assessment of the existing body of knowledge, including availability of experimental thermophysical property data, variable ranges studied, associated uncertainties, state of prediction methods, and parameters for deployment of prediction methods and how these parameters can be obtained using targeted measurements, etc., and, indeed, how the intended measurement may address the underlying scientific or engineering problem under consideration. A second new feature described here is the application of the software capabilities for aid in the design of chemical products through identification of chemical systems possessing desired values of thermophysical properties within defined ranges of tolerance. The algorithms and their software implementation to achieve this are described. Finally, implementation of a new data validation and weighting system is described for vapor-liquid equilibrium (VLE) data, and directions for future enhancements are outlined.

  12. WebStruct and VisualStruct: Web interfaces and visualization for Structure software implemented in a cluster environment.

    Science.gov (United States)

    Jayashree, B; Rajgopal, S; Hoisington, D; Prasanth, V P; Chandra, S

    2008-09-24

    Structure, is a widely used software tool to investigate population genetic structure with multi-locus genotyping data. The software uses an iterative algorithm to group individuals into "K" clusters, representing possibly K genetically distinct subpopulations. The serial implementation of this programme is processor-intensive even with small datasets. We describe an implementation of the program within a parallel framework. Speedup was achieved by running different replicates and values of K on each node of the cluster. A web-based user-oriented GUI has been implemented in PHP, through which the user can specify input parameters for the programme. The number of processors to be used can be specified in the background command. A web-based visualization tool "Visualstruct", written in PHP (HTML and Java script embedded), allows for the graphical display of population clusters output from Structure, where each individual may be visualized as a line segment with K colors defining its possible genomic composition with respect to the K genetic sub-populations. The advantage over available programs is in the increased number of individuals that can be visualized. The analyses of real datasets indicate a speedup of up to four, when comparing the speed of execution on clusters of eight processors with the speed of execution on one desktop. The software package is freely available to interested users upon request.

  13. A software tool for quality assurance of computed/digital radiography (CR/DR) systems

    Science.gov (United States)

    Desai, Nikunj; Valentino, Daniel J.

    2011-03-01

    The recommended methods to test the performance of computed radiography (CR) systems have been established by The American Association of Physicists in Medicine, Report No. 93, "Acceptance Testing and Quality Control of Photostimulable Storage Phosphor Imaging Systems". The quality assurance tests are categorized by how frequently they need to be performed. Quality assurance of CR systems is the responsibility of the facility that performs the exam and is governed by the state in which the facility is located. For Example, the New York State Department of Health has established a guide which lists the tests that a CR facility must perform for quality assurance. This study aims at educating the reader about the new quality assurance requirements defined by the state. It further demonstrates an easy to use software tool, henceforth referred to as the Digital Physicist, developed to aid a radiologic facility in conforming with state guidelines and monitoring quality assurance of CR/DR imaging systems. The Digital Physicist provides a vendor independent procedure for quality assurance of CR/DR systems. Further it, generates a PDF report with a brief description of these tests and the obtained results.

  14. NASE 2015: Implementation of a Management Quality System

    Science.gov (United States)

    Ros, Rosa Maria; García, Beatriz; Santa Maria, Delia

    2015-08-01

    ISO 29990:2010, Learning services for non-formal education and training - Basic requirements for service providers, proposes enhance transparency and allow comparison on a worldwide basis of learning services, offering a single alternative backed by international consensus to the huge variety of national service and management standards which now exists in the field of non-formal learning. The objective of ISO 29990:2010 is to provide a generic model for quality professional practice and performance, and a common reference for learning service providers and their clients in the design, development and delivery of non-formal education, training and development.NASE project, which was born as a consequence of the implementation of IAU-10 Years Strategic Plan and it is in agreement with the proposal to increase the interest and efforts of IAU on schools before the university and constitutes good objective for the implementation of ISO 29990-2010.Taking into account the main objectives of the Strategic Plan, it is possible to remark that NASE acts on all the education levels, and had, during the past 5 years a big impact because Ministries of Education, Universities and Planetariums are involved. After 5 years NASE organized more than 60 courses in about 20 countries distributed in America, Africa, Asia and Europe. In many countries local governments changed their curriculum according NASE course contents (i.e. Nicaragua, Argentina and Honduras).The knowledge obtained by teachers is evaluated by the department of education of the local government or universities who participated. NASE acts as part of the Professional formation of Professors in Science and in this sense is a Program directed to the Tertiary and University Education.After 5 years of development, we present the implementation of a Quality Management System, according to ISO 29990:2010, devoted to Learning services for non-formal education and training. Basic requirements for service providers, and show

  15. Flexible Self-Managing Pipe-line Framework Reducing Development Risk to Improve Software Quality

    Directory of Open Access Journals (Sweden)

    Nitin Deepak

    2015-06-01

    Full Text Available Risk identification and assessment in today‘s scenario play a vital role in any software/web application development industry. Many process models deliver the process related to development life cycle, but the risk assessment at an early stage is still an issue and is a significant subject for research. In this paper, an approach based on MVC architecture by embedding spiral process, which is verified and validated by V-shape model is proposed. By using this approach development efficiency will increase due to less burdened working team(s, reduces stressful maintenance effort that causes reduction in risk factors because of beautifully distributed human effort to improve software quality. Besides, the efficiency of our approach is manifested by the preliminary experiment.

  16. Software quality assurance documentation for the release of NUFT 2.0s for SUN platforms

    Energy Technology Data Exchange (ETDEWEB)

    Fernandez, M.W.; Shaffer, R.J.

    1998-10-01

    The purpose of this Individual Software Plan (ISP) is to plan and organize the activities required to certify the NUFT code for quality affecting work involving problems that include cross drift analysis of the Yucca Mountain Repository facility. NUFT is software for application to the solution of a class of coupled mass and heat transport problems in porous geologic media including Yucca Mountain Repository Cross Drift Problem (YMRCDP- also known as the Enhanced Characterization of the Repository Block (ECRB)). Solution of this class of problems requires a suite of multiphase, multi-component models for numerical solution of non-isothermal flow and transport in porous media with applications to subsurface contaminant transport problems.

  17. Implementing clinical governance in English primary care groups/trusts: reconciling quality improvement and quality assurance

    Science.gov (United States)

    Campbell, S; Sheaff, R; Sibbald, B; Marshall, M; Pickard, S; Gask, L; Halliwell, S; Rogers, A; Roland, M

    2002-01-01

    Objectives: To investigate the concept of clinical governance being advocated by primary care groups/trusts (PCG/Ts), approaches being used to implement clinical governance, and potential barriers to its successful implementation in primary care. Design: Qualitative case studies using semi-structured interviews and documentation review. Setting: Twelve purposively sampled PCG/Ts in England. Participants: Fifty senior staff including chief executives, clinical governance leads, mental health leads, and lay board members. Main outcome measures: Participants' perceptions of the role of clinical governance in PCG/Ts. Results: PCG/Ts recognise that the successful implementation of clinical governance in general practice will require cultural as well as organisational changes, and the support of practices. They are focusing their energies on supporting practices and getting them involved in quality improvement activities. These activities include, but move beyond, conventional approaches to quality assessment (audit, incentives) to incorporate approaches which emphasise corporate and shared learning. PCG/Ts are also engaged in setting up systems for monitoring quality and for dealing with poor performance. Barriers include structural barriers (weak contractual levers to influence general practices), resource barriers (perceived lack of staff or money), and cultural barriers (suspicion by practice staff or problems overcoming the perceived blame culture associated with quality assessment). Conclusion: PCG/Ts are focusing on setting up systems for implementing clinical governance which seek to emphasise developmental and supportive approaches which will engage health professionals. Progress is intentionally incremental but formidable challenges lie ahead, not least reconciling the dual role of supporting practices while monitoring (and dealing with poor) performance. PMID:12078380

  18. Empirically Validating an Analytical Method for Assessing the Impact of Design Patterns on Software Quality : Three Case Studies

    NARCIS (Netherlands)

    Ampatzoglou, Apostolos; Avgeriou, Paris; Arvanitou, Elvira Maria; Stamelos, Ioannis

    2013-01-01

    This technical report has been created as support material for the paper entitled “Empirically Validating an Analytical Method for Assessing the Impact of Design Patterns on Software Quality: A Case Study” that has been submitted in ACM Transactions on Software Engineering. The corresponding paper a

  19. Implementation of the boundary element method in a practical reservoir engineering software application

    Energy Technology Data Exchange (ETDEWEB)

    Kryuchkov, S.; Sanger, S.; Barden, R. [Vertex Petroleum Systems, Englewood, CO (United States)

    2001-06-01

    The mathematical basis of a newly developed reservoir modeling software based on the Boundary Element Method (BEM) was presented. The software includes a fully graphical interface which provides accurate and fast solutions for most engineering problems. The model capabilities include modeling of arbitrary shaped heterogenous oil and gas reservoirs with fractured, radial and horizontal wells. In addition, the software can be used to model water injection and edge water drive. The model is suitable for managing small and midsize oil and gas fields, and is particularly useful for performing case studies at each field in real time. A comparison was also conducted between the BEM model and other well known analytical solutions such as steady state and transient solutions for standard reservoirs. Results showed good agreement between the two modeling methods. for vertical, fractured and horizontal wells. 24 refs., 8 figs.

  20. Design and Implementation of an Efficient Software Communications Architecture Core Framework for a Digital Signal Processors Platform

    Directory of Open Access Journals (Sweden)

    Wael A. Murtada

    2011-01-01

    Full Text Available Problem statement: The Software Communications Architecture (SCA was developed to improve software reuse and interoperability in Software Defined Radios (SDR. However, there have been performance concerns since its conception. Arguably, the majority of the problems and inefficiencies associated with the SCA can be attributed to the assumption of modular distributed platforms relying on General Purpose Processors (GPPs to perform all signal processing. Approach: Significant improvements in cost and power consumption can be obtained by utilizing specialized and more efficient platforms. Digital Signal Processors (DSPs present such a platform and have been widely used in the communications industry. Improvements in development tools and middleware technology opened the possibility of fully integrating DSPs into the SCA. This approach takes advantage of the exceptional power, cost and performance characteristics of DSPs, while still enjoying the flexibility and portability of the SCA. Results: This study presents the design and implementation of an SCA Core Framework (CF for a TI TMS320C6416 DSP. The framework is deployed on a C6416 Device Cycle Accurate Simulator and TI C6416 Development board. The SCA CF is implemented by leveraging OSSIE, an open-source implementation of the SCA, to support the DSP platform. OIS’s ORBExpress DSP and DSP/BIOS are used as the middleware and operating system, respectively. A sample waveform was developed to demonstrate the framework’s functionality. Benchmark results for the framework and sample applications are provided. Conclusion: Benchmark results show that, using OIS ORBExpress DSP ORB middleware has an impact for decreasing the Software Memory Footprint and increasing the System Performance compared with PrismTech's e*ORB middleware.

  1. Software-Defined GPS Receiver Implemented on the Parallella-16 Board

    DEFF Research Database (Denmark)

    Olesen, Daniel Madelung; Jakobsen, Jakob; Knudsen, Per

    2015-01-01

    This paper describes a GPS software receiver design made of inexpensive and physically small hardware components. The small embedded platform, known as the Parallella-16 computer has been utilized in conjunction with a commercial RF front-end to construct a 4-channel real time software GPS receiv....... The total cost of the hardware is below 150$ and the size is comparable to a credit-card. The receiver has been developed for research in GNSS/INS integration on small Unmanned Aerial Vehicles (UAVs)....

  2. PHASIM, a sophisticated phased array antenna software simulator implemented in MATLAB 5.2

    NARCIS (Netherlands)

    Keizer, W.P.M.N.

    1999-01-01

    A sophisticated phased array simulator software package for the design and analysis of planar phased array antennas is presented. This simulator is coded in MATLAB version 5.2. Using MATLAB, numerical engineering problems can be solved in a fraction of time of time required by programs coded in FORT

  3. PHASIM, a sophisticated phased array antenna software simulator implemented in MATLAB 5.2

    NARCIS (Netherlands)

    Keizer, W.P.M.N.

    1999-01-01

    A sophisticated phased array simulator software package for the design and analysis of planar phased array antennas is presented. This simulator is coded in MATLAB version 5.2. Using MATLAB, numerical engineering problems can be solved in a fraction of time of time required by programs coded in FORT

  4. Improvement of CSCW Software Implementation in NPD: The CAM Mechanism for a Better Adoption by Users

    Science.gov (United States)

    Restrepo, Tomas; Arbelaez, Natalia; Millet, Dominique; Gidel, Thierry

    2010-01-01

    Cooperation between disseminated actors is a key factor in improving new product development (NPD) performance. In the last years, numerous CSCW software applications have been introduced in the industry to support NPD with a low success rate. This is partly due to the limited insight of the organisational and human factors influencing user…

  5. Optimal parameters for clinical implementation of breast cancer patient setup using Varian DTS software.

    Science.gov (United States)

    Ng, Sook Kien; Zygmanski, Piotr; Jeung, Andrew; Mostafavi, Hassan; Hesser, Juergen; Bellon, Jennifer R; Wong, Julia S; Lyatskaya, Yulia

    2012-05-10

    Digital tomosynthesis (DTS) was evaluated as an alternative to cone-beam computed tomography (CBCT) for patient setup. DTS is preferable when there are constraints with setup time, gantry-couch clearance, and imaging dose using CBCT. This study characterizes DTS data acquisition and registration parameters for the setup of breast cancer patients using nonclinical Varian DTS software. DTS images were reconstructed from CBCT projections acquired on phantoms and patients with surgical clips in the target volume. A shift-and-add algorithm was used for DTS volume reconstructions, while automated cross-correlation matches were performed within Varian DTS software. Triangulation on two short DTS arcs separated by various angular spread was done to improve 3D registration accuracy. Software performance was evaluated on two phantoms and ten breast cancer patients using the registration result as an accuracy measure; investigated parameters included arc lengths, arc orientations, angular separation between two arcs, reconstruction slice spacing, and number of arcs. The shifts determined from DTS-to-CT registration were compared to the shifts based on CBCT-to-CT registration. The difference between these shifts was used to evaluate the software accuracy. After findings were quantified, optimal parameters for the clinical use of DTS technique were determined. It was determined that at least two arcs were necessary for accurate 3D registration for patient setup. Registration accuracy of 2 mm was achieved when the reconstruction arc length was > 5° for clips with HU ≥ 1000; larger arc length (≥ 8°) was required for very low HU clips. An optimal arc separation was found to be ≥ 20° and optimal arc length was 10°. Registration accuracy did not depend on DTS slice spacing. DTS image reconstruction took 10-30 seconds and registration took less than 20 seconds. The performance of Varian DTS software was found suitable for the accurate setup of breast cancer patients

  6. Air quality improvements following implementation of Lisbon's Low Emission Zone

    Science.gov (United States)

    Ferreira, F.; Gomes, P.; Tente, H.; Carvalho, A. C.; Pereira, P.; Monjardino, J.

    2015-12-01

    Air pollution levels within Lisbon city limits have been exceeding the limit values established in European Union and national legislation since 2001, with the most problematic cases related to the levels of fine particles (PM10) and nitrogen dioxide (NO2), mainly originated by road traffic. With the objective of answering this public health issue, an Air Quality Action Plan was developed in 2006 and the respective Enforcement Plan was published in 2009. From the overall strategy, one of the major measures presented in this strategy was the creation of a Low Emission Zone (LEZ) in Lisbon, which has been operating since July 2011. Implemented at different stages it has progressively expanded its area, including more vehicle types and adopting more stringent requirements in terms of minimum emission standards (currently LEZ phase 2 with EURO 2 in the city center - zone 1 and EURO 1 in the rest of the LEZ area - zone 2). At the same time the road axis comprised of Marquês de Pombal square and Avenida da Liberdade was subjected to profound changes in its traffic circulation model, reducing road traffic volumes. The analysis of the air quality data before and after the LEZ phase 2 has shown positive evolution when comparing the period between 2011 (before measures) and 2013 (after measures). In 2013, there was a reduction in PM10 annual average concentration of 23% and NO2 annual average concentrations of 12%, compared with the year 2011. Although PM10 reductions were more significant inside the LEZ area, the same was not valid for NO2, suggesting that the implementation of these measures was not as effective in reducing NO2 levels as shown by results in other cities like Berlin and London. The results from road traffic characterization indicate a relevant effect on fleet renewal with an overall decrease in the relative weight of pre-EURO 2 vehicles in 2012/2013, compared with data from 2011. An important increase in the share of EURO 4 and EURO 5 vehicles was also

  7. 78 FR 22840 - Approval and Promulgation of Air Quality Implementation Plans; Delaware; State Board Requirements

    Science.gov (United States)

    2013-04-17

    ... requirements of the Clean Air Act (CAA) for all criteria pollutants of the national ambient air quality... Promulgation of Air Quality Implementation Plans; Delaware; State Board Requirements,'' that is located in the... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Delaware; State...

  8. 76 FR 14584 - Approval and Promulgation of Air Quality Implementation Plans; Montana; Attainment Plan for Libby...

    Science.gov (United States)

    2011-03-17

    ... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Montana; Attainment... ) national ambient air quality standard (NAAQS) for the Libby nonattainment area. The plan revision, herein... and transit projects are consistent with (``conform to'') the state air quality implementation plan...

  9. An evaluation approach to software quality based on quality cloud%一种基于质量云的软件质量评估方法

    Institute of Scientific and Technical Information of China (English)

    周伟良; 辜丽川; 张友华

    2011-01-01

    软件质量评估是软件质量保证的关键技术,具有多属性、模糊性和不确定性的特点.基于云模型在非规范知识的定性、定量表示及其相互转换过程中的优良特征,提出了一种基于云模型的软件质量量化评估方法.该方法使用软件质量云的期望和超熵对软件质量进行定量评估,利用质量变化云刻划软件质量的变化情况,为进一步的质量评估提供依据.实验结果表明,该方法能够有效地支持软件质量评估决策过程,在有效性和精确度上有显著优势.%Quality evaluation is the key technology for software quality assurance, with characteristics of fuzziness, uncertainty and multi-attribute.In view of the outstanding characteristics of the cloud model during the process of transforming a qualitative concept into a set of quantitative numerical values, a novel software quality evaluation method based on quality cloud and quality change cloud is proposed.The expected value and hyper entropy of software quality cloud is used to evaluate software quality, and a software quality change cloud model is designed to depict changes in software quality so as to provide additional evidence.The empirical study indicates that the proposed method is effective in assessing software quality, with an apparent advantage in effectiveness and precision.

  10. The role of metrics and measurements in a software intensive total quality management environment

    Science.gov (United States)

    Daniels, Charles B.

    1992-01-01

    Paramax Space Systems began its mission as a member of the Rockwell Space Operations Company (RSOC) team which was the successful bidder on a massive operations consolidation contract for the Mission Operations Directorate (MOD) at JSC. The contract awarded to the team was the Space Transportation System Operations Contract (STSOC). Our initial challenge was to accept responsibility for a very large, highly complex and fragmented collection of software from eleven different contractors and transform it into a coherent, operational baseline. Concurrently, we had to integrate a diverse group of people from eleven different companies into a single, cohesive team. Paramax executives recognized the absolute necessity to develop a business culture based on the concept of employee involvement to execute and improve the complex process of our new environment. Our executives clearly understood that management needed to set the example and lead the way to quality improvement. The total quality management policy and the metrics used in this endeavor are presented.

  11. An Evaluation of Output Quality of Machine Translation (Padideh Software vs. Google Translate

    Directory of Open Access Journals (Sweden)

    Haniyeh Sadeghi Azer

    2015-08-01

    Full Text Available This study aims to evaluate the translation quality of two machine translation systems in translating six different text-types, from English to Persian. The evaluation was based on criteria proposed by Van Slype (1979. The proposed model for evaluation is a black-box type, comparative and adequacy-oriented evaluation. To conduct the evaluation, a questionnaire was assigned to end-users to evaluate the outputs to examine and determine, if the machine-generated translations are intelligible and acceptable from their point of view and which one of the machine-generated translations produced by Padideh software and Google Translate is more acceptable and useful from the end-users point of view. The findings indicate that, the machine-generated translations are intelligible and acceptable in translating certain text-types, for end-users and Google Translate is more acceptable from end-users point of view. Keywords: Machine Translation, Machine Translation Evaluation, Translation Quality

  12. An implementation of Software Defined Radios for federated aerospace networks: Informing satellite implementations using an inter-balloon communications experiment

    Science.gov (United States)

    Akhtyamov, Rustam; Cruz, Ignasi Lluch i.; Matevosyan, Hripsime; Knoll, Dominik; Pica, Udrivolf; Lisi, Marco; Golkar, Alessandro

    2016-06-01

    Novel space mission concepts such as Federated Satellite Systems promise to enhance sustainability, robustness, and reliability of current missions by means of in-orbit sharing of space assets. This new paradigm requires the utilization of several technologies in order to confer flexibility and re-configurability to communications systems among heterogeneous spacecrafts. This paper illustrates the results of the experimental demonstration of the value proposition of federated satellites through two stratospheric balloons interoperating with a tracking ground station through Commercial Off-The-Shelf Software Defined Radios (SDRs). The paper reports telemetry analysis and characterizes the communications network that was realized in-flight. Furthermore, it provides details on an in-flight anomaly experienced by one of the balloons, which was recovered through the use of the federated technology that has been developed. The anomaly experienced led to the early loss of the directional link from the ground station to the affected stratospheric balloon node after 15 min in flight. Nevertheless, thanks to the federated approach among the systems, the ground station was still able to retrieve the balloon's data in real time through the network system, for which the other balloon operated as a federated relay for 45 min in flight, uninterrupted. In other words, the federated approach to the system allowed triplicating the useful lifetime of the defective system, which would have not been possible to realize otherwise. Such anomaly coincidentally demonstrated the value of the federated approach to space systems design. The paper paves the way for future tests on space assets.

  13. SIFT - Design and analysis of a fault-tolerant computer for aircraft control. [Software Implemented Fault Tolerant systems

    Science.gov (United States)

    Wensley, J. H.; Lamport, L.; Goldberg, J.; Green, M. W.; Levitt, K. N.; Melliar-Smith, P. M.; Shostak, R. E.; Weinstock, C. B.

    1978-01-01

    SIFT (Software Implemented Fault Tolerance) is an ultrareliable computer for critical aircraft control applications that achieves fault tolerance by the replication of tasks among processing units. The main processing units are off-the-shelf minicomputers, with standard microcomputers serving as the interface to the I/O system. Fault isolation is achieved by using a specially designed redundant bus system to interconnect the processing units. Error detection and analysis and system reconfiguration are performed by software. Iterative tasks are redundantly executed, and the results of each iteration are voted upon before being used. Thus, any single failure in a processing unit or bus can be tolerated with triplication of tasks, and subsequent failures can be tolerated after reconfiguration. Independent execution by separate processors means that the processors need only be loosely synchronized, and a novel fault-tolerant synchronization method is described.

  14. Software Quality and Security in Teachers' and Students' Codes When Learning a New Programming Language

    Directory of Open Access Journals (Sweden)

    Arnon Hershkovitz

    2015-09-01

    Full Text Available In recent years, schools (as well as universities have added cyber security to their computer science curricula. This topic is still new for most of the current teachers, who would normally have a standard computer science background. Therefore the teachers are trained and then teaching their students what they have just learned. In order to explore differences in both populations’ learning, we compared measures of software quality and security between high-school teachers and students. We collected 109 source files, written in Python by 18 teachers and 31 students, and engineered 32 features, based on common standards for software quality (PEP 8 and security (derived from CERT Secure Coding Standards. We use a multi-view, data-driven approach, by (a using hierarchical clustering to bottom-up partition the population into groups based on their code-related features and (b building a decision tree model that predicts whether a student or a teacher wrote a given code (resulting with a LOOCV kappa of 0.751. Overall, our findings suggest that the teachers’ codes have a better quality than the students’ – with a sub-group of the teachers, mostly males, demonstrate better coding than their peers and the students – and that the students’ codes are slightly better secured than the teachers’ codes (although both populations show very low security levels. The findings imply that teachers might benefit from their prior knowledge and experience, but also emphasize the lack of continuous involvement of some of the teachers with code-writing. Therefore, findings shed light on computer science teachers as lifelong learners. Findings also highlight the difference between quality and security in today’s programming paradigms. Implications for these findings are discussed.

  15. SU-E-T-27: A Tool for Routine Quality Assurance of Radiotherapy Dose Calculation Software

    Energy Technology Data Exchange (ETDEWEB)

    Popple, R; Cardan, R; Duan, J; Wu, X; Shen, S; Brezovich, I [The University of Alabama at Birmingham, Birmingham, AL (United States)

    2014-06-01

    Purpose: Dose calculation software is thoroughly evaluated when it is commissioned; however, evaluation of periodic software updates is typically limited in scope due to staffing constraints and the need to quickly return the treatment planning system to clinical service. We developed a tool for quickly and comprehensively testing and documenting dose calculation software against measured data. Methods: A tool was developed using MatLab (The MathWorks, Natick, MA) for evaluation of dose calculation algorithms against measured data. Inputs to the tool are measured data, reference DICOM RT PLAN files describing the measurements, and dose calculations in DICOM format. The tool consists of a collection of extensible modules that can perform analysis of point dose, depth dose curves, and profiles using dose difference, distance-to-agreement, and the gamma-index. Each module generates a report subsection that is incorporated into a master template, which is converted to final form in portable document format (PDF). Results: After each change to the treatment planning system, a report can be generated in approximately 90 minutes. The tool has been in use for more than 5 years, spanning 5 versions of the eMC and 4 versions of the AAA. We have detected changes to the algorithms that affected clinical practice once during this period. Conclusion: Our tool provides an efficient method for quality assurance of dose calculation software, providing a complete set of tests for an update. Future work includes the addition of plan level tests, allowing incorporation of, for example, the TG-119 test suite for IMRT, and integration with the treatment planning system via an application programming interface. Integration with the planning system will permit fully-automated testing and reporting at scheduled intervals.

  16. genepop'007: a complete re-implementation of the genepop software for Windows and Linux.

    Science.gov (United States)

    Rousset, François

    2008-01-01

    This note summarizes developments of the genepop software since its first description in 1995, and in particular those new to version 4.0: an extended input format, several estimators of neighbourhood size under isolation by distance, new estimators and confidence intervals for null allele frequency, and less important extensions to previous options. genepop now runs under Linux as well as under Windows, and can be entirely controlled by batch calls.

  17. The design and implementation of pedagogical software for multi-backend/multi-lingual database system.

    OpenAIRE

    Little, Craig W.

    1987-01-01

    Approved for public release; distribution is unlimited Traditionally, courses in database systems do not use pedagogical software for the purpose of instructing the database systems, despite the progress made in modem database architecture. In this thesis, we present a working document to assist in the instruction of a new database system, the Multi-Backend Database System (MBDS)-and the Multi-Lingual Database System (MLDS). As the course of instruction describes the creatio...

  18. Implementation of complex digital PLL for phase detection in software defined radar

    OpenAIRE

    2016-01-01

    Software defined radar (SDR) has been the latest trend in developing enhanced radar signal processing techniques for state-of-the-art radar systems. SDR provides tremendous flexibility in reconfigurable design and rapid prototyping capabilities on FPGA platform. To cater real-time processing for high-speed radar, COordinate Rotation Digital Computer (CORDIC) unit has been utilized as a core processing element in a complex digital phase locked loop (DPLL) for digital demodulation of received s...

  19. Global Forest Products Model software design and implementation (GFPM version 2014 with BPMP)

    Science.gov (United States)

    Shushuai Zhu; James Turner; Joseph   Buongiorno

    2014-01-01

    An overview of the GFPM software structure is given in Section 1.1 in terms of the overall processing flows and the main components of the GFPM. Section 1.2 describes the role of batch files in controlling the execution of the GFPM programs, and details of the sequence of program execution corresponding to each of the “Main Menu” options of the GFPM. Next, each...

  20. ParseCNV integrative copy number variation association software with quality tracking.

    Science.gov (United States)

    Glessner, Joseph T; Li, Jin; Hakonarson, Hakon

    2013-03-01

    A number of copy number variation (CNV) calling algorithms exist; however, comprehensive software tools for CNV association studies are lacking. We describe ParseCNV, unique software that takes CNV calls and creates probe-based statistics for CNV occurrence in both case-control design and in family based studies addressing both de novo and inheritance events, which are then summarized based on CNV regions (CNVRs). CNVRs are defined in a dynamic manner to allow for a complex CNV overlap while maintaining precise association region. Using this approach, we avoid failure to converge and non-monotonic curve fitting weaknesses of programs, such as CNVtools and CNVassoc, and although Plink is easy to use, it only provides combined CNV state probe-based statistics, not state-specific CNVRs. Existing CNV association methods do not provide any quality tracking information to filter confident associations, a key issue which is fully addressed by ParseCNV. In addition, uncertainty in CNV calls underlying CNV associations is evaluated to verify significant results, including CNV overlap profiles, genomic context, number of probes supporting the CNV and single-probe intensities. When optimal quality control parameters are followed using ParseCNV, 90% of CNVs validate by polymerase chain reaction, an often problematic stage because of inadequate significant association review. ParseCNV is freely available at http://parsecnv.sourceforge.net.

  1. 75 FR 18782 - Approval and Promulgation of Air Quality Implementation Plans; Indiana; Alternate Monitoring...

    Science.gov (United States)

    2010-04-13

    ... AGENCY 40 CFR Part 52 Approval and Promulgation of Air Quality Implementation Plans; Indiana; Alternate Monitoring Requirements for Indianapolis Power and Light--Harding Street Station AGENCY: Environmental... State Implementation Plan alternative monitoring requirements for Indianapolis Power and Light...

  2. Implementation of quality management systems: the role of hospital (management) boards.

    NARCIS (Netherlands)

    Botje, D.; Klazinga, N.S.; Suñol, R.; Wagner, C.

    2013-01-01

    Objectives: Hospitals are putting tremendous efforts in implementing evidence-based management systems and organisational innovations for patient-centred care. Having a hospital quality management system is a prerequisite to successfully implement these innovations. Previous studies showed that the

  3. Knowledge work productivity effect on quality of knowledge work in software development process in SME

    Science.gov (United States)

    Yusoff, Mohd Zairol; Mahmuddin, Massudi; Ahmad, Mazida

    2016-08-01

    Knowledge and skill are necessary to develop the capability of knowledge workers. However, there is very little understanding of what the necessary knowledge work (KW) is, and how they influence the quality of knowledge work or knowledge work productivity (KWP) in software development process, including that in small and medium-sized (SME) enterprise. The SME constitutes a major part of the economy and it has been relatively unsuccessful in developing KWP. Accordingly, this paper seeks to explore the influencing dimensions of KWP that effect on the quality of KW in SME environment. First, based on the analysis of the existing literatures, the key characteristics of KW productivity are defined. Second, the conceptual model is proposed, which explores the dimensions of the KWP and its quality. This study analyses data collected from 150 respondents (based on [1], who involve in SME in Malaysia and validates the models by using structural equation modeling (SEM). The results provide an analysis of the effect of KWP on the quality of KW and business success, and have a significant relevance for both research and practice in the SME

  4. An Implementation Model for Production Quality Management in ...

    African Journals Online (AJOL)

    PROF. O. E. OSUAGWU

    2013-06-01

    Jun 1, 2013 ... Key words: Quality, Management, Iranian Industrial Firms, production. Introduction ... services and products with high quality ... these needs have important effects on quality for ... quality service to customers. ... in the process of analysis, enterprises try to ... through gaining satisfaction of consumers.

  5. Software quality assurance documentation for the release of NUFT 2.0 for HP platforms

    Energy Technology Data Exchange (ETDEWEB)

    Fernandez, Michael W. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Johnson, Gary L. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Preckshot, Gary G. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    1998-08-28

    This document is the Individual Software Plan (ISP) for version 2.0 of the Non-isothermal Unsaturated-saturated Flow and Transport (NUFT.) analysis computer program. This document addresses the applicable requirements of LLNL YMP procedure 033-YMP-QP 3.2, Section 4.2.1.1. The purpose of this ISP is to plan and organize the activities required to certify the NUFT code for quality affecting work involving problems that include cross drift analysis of the Yucca Mountain Repository facility. NUFT is software for application to the solution of a class of coupled mass and heat transport problems in porous geologic media including Yucca Mountain Repository Cross Drift Problem (YMRCDP- also known as the Enhanced Characterization of the Repository Block (ECRB)). Solution of this class of problems requires a suite of multiphase, multi-component models for numerical solution of non- isothermal flow and transport in porous media with applications to subsurface contaminant transport problems. NUFT is a suite of multiphase, multi-component models for numerical solution of non- isothermal flow and transport in porous media, with application to subsurface contaminant transport problems, and in particular, to the hydrology in and about the Yucca Mountain Repository Site. NUFI is acquired software, as defined by 033-YMP-QP 3.2, and a preliminary baseline of source code, electronic documentation, and paper documentation has been established as required by 033-YMP-QP 3.2, Section 4.1. NUFT runs on Sun Unix platforms, Solaris operating system version 5.5 and HP-UX with operating system version 10.20. The product to be qualified under this ISP is the version running on HP- UX. The HP version will be labeled Version 2.0h. The "h" is included to distinguish the HP version from possible future versions qualified for Sun or other platforms. The scope of the plans and procedures outlined in this ISP is limited to the effort required to qualify NUFT for the class of problems identified in

  6. 77 FR 65133 - Revisions to the California State Implementation Plan, Mojave Desert Air Quality Management District

    Science.gov (United States)

    2012-10-25

    ... Identification of plan. * * * * * (c) * * * (379) * * * (i) * * * (E) Mojave Desert Air Quality Management... AGENCY 40 CFR Part 52 Revisions to the California State Implementation Plan, Mojave Desert Air Quality Management District AGENCY: Environmental Protection Agency (EPA). ACTION: Final rule. SUMMARY: EPA...

  7. Quality of radiomic features in glioblastoma multiforme: Impact of semi-automated tumor segmentation software

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Myung Eun; Kim, Jong Hyo [Center for Medical-IT Convergence Technology Research, Advanced Institutes of Convergence Technology, Seoul National University, Suwon (Korea, Republic of); Woo, Bo Yeong [Dept. of Transdisciplinary Studies, Graduate School of Convergence Science and Technology, Seoul National University, Suwon (Korea, Republic of); Ko, Micheal D.; Jamshidi, Neema [Dept. of Radiological Sciences, University of California, Los Angeles, Los Angeles (United States)

    2017-06-15

    The purpose of this study was to evaluate the reliability and quality of radiomic features in glioblastoma multiforme (GBM) derived from tumor volumes obtained with semi-automated tumor segmentation software. MR images of 45 GBM patients (29 males, 16 females) were downloaded from The Cancer Imaging Archive, in which post-contrast T1-weighted imaging and fluid-attenuated inversion recovery MR sequences were used. Two raters independently segmented the tumors using two semi-automated segmentation tools (TumorPrism3D and 3D Slicer). Regions of interest corresponding to contrast-enhancing lesion, necrotic portions, and non-enhancing T2 high signal intensity component were segmented for each tumor. A total of 180 imaging features were extracted, and their quality was evaluated in terms of stability, normalized dynamic range (NDR), and redundancy, using intra-class correlation coefficients, cluster consensus, and Rand Statistic. Our study results showed that most of the radiomic features in GBM were highly stable. Over 90% of 180 features showed good stability (intra-class correlation coefficient [ICC] ≥ 0.8), whereas only 7 features were of poor stability (ICC < 0.5). Most first order statistics and morphometric features showed moderate-to-high NDR (4 > NDR ≥1), while above 35% of the texture features showed poor NDR (< 1). Features were shown to cluster into only 5 groups, indicating that they were highly redundant. The use of semi-automated software tools provided sufficiently reliable tumor segmentation and feature stability; thus helping to overcome the inherent inter-rater and intra-rater variability of user intervention. However, certain aspects of feature quality, including NDR and redundancy, need to be assessed for determination of representative signature features before further development of radiomics.

  8. A Model to Calculate the Return on Investment After a Software Implementation

    Directory of Open Access Journals (Sweden)

    PADUAM, T. C.

    2015-06-01

    Full Text Available Currently the organization has been concerned with the analysis of the impact of IT investments. Economic pressures, combined with years of significant IT spending without demonstrating clear returns, forced companies to improve their financial practices and justify better and more clearly every penny invested way. Thus, this article presents the model to calculate the return on investment after deploying software. This model was generated from two experiments conducted in the laboratory and in the field, applied in southern Brazil, which showed effective action to catch the post-deployment time metrics. Nevertheless, this article may be applicable to all companies wishing to calculate the temporal return from a deployment

  9. Open source software implementation of an integrated testing strategy for skin sensitization potency based on a Bayesian network.

    Science.gov (United States)

    Pirone, Jason R; Smith, Marjolein; Kleinstreuer, Nicole C; Burns, Thomas A; Strickland, Judy; Dancik, Yuri; Morris, Richard; Rinckel, Lori A; Casey, Warren; Jaworska, Joanna S

    2014-01-01

    An open-source implementation of a previously published integrated testing strategy (ITS) for skin sensitization using a Bayesian network has been developed using R, a free and open-source statistical computing language. The ITS model provides probabilistic predictions of skin sensitization potency based on in silico and in vitro information as well as skin penetration characteristics from a published bioavailability model (Kasting et al., 2008). The structure of the Bayesian network was designed to be consistent with the adverse outcome pathway published by the OECD (Jaworska et al., 2011, 2013). In this paper, the previously published data set (Jaworska et al., 2013) is improved by two data corrections and a modified application of the Kasting model. The new data set implemented in the original commercial software package and the new R version produced consistent results. The data and a fully documented version of the code are publicly available (http://ntp.niehs.nih.gov/go/its).

  10. Advanced fingerprint verification software

    Science.gov (United States)

    Baradarani, A.; Taylor, J. R. B.; Severin, F.; Maev, R. Gr.

    2016-05-01

    We have developed a fingerprint software package that can be used in a wide range of applications from law enforcement to public and private security systems, and to personal devices such as laptops, vehicles, and door- locks. The software and processing units are a unique implementation of new and sophisticated algorithms that compete with the current best systems in the world. Development of the software package has been in line with the third generation of our ultrasonic fingerprinting machine1. Solid and robust performance is achieved in the presence of misplaced and low quality fingerprints.

  11. SPC in Software Quality Management%软件质量管理中的统计过程控制

    Institute of Scientific and Technical Information of China (English)

    赖旭芳; 王青

    2002-01-01

    SPC(Statistical Process Control) is a process control method which can help the organizations to meausure and control the quality of the processes,and then to achieve the goal of improving the process stability.Thereby,the quality of products and service is advanced.SPC is used widely and maturely in manufacturing,but it is in inchoative in software engineering.In this paper,we discuss the applying method and foreground of SPC used to manage and control the software quality.Some applying examples are given.

  12. Implementation of a GPU accelerated total focusing reconstruction method within CIVA software

    Science.gov (United States)

    Rougeron, Gilles; Lambert, Jason; Iakovleva, Ekaterina; Lacassagne, Lionel; Dominguez, Nicolas

    2014-02-01

    This paper presents results of a TFM implementation for Full Matrix Capture acquisitions in CIVA, proposed as a post-processing tool for accurate analysis. This implementation has been made on GPU architecture with OpenCL to minimize the processing time and offer computational device flexibility (GPU/CPU). Examples on immersion configurations on isotropic 2D CAD specimen with planar extrusion are proposed to illustrate the performances. Reconstructions on 2D or 3D areas of direct echoes with mode conversion are allowed. Probe scanning can also be taken into account. Reconstruction results and a benchmark explaining the speedup are presented. Further improvements are also reviewed.

  13. Evaluation of the Influence of Selected Factors on a Successful ERP Software Implementation

    Directory of Open Access Journals (Sweden)

    Grażyna Hołodnik-Janczura

    2010-01-01

    Full Text Available Based on a pilot study, a list of critical success factors for an IT project was developed. These factors were subjected to statistical analysis on the basis of the results of a questionnaire. Two samples of projects, called "success" and "failure" respectively, were analysed. The groups were distinguished in order to evaluate the influence of selected factors on the success of an implementation project. Furthermore, a non-parametric test was applied, in order to conduct statistical data analysis, which confirmed the influence of the majority of the selected factors on the success of an implementation project. (original abstract

  14. Web Implementation of Quality Assurance (QA) for X-ray Units in Balkanic Medical Institutions.

    Science.gov (United States)

    Urošević, Vlade; Ristić, Olga; Milošević, Danijela; Košutić, Duško

    2015-08-01

    Diagnostic radiology is the major contributor to the total dose of the population from all artificial sources. In order to reduce radiation exposure and optimize diagnostic x-ray image quality, it is necessary to increase the quality and efficiency of quality assurance (QA) and audit programs. This work presents a web application providing completely new QA solutions for x-ray modalities and facilities. The software gives complete online information (using European standards) with which the corresponding institutions and individuals can evaluate and control a facility's Radiation Safety and QA program. The software enables storage of all data in one place and sharing the same information (data), regardless of whether the measured data is used by an individual user or by an authorized institution. The software overcomes the distance and time separation of institutions and individuals who take part in QA. Upgrading the software will enable assessment of the medical exposure level to ionizing radiation.

  15. Towards Quality of Service and Resource Aware Robotic Systems through Model-Driven Software Development

    CERN Document Server

    Steck, Andreas

    2010-01-01

    Engineering the software development process in robotics is one of the basic necessities towards industrial-strength service robotic systems. A major challenge is to make the step from code-driven to model-driven systems. This is essential to replace hand-crafted single-unit systems by systems composed out of components with explicitly stated properties. Furthermore, this fosters reuse by separating robotics knowledge from short-cycled implementational technologies. Altogether, this is one but important step towards "able" robots. This paper reports on a model-driven development process for robotic systems. The process consists of a robotics metamodel with first explications of non-functional properties. A model-driven toolchain based on Eclipse provides the model transformation and code generation steps. It also provides design time analysis of resource parameters (e.g. schedulability analysis of realtime tasks) as a first step towards overall resource awareness in the development of integrated robotic syste...

  16. PROCESS QUALITY ANALYSIS OF PERFECTIVE MAINTAINABILITY FOR COMPONENT-BASED SOFTWARE SYSTEMS USING ASPECT-ORIENTED PROGRAMMING TECHNIQUES

    Directory of Open Access Journals (Sweden)

    Jyothi R

    2011-12-01

    Full Text Available Maintainability occupy’s the major role in the software development life cycle (SDLC. Once the software product comes out of this SDLC, major cost and effort goes to the modification/ enhancement of the different components in a component-based software systems. This research presents the modeling work and prototyping techniques, which highlights the importance of process quality analysis for perfective maintainability. This analysis comprising of time, quality and efficiency of the derived solution and it is based on the type ofmodification/enhancement request from the customers of the software product. Here we are proposing a mathematical approach for time evaluation, which includes the summation of response time and the solution production time. Quality analysis is based on the qualitative and quantitative approach. Efficiency analysisrequires the potential operational performance for component-based software systems. This needs high execution speed for handling complex algorithms and huge volumes of data. For this we are proposing an aspect-oriented programming techniques which increases the development speed, modularity, as well as it provides the best software efficiency and design quality.

  17. Software Design Challenges in Time Series Prediction Systems Using Parallel Implementation of Artificial Neural Networks

    Directory of Open Access Journals (Sweden)

    Narayanan Manikandan

    2016-01-01

    Full Text Available Software development life cycle has been characterized by destructive disconnects between activities like planning, analysis, design, and programming. Particularly software developed with prediction based results is always a big challenge for designers. Time series data forecasting like currency exchange, stock prices, and weather report are some of the areas where an extensive research is going on for the last three decades. In the initial days, the problems with financial analysis and prediction were solved by statistical models and methods. For the last two decades, a large number of Artificial Neural Networks based learning models have been proposed to solve the problems of financial data and get accurate results in prediction of the future trends and prices. This paper addressed some architectural design related issues for performance improvement through vectorising the strengths of multivariate econometric time series models and Artificial Neural Networks. It provides an adaptive approach for predicting exchange rates and it can be called hybrid methodology for predicting exchange rates. This framework is tested for finding the accuracy and performance of parallel algorithms used.

  18. Software Design Challenges in Time Series Prediction Systems Using Parallel Implementation of Artificial Neural Networks

    Science.gov (United States)

    Manikandan, Narayanan; Subha, Srinivasan

    2016-01-01

    Software development life cycle has been characterized by destructive disconnects between activities like planning, analysis, design, and programming. Particularly software developed with prediction based results is always a big challenge for designers. Time series data forecasting like currency exchange, stock prices, and weather report are some of the areas where an extensive research is going on for the last three decades. In the initial days, the problems with financial analysis and prediction were solved by statistical models and methods. For the last two decades, a large number of Artificial Neural Networks based learning models have been proposed to solve the problems of financial data and get accurate results in prediction of the future trends and prices. This paper addressed some architectural design related issues for performance improvement through vectorising the strengths of multivariate econometric time series models and Artificial Neural Networks. It provides an adaptive approach for predicting exchange rates and it can be called hybrid methodology for predicting exchange rates. This framework is tested for finding the accuracy and performance of parallel algorithms used. PMID:26881271

  19. Software Design Challenges in Time Series Prediction Systems Using Parallel Implementation of Artificial Neural Networks.

    Science.gov (United States)

    Manikandan, Narayanan; Subha, Srinivasan

    2016-01-01

    Software development life cycle has been characterized by destructive disconnects between activities like planning, analysis, design, and programming. Particularly software developed with prediction based results is always a big challenge for designers. Time series data forecasting like currency exchange, stock prices, and weather report are some of the areas where an extensive research is going on for the last three decades. In the initial days, the problems with financial analysis and prediction were solved by statistical models and methods. For the last two decades, a large number of Artificial Neural Networks based learning models have been proposed to solve the problems of financial data and get accurate results in prediction of the future trends and prices. This paper addressed some architectural design related issues for performance improvement through vectorising the strengths of multivariate econometric time series models and Artificial Neural Networks. It provides an adaptive approach for predicting exchange rates and it can be called hybrid methodology for predicting exchange rates. This framework is tested for finding the accuracy and performance of parallel algorithms used.

  20. Helping organizations to address their effort toward the implementation of improvements in their software process

    Directory of Open Access Journals (Sweden)

    Mirna Ariadna Muñoz-Mata

    2015-01-01

    Full Text Available Debido a la importancia de las pequeñas y medianas empresas (pymes de desarrollo de software para la industria del software, es importante garantizar la calidad de sus productos, para lo cual la mejora de sus procesos ofrece un camino atractivo para lograrlo. Desafortunadamente, la falta de conocimientos que existe en las pymes sobre cómo o dónde dirigir su esfuerzo de mejora hace que la implementación de mejoras sea un camino lleno de obstáculos, la mayoría de veces imposible de lograr. Este artículo presenta un marco de trabajo cuyo objetivo es ayudar a las pymes a encontrar un punto de inicio para dirigir sus esfuerzos de mejora, basado en la identifi cación de sus problemas principales. Además, el marco de trabajo proporciona una guía para iniciar la implementación de la mejora (punto de inicio, referente a los modelos, estándares y/o metodologías ágiles que pueden ser usados como referencia, basada en sus necesidades, características y cultura de trabajo. Finalmente, el artículo incluye un caso de estudio en el que se muestran los resultados de la implementación del marco de trabajo en cuatro pymes de la región de Zacatecas.